© Copyright. Saul Carliner. 2000-2002. We Are What We Measure: Productivity and Effectiveness...
Transcript of © Copyright. Saul Carliner. 2000-2002. We Are What We Measure: Productivity and Effectiveness...
© Copyright. Saul Carliner. 2000-
2002.
Agenda
ExerciseAbout the Study12 Lessons and InsightsApplying the Lessons
© Copyright. Saul Carliner. 2000-
2002.
Clarifying
“We”Helen ConstantinedesKirk St.AmantCatherine WalstadSTC
© Copyright. Saul Carliner. 2000-
2002.
Exercise 1: The Request You’re Director of Technical Communication, reporting to
the Vice-President of Development for a telecommunications firm.
Your staff: numbers 76, including 5 managers, 50 writers, 5 planners, 5 editors, 5 programmers, 5 graphic illustrators, and the Director.
Your work: mostly PDF files—about 75,000 pages per year. Also maintains 12 technical web sites (average of 37 topics each) , participate in the design of user interfaces for some of the software, and produce online marketing materials (prepared one such project last year).
By his own admission, the new Vice-President has a limited understanding of technical communication and has suggested that you track metrics for your organization.
What do you track?
© Copyright. Saul Carliner. 2000-
2002.
Some Definitions
ProductivityEffectivenessMetrics AssessmentEvaluationProof
© Copyright. Saul Carliner. 2000-
2002.
About the Studies
Personal desire to find out how well I’m doing
Best practices often emerge from non-competitors
© Copyright. Saul Carliner. 2000-
2002.
Details: Part One
Beginning of a larger study Purpose
– Identify metrics used by related professions
– Assess their transferability to technical communication
Reviewed literature in software engineering, training, marketing and technical communication, public relations
© Copyright. Saul Carliner. 2000-
2002.
Details: Part Two
Studied larger technical communication departments
Questioned about project, people, and business management
Insufficient response for statistical validity but..
Some responses were abundantly clear
© Copyright. Saul Carliner. 2000-
2002.
2. Metrics Involve Several Measures
No single measure suffices Differences in assessing
Types of products IndividualsTypes of projects DepartmentsEffectiveness Productivity
© Copyright. Saul Carliner. 2000-
2002.
3. Simple Measures Don’t
(in fact, no standard practicesacross organizations in technicalcommunication) Page counts Topics written Lines of codes written Student days
© Copyright. Saul Carliner. 2000-
2002.
Issues
Producing better often means producing less
As number of media grow, no single count works
© Copyright. Saul Carliner. 2000-
2002.
4. You Cannot Wait Until Completion to Begin
Showing a change in performance usually requires a before:after comparison: – Effectiveness of communication product– Productivity of development effort
To be credible, measurement cannot begin after-the-fact
© Copyright. Saul Carliner. 2000-
2002.
In Technical Communication
Evaluation is not a standard business activity.
Until that changes, measures unlikely to have credibility.
© Copyright. Saul Carliner. 2000-
2002.
5. Metrics Assume You Follow a Documented
Process Preferably, an industry-standard one Software Capability Maturity Model Instructional Systems Design
Process Analysis-Design-Placement
Approach
© Copyright. Saul Carliner. 2000-
2002.
In Technical Communication
A standard process seems to exist Focused on project management
rather than design and development issues
Requirements-Development- “Checks”-Release
© Copyright. Saul Carliner. 2000-
2002.
6. (starting it, anyway)
Why do organizations commission us to write information for users?
Hint: It’s not for usability.
© Copyright. Saul Carliner. 2000-
2002.
Organizations Hire Us to Produce Business Results
Why we’re commissionedGenerate revenueContain expensesComply with regulations
© Copyright. Saul Carliner. 2000-
2002.
6. To Assess Effectiveness, Must Begin with Clear
GoalsState goals in observable and measurable terms: For intended business results For the content
© Copyright. Saul Carliner. 2000-
2002.
In an Interesting Paradox in Technical Communication
“Users performing intended tasks” and “business results” are commonly stated measures of effectiveness
Projects do not begin with content or business objectives
Usability testing is rarely performed
© Copyright. Saul Carliner. 2000-
2002.
7. To Assess Productivity, You Must Benchmark Against the
Industry Industry surveys provide benchmarks: TRAINING Industry Survey IDG Surveys Merrill Lynch and other investment
firms AD Week AC Nielsen
© Copyright. Saul Carliner. 2000-
2002.
Unfortunately
We have no comparable survey in technical communication.
© Copyright. Saul Carliner. 2000-
2002.
8. Industry Standard Models Build Acceptance
They also improve methods Examples
– Kirkpatrick Model – Direct response measures– cpm
© Copyright. Saul Carliner. 2000-
2002.
Question
You receive a resume. The applicant is one of the best you’ve seen.
But you find a visible typo on the resume.
Would you hire the candidate or not?
© Copyright. Saul Carliner. 2000-
2002.
In Technical Communication
Most common method of assessing effectiveness is technical reviews
Use of Readers’ Comment Forms is low
© Copyright. Saul Carliner. 2000-
2002.
9. Measurements of Technique Are Not Persuasive
They assess neither effectiveness nor productivity
Although research might indicate which characteristics correlate with effectiveness, they do not guarantee it
© Copyright. Saul Carliner. 2000-
2002.
Examples
White space Typographical errors Number of index entries per
page Number of illustrations per
page Choice of type font
© Copyright. Saul Carliner. 2000-
2002.
When Possible, Use Financially Based Measures
Sales generated Billings Registrations Billings
© Copyright. Saul Carliner. 2000-
2002.
10. Intangibles Matter to Executives
PerceptionsFeelingsFrustrationName recognitionAlso collect descriptive data
in addition to numeric data
© Copyright. Saul Carliner. 2000-
2002.
Use of Data
Numerical data indicates the extent
Descriptive data tells the impact
© Copyright. Saul Carliner. 2000-
2002.
11. Collect Data on an Ongoing Basis
Devote between 2 and 5 percent of a project budget to evaluation
© Copyright. Saul Carliner. 2000-
2002.
Issues
One-time studiesDesperation studiesMetrics for tyranny rather
than as a tool for continuous improvement
© Copyright. Saul Carliner. 2000-
2002.
12. On Their Own, Metrics Prove Nothing
Metrics build perceptions Data supports those perceptions Data promotes informed decision-
making Part of a larger system of building
and maintaining relationships with sponsors
© Copyright. Saul Carliner. 2000-
2002.
In Other Words
Metrics are a value system.
We choose to measure things we believe to be important.
© Copyright. Saul Carliner. 2000-
2002.
A Successful Metrics Program
Tracks measurements that are important to us as technical communicators and meaningful to our sponsors