Collecting Data from Users. Uses User and task analysis Prototype testing On-going evaluation and...

63
Collecting Data from Users

Transcript of Collecting Data from Users. Uses User and task analysis Prototype testing On-going evaluation and...

Collecting Data from Users

Uses

• User and task analysis• Prototype testing• On-going evaluation and re-design

Data collection methods

• Questionnaires

• Interviews

• Focus groups

Definitions• Survey:

– (n): A gathering of a sample of data or opinions considered to be representative of a whole.

– (v): To conduct a statistical survey on. • Questionnaire: (n) A form containing a set of

questions, especially one addressed to a statistically significant number of subjects as a way of gathering information for a survey.

• Interview – (n): A conversation, such as one conducted by a

reporter, in which facts or statements are elicited from another.

– (v) To obtain an interview from. – American Heritage Dictionary

Surveying

• Sample selection• Questionnaire construction• Data collection• Data analysis

Surveys – detailed steps• determine purpose, information needed• identify target audience(s)• Select method of administration • design sampling method• design prelim questionnaire

– including analysis– Often based on unstructured or semi-structured

interviews with people like your respondents

• pretest• Revise, pretest…• administer• analyze

Why survey as method?

• Answers from many people, including those at a distance

• Relatively easy to administer, analyze

• Can continue for a long time

Surveys can collect data on:• Facts

– Characteristics of respondents– Self-reported behavior

• This instance• Generally/usually

• Opinions and attitudes:– Preferences, opinions, satisfaction,

concerns

Some Limits of Surveys

• Reaching users is easier than non-users• Relies on voluntary cooperation, possibly

biasing the sample• Questions have to be unambiguous,

amenable to short answers• You only get answers to the questions you

ask; you generally don’t get explanations• The longer or more complex the survey the

less cooperation

Some sources of error

• Sample• Question choice

– Can respondents answer?

• Question wording• Method of administration• Inferences from the data• Users’ interests in influencing results

– “vote and view the results”CNN quick vote: http://www.cnn.com/

When to do interviews?

• Need details that can’t get from survey• Need more open-ended discussions with

users• Small #s OK• Can identify and gain cooperation from

target group• Sometimes: want to influence

respondents as well as get info from them

Sample selection

Targeting respondents

• About whom do you want information?

• About whom can you get information?– E.g. non-users are hard to reach

Sampling terminology

• Element: the unit about which info is collected; basis of analysis. E.g., “User”

• Universe: hypothetical aggregation of all elements. “All users”

• Population: a specified aggregation of survey elements. “People who have used this service at least once in the last year.”

• Survey population: aggregate of elements from which the sample is selected. “People who use the service at least once during the survey period.”

Terminology, cont.• Sampling unit: elements considered for selection.

• Sampling frame: list of sampling units. • Observation unit: unit about which data is collected.

Often the same as unit of analysis but not always. E.g. one person (observational unit) may be asked about the household (unit of analysis).

• Variable: a set of mutually exclusive characteristics such as sex, age, employment status.

• Parameter: summary description of a given variable in a population.

• Statistics: summary description of a given variable in a sample.

Sample design• Probability samples

– random– stratified random– cluster– Systematic– Size: if 10/90% split, 100; if 50/50, 400;

• Yale’s slide• 30-50 in each cell

– GOAL: Representative sample

• Non-probability sampling– convenience sampling– purposive sampling– quota sampling

Representative samples

• Which characteristics matter? • Want the sample to be roughly

proportional to the population in terms of groups that matter

• E.g., students by gender and grad/undergrad status:

• http://opa.vcbf.berkeley.edu/IC/Campus.Stats/CampStats_F00/CS.F00.Table.F2.htmgraduate/undergraduate…a

Active vs passive sampling

• active: solicit respondents– Send out email– Phone– Otherwise reach out to them– Follow up on non-respondents if possible

• passive – e.g. on web site– Response rate may be unmeasurable– heavy users may be over-represented– Disgruntled and/or happy users over-

represented

Response Rates

• low rates may > bias– Whom did you miss? Why?

• How much is enough? – Babbie: 50% is adequate; 70% is very good

• May help if they understand purpose– Don’t underestimate altruism

• Incentives may increase response– Reporting back to respondents as a way of getting

response

Example of a careful sample design

• http://www.pewinternet.org/reports/reports.asp?Report=55&Section=ReportLevel1&Field=Level1ID&ID=248

RECAP

• We collect data from users/potential users for:– User and task analysis– Prototype testing– On-going evaluation and re-design

• Methods include:– Surveys, interviews, focus groups

• Different methods useful for different purposes

Some Issues Common to Different Methods

• Know your purpose! – Match method to purpose and

feasibility• Whom do you need/can you get to

participate?– Population, sample composition,

sample selection methods, size– Response rate, respondent

characteristics (bias)

Common Issues, cont.

• What do you need to know?– What can/will respondents tell you?– How will it help?

• How do you ask what you want to know?– Question construction, wording– Question ordering

• What do you do with results?– Reporting and analysis

Types of web surveys

• Comprehensive • Quick polls – focused, one or few

questions– http://www.gomez.com/ratings/index.cfm?

topcat_id=19&firm_id=1768&CFID=295681&CFTOKEN=8136111

• Short, focused surveys• Guestbooks, user registration, user

feedback– http://www.bookfinder.com/interact/comments/

Uses of surveys of web sites

• Identify users– Describe their characteristics

• Describe their behavior• Ask their needs, preferences• Assess user satisfaction/response• Identify user problems,

dissatisfactions• Solicit ideas for improvement

Problems with Web Surveys

• Population?– Size– Characteristics

• Response rate? – Multiple responses from same person OK?

• Biased sample?• Users competent to answer?

– E.g., will they answer after they have used the site enough to be able to judge?

• Non-users not represented; infrequent users under-represented?

Questionnaire construction

Questionnaire construction

• Content– Goals of study: What do you need to know?– What can people tell you?

• Conceptualization • Operationalization – e.g., how do you

define “user”? • Question design• Question ordering

Topics addressed by surveys

• Respondent (user) characteristics• Respondent behavior• Respondent opinions, perceptions,

preferences, evaluations

Respondent characteristics• Demographics

– General http://www.gvu.gatech.edu/user_surveys/survey-1998-10/questions/general.html

– Professional role (manager, student..)

• User role (e.g., buyer, browser…)• Expertise – hard to ask

– Domain– technology

• http://www.gvu.gatech.edu/user_surveys/survey-1998-10/questions/general.html

• http://www.gvu.gatech.edu/user_surveys/survey-1998-10/questions/general.html

– System

Behavior

• Tasks (e.g., for what are they using this site?)

• Site usage, activity– Frequency; common functions – hard

to answer accurately– Self-reports vs observations

• Web and internet use: Pew study

Opinions, preferences, concerns

• content• Organization, architecture• interface• perceived needs • concerns

– http://www.gvu.gatech.edu/user_surveys/survey-1998-10/questions/privacy.html

• Success, satisfaction– Subdivided by part of site, task, purpose…

• other requirements• Suggestions

Problems with questions

• Social acceptability• Privacy

– “Do you use the internet to view pornography?”

• Difficulty wording unambiguously• International concerns

– Language– Social acceptability

Problems: topics hard to conceptualize, operationalize

• e.g., “Why did you use the CDL today?”– Teaching, research…

• my category of work or task– to find an electronic journal, locate a book, find a

citation…• To locate kind of resource

– to find material on x topic• the subject area

– to save me time• i.e. I used this rather than another way of

accessing same resource

Question construction

Question formulation• match respondents’ language• match respondents’ behavior• what do they want to tell you?• What can they tell you?

– Recent CNN.com poll: “do you think the sentence given to x was too long, about right, too short.” What was the sentence? (What was the crime?)

• Beware of compound questions, hidden assumptions:– ‘did you order something? If so, did you….’ –

what if only browsing?

Question formulation: simplicity and clarity

Complete the following sentence in the way that comes closest to your own views: 'Since getting on the Internet, I have ...'

• ... become MORE connected with people like me.

• ... become LESS connected with people like me.

• ... become EQUALLY connected with people like me.

• ... Don't know/No answer.http://www.gvu.gatech.edu/user_surveys/survey-1998-10/questions/general.html

Survey questions – format• open-ended

http://www.bookfinder.com/interact/comments/– “What is your job title?” _______– “What do you find most difficult about your

job?”• closed-ended (one answer; multiple answers)

– http://www.useit.com/papers/surveys.html“Select the range that best represents the

total number of staff…”1-2 3-5 6-10…

– paired characteristics/semantic differentialFriendly__|___|___|___|____| Unfriendly

Question format, cont.

• Ordinal scale/ “Likert scale”Strongly agree, agree, neutral, disagree,

strongly disagreehttp://www.amazon.com/exec/obidos/tg/

stores/detail/-/books/073571102X/rate-this-item/104-0765616-4703139

– Rating scale – matrix (usability survey, qn 5)

• Always include neutral, “other,” “N/A” (not applicable)

Closed-ended questions• Answers need to be comprehensive and mutually

exclusive;• OR

– Allow people to give more than one answer– Tell them how to choose“Why did you come to SIMS?”

• How many responses to allow? – As many as your respondent needs– Likert scale-like questions:

• 5 OR 7 is usual; can respondent differentiate 7?• Very strongly agree; strongly agree; agree;

neutral; disagree; strongly disagree; very strongly disagree?

– Odd vs even: even allows neutral, odd forces a choices.

Filter and Contingency questions

• 1. “Have you ever used our competitor?”– “If no, go to question 3.”– 2. “If yes, how would you rate….”

• Did you apply to any graduate program other than SIMS? – If yes…– If no…

Layout: consistency

Circle the answer that best matches your opinion.

SIMS is a great place to study.Strongly agree Agree Neutral Disagree Strongly disagree

UC Berkeley is a great university.Strongly agree Agree Neutral Disagree Strongly disagree

IS214 is a great course.Strongly disagree Disagree Neutral Agree Strongly agree

Include Instructions!

• One answer to each question, or multiples? – If online, can program to test and let user

know they have violated guidelines.

• Respond every time you visit this site?• Be sure it’s clear what to do if a

question does not apply.– “Why did you choose SIMS over any other

programs that accepted you?”

Question ordering

• Group similar questions• Using headings to label parts of

survey, topics• Funneling:

– General to particular– Particular to general

• Keep it short! Response falls off with length.

Pre-test, pre-test, pre-test!

• With people like expected respondents• Looking for:

– Ambiguous wording– Missing responses– Mismatch between your expectations and

their reality– Any other difficulties respondents will have

answering, or you will have interpreting their responses

Web Survey Problems• Who is the population? Self-selected sample.• Stuffing the ballot box

– Cnn.com polls• How to know what response rate is• How to get responses (1) after they have

used site (2) before they leave• What are you assessing and what are

they responding to?– E.g., design of site, or content? Presentation

of content, or content of content?• Loss of context – what exactly are you

asking about, what are they responding to?– Are you reaching them at the appropriate

point in their interaction with site?

Web Survey Problems II

• Incomplete responses -- avoiding blank responses – and annoying users

• Respondents may not know how long the survey is– Let people know

• Multiple submissions

Survey administration

Human subjects considerations

• not hurt by the process itself (e.g., uncomfortable questions)

• not hurt by uses of the data– job etc not in jeopardy– not at risk for criminal prosecution

• confidentiality or anonymity• Consent, voluntary participation• http://cphs.berkeley.edu

Ways of administering surveys• On a web site

– Open access– Limited access

• Email (with URL)• Phone• Paper

– Mail– Users pick up on their own– Hand out in person– Fax

• Face to face

Resources for online surveys

Examples:• Zoomerang

– http://www.zoomerang.com/

• Perseus Development Corp.– http://www.perseusdevelopment.com/

• Websurveys.net – http://www.web-surveys.net/

Telephone surveys

• Reach a broader cross-section, more representative sample– Non-internet users, people outside

your usual group (whatever that is)– More chance of persuading people to

participate• More labor intensive• More intrusive

Other Active Methods of Getting User Info

• User journals (e.g., Arbitron ratings)

Keeping track of responses

• Mail or phone procedure: If you have a list of people contacted, and a way of knowing who has responded, can follow up with those who have not.– Pew tried phone survey 10 times, varying

times of day.– Trying to maximize response rate– Characterizing non-respondents– Privacy sensitivity? E.g., sealed envelopes.

• If no list, a reminder that says, if you have not yet responded, please do so.

Reporting and analysis

Reporting

• Purpose• Methods• Data• Interpretation• Conclusions, implications

Easy method: Questionnaire with the responses filled in

Q5 Do you use a computer at your workplace, at school, at home, or anywhere else on at least an occasional basis?

% 64 Yes 36 No * Don’t know/Refused

Q6 Do you ever go online to access the Internet or World Wide Web or to send and receive email?

% 49 Goes online 15 Does not go online * Don’t know 36 Not a computer user

http://www.pewinternet.org/reports/pdfs/Report1Question.pdf

Analysis• Univariate descriptive statistics

– Frequency;• table: note that they quote the question exactly as asked

– http://www.harrisinteractive.com/harris_poll/index.asp?PID=260

• Pie chart http://www.cc.gatech.edu/gvu/user_surveys/papers/1997-10/sld015.htm

– http://pm.netratings.com/nnpm/owa/NRpublicreports.toppropertiesweekly

– Mean– Standard deviation– Quartiles– Max, min

Analysis

• Bi-variate– Cross-tabs http://www.census.gov/population/estimates/state/st-99-1.txt,

• http://www.cc.gatech.edu/gvu/user_surveys/papers/1997-10/sld019.htm

– Correlations– Graphs

http://www.cc.gatech.edu/gvu/user_surveys/papers/1997-10/sld006.htm

• http://www.cc.gatech.edu/gvu/user_surveys/survey-1998-10/graphs/general/q54.htm

Types of measures• Dichotomous

– Yes/no; male/female– Can report: frequencies

• Nominal– Favorite color; most frequent activity; race/ethnicity– Frequencies, mode

• Ordinal– Strongly agree, agree, neutral, disagree, strongly disagree– Frequencies, mode, median

• Integer/ratio– # of staff; salary– can compare values not only in terms of which is larger or

smaller, but also how much larger or smaller one is.– ratio variables that have a natural zero point, e.g., weight,

length– Integer variables do not, e.g. temperature. 100°F is not twice

as warm as 50°F.– Frequencies, mode, median, mean.

CrosstabsUndergrads(n=120)

%

Grads(n=200)

%

Total(n=320)

%

Satisfied

6071

1325

3196

Dissatis. 4047

87165

69127

Total 100%n = 118

100%n = 190

100%n = 308

No ans. n = 2 n = 10 n= 12

Confidence intervals

• http://www.pewinternet.org/reports/reports.asp?Report=55&Section=ReportLevel1&Field=Level1ID&ID=248