Tomas Petricek
email: t.petricek@kent.ac.uk
twitter: @tomaspetricek
office: S129A
1. Usability | 2. Accessibility |
3. Operations | 4. Ethics |
1. Usability | 2. Accessibility |
3. Operations | 4. Ethics |
Does the user interface work?
Can users enable cookies they want?
Are there any usability issues?
Are users more efficient with a new product?
Does a design encourage new way of thinking?
How do we know that what we've done works?
1. Usability | 2. Accessibility |
3. Operations | 4. Ethics |
Taking accessibility seriously
What if you need to develop a web that works for everyone?
Partial list from government accessibility site
1. Usability | 2. Accessibility |
3. Operations | 4. Ethics |
Framing - what we consider and what we ignore
Alternatives - what decisions can you make?
Uncertainties - could the outcome be by accident?
Value - how do you evaluate outcomes?
Decision maker - need one person to decide
Commitment to action - will we actually do it?
1. Usability | 2. Accessibility |
3. Operations | 4. Ethics |
Should we run all experiments?
But we're just evaluating two versions of an algorithm!
Studies directly involving users
Analytics and remote testing
Data is what we collect
Usage statistics, survey data, performance
For example, visited pages on a web site
Information is what we get by analysis
What groups of users follow what paths
Say, phone users never reach a certain page
Conclusions is what we decide to do
Interpretation of reasons behind information
For example, make a prominent navigation link!
Data can range from numbers to written reports
Interviews (structured, semi-structured or open)
Questionnaires (offline or online)
Observation (in the lab or in the wild)
Logging (usage data about computer system)
Qualitative data analysis
Quantitative data analysis
Qualitative data analysis
Quantitative data analysis
Testing a hypothesis
Producing reproducible results
Detecting a tank using neural networks
Tank photos on sunny day, other photos rainy
Time to completion, number or scale of errors
Recall of presented information
Labs for fully controlled environment
Amazon Mechanical Turk for online studies
Making sense of numbers
What information people know?
What is a good analogy?
Does analogy help later recall?
About 350 Amazon MTurk workers
E1: Compare different multipliers and states
E2: Is best numerical fit or home state better?
E3: Repeat after 6 weeks without perspective
Does interactivity make you think?
https://dataviz-study.azurewebsites.net/demo/step1
Test recall of numerical values from a newspaper article
Interactive, static and text-only versions
"The article listed top 5 categories of products (...).
The two largest categories accounted for roughly
the same amount of total exports. Please select
the two top categories."
Group | Both | Mach | Trans | None |
---|---|---|---|---|
Interactive | 20 | 9 | 1 | 1 |
Static chart | 17 | 14 | 1 | 0 |
Text only | 11 | 19 | 4 | 0 |
Is the interactive version significantly better?
The p-value is the probability of finding the observed, or more extreme, results when the null hypothesis of a study question is true.
A small p-value (\(\leq\) 0.05) indicates strong evidence against the null hypothesis, so you reject the null hypothesis.
See also: Analysis of variance (ANOVA), Chi-squared test
Taking evaluation seriously
Medical registered randomized control trials
Compare against control group with placebo
Assign participants randomly to groups
Use consistent categorization of effects
Register study protocol in advance
What is the purpose?
Study methodology
Giving observation more structure
How do people
learn advanced
Excel features?
Observe users
during their work
Ask follow-up interview questions to clarify
Real-world use of an early prototype
Collect quantitative and qualitative data
Theoretical frameworks to guide the study
Think-aloud, diaries, logs and analytics
Evaluation without participants
Experimental methods
Can be used with early prototypes
Heuristic evaluation using 10 rules
Simulating user's problem solving process
Five evaluators find 75% of problems
Different rules for different products
Web page analytics
Statistics on visitors and page views
Where they come from, how they behave?
Pages with the largest number of views
Language, country and device used
Number of visitors, visits and views
Traffic sources and search keywords
A/B testing
What version is more effective?
Google in 2001: How many results to show?
Measure clicks, purchases or likes
Needs a product you can change and users
Statistical analysis of significance
What to evaluate
Lo-fi prototype - limited, but easy to do early
Running software - need to write it first, but realistic
Where to evaluate
In a lab - controlled, but may not be relevant
In the wild - less precise, but may say useful things
When to evaluate
Early in the process - get feedback for development
Finished product - assess the quality of the work
If I had asked people what they wanted, they would have said faster horses.
(Incorrectly attributed
to Henry Ford)
You can test only what you can imagine
Users are talking in well understood terms
Quantitative can only help you choose
Can make it hard to see new ideas
Evaluation in controlled setting
Precise answers to narrow questions
Test usability, compare methods
Evaluation in natural setting
Broad answers to interesting questions
How people really use your product
Evaluation without direct participants
Less expensive, but limited questions
Expert analysis or data collection
What you should remember from this lecture
Tomas Petricek
t.petricek@kent.ac.uk | @tomaspetricek
Books
Papers and links