lecture 14: evaluation techniques november 30, 2015 sds 235 visual analytics
TRANSCRIPT
![Page 1: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/1.jpg)
LECTURE 14:
EVALUATION TECHNIQUES
November 30, 2015
SDS 235
Visual Analytics
![Page 2: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/2.jpg)
Announcements 1/2
• Friday 4:30pm: Amanda Cox invited lecture – Ford 240 (extra credit for attending and posting to Piazza)
• Schedule change: - Evaluation Methods today- Self-Critique on Wednesday
• Final Project Demonstration will be held the evening of December 14th, time TBA (regularly scheduled class will be cancelled)
![Page 3: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/3.jpg)
Announcements 2/2
• End of the semester is nearly upon us (T-minus 2 weeks!)
• Some of you might be starting to stress about grades
• However: if you would like a personalized PDF grade sheet of your progress in the course so far, email me
#ProTip
![Page 4: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/4.jpg)
Discussion
How do we measure the effectiveness
of a visualization system?
![Page 5: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/5.jpg)
Tufte, 1983
“Above all else, show the data.”
![Page 6: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/6.jpg)
Tufte, 1983
![Page 7: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/7.jpg)
Examples
![Page 8: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/8.jpg)
Examples
![Page 9: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/9.jpg)
Discussion
• What do you think of the Data-Ink Ratio?
• Consider ways to maximize the ratio…
![Page 10: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/10.jpg)
Flashback: Epistemic Action
The purpose of some actions is not the effect they have on the environment but the effect they have on the humans.
![Page 11: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/11.jpg)
ChartJunk and Recall
Bateman et al. “Useful Junk? The Effects of Visual Embellishment on Comprehension and Memorability of Charts”, CHI 2010
![Page 12: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/12.jpg)
ChartJunk and Eye Gaze
![Page 13: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/13.jpg)
Discussion
• Have you seen particularly compelling examples of “visual embellishment”? Tragic ones?
• What’s the right balance between Tufte and ChartJunk?
![Page 14: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/14.jpg)
Nested Model of VIS Design (Munzner, 2009)
Munzner, Tamara. "A nested model for visualization design and validation." Visualization and Computer Graphics, IEEE Transactions on 15.6 (2009): 921-928.
![Page 15: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/15.jpg)
Threats
![Page 16: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/16.jpg)
Evaluation “Threats”
• Mismatch: a common problem in evaluating VIS systems• Examples:
- The contribution of a new visual encoding cannot be tested using a quantitative timing of the algorithm
- A mischaracterized task cannot be addressed in a formal lab study
![Page 17: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/17.jpg)
Matching Methods and Metrics
![Page 18: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/18.jpg)
Insight-Based Evaluation (North et. al, 2005)
Measure the utility of a visualization system by
counting the number of insights generated by the user
while using it
![Page 19: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/19.jpg)
Insight-Based Evaluation Method
• No “benchmark tasks”• Training on data and visualization for 15 minutes• Participants list questions that they would like to pursue• Asked to examine the data for as long as necessary until
no new insights can be gained• During analysis, the users asked to comment on their
observations, inferences, and conclusions
![Page 20: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/20.jpg)
Evaluating the Results
• The number of insights are tallied• Insights: distinct observations about the data by each
participant• Collect all insights generated by all participants as a
baseline• Various quantitative statistics collected on insight
generation (time spent, time to first insight, etc.)
![Page 21: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/21.jpg)
What does insight-based evaluation address?
![Page 22: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/22.jpg)
Problem: defining “insight”
North’s definition:
“[Insight is] an individual observation about the data by the participant, a unit of discovery. It is straightforward to recognize insight occurrences in a think-aloud protocol as any data observation that the user mentions is considered an insight.”
![Page 23: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/23.jpg)
Example 1
“Our tool allows the biologists to interactively visualize and explore the whole set of trees, providing insight into the overall distribution and possible conflicting hypothesis”
Insight = knowledge about the overall distribution
![Page 24: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/24.jpg)
Example 2
“The analyst determined the answers to these questions, but also came up with further insights that she shared with people from other administrative units. She used the discovered information to advise other administrators of certain previously unknown relationships in their data”
Insight = information about previously unknown relationships
![Page 25: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/25.jpg)
Cognitive Science Definition
• Something measureable in the front lobes and the temporal lobes (superior temporal gyrus).
• Spontaneous Insight vs. Model-building Insight
![Page 26: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/26.jpg)
Disambiguating “Insight”
• Knowledge-building insight:- Discovering insight, gaining insight, and providing insight - Insight as a substance, that accumulates over time and could
be measured/quantified
• Spontaneous insight:- Experiencing insight, having an insight, or a moment of insight - Insight as a discrete event, that occurs at a specific moment in
time and could be observed
![Page 27: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/27.jpg)
Discussion
• Can we measure knowledge-building insight?• Can we measure spontaneous insight?• Are they related?
![Page 28: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/28.jpg)
Questions?
![Page 29: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/29.jpg)
• Multi-dimensional In-depth Long-term Case studies• Hypothesis: the efficacy of tools can be assessed by
documenting:- Usage (observations, interviews, surveys, logging, etc.)- How successful the users are in achieving their professional goals
MILCs – Shneiderman and Plaisant (2006)
![Page 30: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/30.jpg)
What do MILCs address?
![Page 31: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/31.jpg)
Definition
• Multi-dimensional: using observations, interviews, surveys, and loggers
• In-Depth: intense engagement of the researchers with the expert users to the point of becoming a partner or assistant
• Long-term: longitudinal studies that begin with training in use of a specific tool through proficient usage that leads to strategy changes for the expert users.
• Case studies: detailed reporting about a small number of individuals working on their own problems, in their own environment
![Page 32: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/32.jpg)
Motivation
• MILCs has been embraced by a small community of researchers interested in studying creativity support tools.
• Challenges: - Cannot control for the users - Cannot control for the tasks- Toy problems in laboratories are not indicative of real-world
problems and environments
![Page 33: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/33.jpg)
Execution issues with MILCs
• Duration is always a problem• Number of participants has to be small• Familiarities are difficult
- Understand organization policies and work culture- Gain access and permission to observe or interview- Observe users in their workplace, and collect subjective and
objective quantitative and qualitative data.- Compile data of all types in all dimensions- Interpret the results- Isolate factors- Need to repeat the process
![Page 34: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/34.jpg)
Questions?
![Page 35: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/35.jpg)
Grounded Evaluation
• Proposed by Petra Isenberg et al. (2008)
• Definition:• A process that attempts to ensure that the evaluation of an
information visualization is situated within the context of its intended use
![Page 36: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/36.jpg)
Call For Qualitative Measurements
![Page 37: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/37.jpg)
Questions?
![Page 38: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/38.jpg)
Learning-Based Evaluation (Chang, 2010)
• Working assumption: “the goal of visualization is to gain insight and knowledge”
• Proposal: we should evaluate a visualization based on whether or not the user actually gains insight or knowledge after using a visualization
![Page 39: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/39.jpg)
Similar to Learning in Education
• How would an instructor choose between two textbooks for a course?
• We could:- Ask the students which book they prefer- Issue: they might like a book because its cover is pretty
- Ask colleagues what book they prefer- Issue: different students in different environments
- Ask the students to find some information in the book and measure how quickly they can perform the task
- Issue: this only demonstrates how well the book is organized.
![Page 40: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/40.jpg)
Metaphor for Visualization Evaluation
• In a best case scenario, we would:- Ask half of the student to use book one to learn a subject- Ask other half to use another book to learn the same subject
• Then we give the two groups the same test, and whichever scores higher “wins”
![Page 41: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/41.jpg)
Flow Chart: Traditional LBE
![Page 42: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/42.jpg)
Discussion
• Do you see any problems with this method?
![Page 43: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/43.jpg)
Flow Chart: Single-System LBE
![Page 44: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/44.jpg)
Discussion
How should we evaluate your final projects?
![Page 45: LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics](https://reader035.vdocument.in/reader035/viewer/2022062519/5697bfb61a28abf838c9e1f9/html5/thumbnails/45.jpg)
Reminders / Administrivia
• On Wednesday: “Self-critique and feedback”
• Final Project Demonstration Day: 14 December, time TBA- Feel free to invite friends, colleagues, professors, etc.- If you need additional hardware (monitor, mouse, etc.), email me
• Final deliverable:- Short (~2-page) write up on your completed project- Guidelines will be posted to course website next week- Officially due last day of finals: 22 December by 5pm- Consider submitting to conferences and undergrad research
competitions! More details on venues to come