measuring collaborative measuring collaborative cognition

51
Measuring Collaborative Measuring Collaborative Cognition Cognition Nancy J. Cooke Nancy J. Cooke Arizona State Arizona State University University CKM Workshop CKM Workshop January 14 January 14 - - 16, 2003 16, 2003

Upload: others

Post on 22-Nov-2021

8 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Measuring Collaborative Measuring Collaborative Cognition

Measuring Collaborative Measuring Collaborative CognitionCognition

Nancy J. CookeNancy J. CookeArizona State Arizona State

UniversityUniversity

CKM WorkshopCKM WorkshopJanuary 14January 14--16, 200316, 2003

Page 2: Measuring Collaborative Measuring Collaborative Cognition

Report Documentation Page Form ApprovedOMB No. 0704-0188

Public reporting burden for the collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering andmaintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this collection of information,including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 1215 Jefferson Davis Highway, Suite 1204, ArlingtonVA 22202-4302. Respondents should be aware that notwithstanding any other provision of law, no person shall be subject to a penalty for failing to comply with a collection of information if itdoes not display a currently valid OMB control number.

1. REPORT DATE JAN 2003 2. REPORT TYPE

3. DATES COVERED 00-00-2003 to 00-00-2003

4. TITLE AND SUBTITLE Measuring Collaborative Cognition

5a. CONTRACT NUMBER

5b. GRANT NUMBER

5c. PROGRAM ELEMENT NUMBER

6. AUTHOR(S) 5d. PROJECT NUMBER

5e. TASK NUMBER

5f. WORK UNIT NUMBER

7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Arizona State University,Tempe,AZ,85287

8. PERFORMING ORGANIZATIONREPORT NUMBER

9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) 10. SPONSOR/MONITOR’S ACRONYM(S)

11. SPONSOR/MONITOR’S REPORT NUMBER(S)

12. DISTRIBUTION/AVAILABILITY STATEMENT Approved for public release; distribution unlimited

13. SUPPLEMENTARY NOTES Collaboration and Knowledge Management (CKM) Workshop, 14-16 Jan 2003, College Park, MD. U.S.Government or Federal Rights License

14. ABSTRACT

15. SUBJECT TERMS

16. SECURITY CLASSIFICATION OF: 17. LIMITATION OF ABSTRACT Same as

Report (SAR)

18. NUMBEROF PAGES

50

19a. NAME OFRESPONSIBLE PERSON

a. REPORT unclassified

b. ABSTRACT unclassified

c. THIS PAGE unclassified

Standard Form 298 (Rev. 8-98) Prescribed by ANSI Std Z39-18

Page 3: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 22

AcknowledgementsAcknowledgements NMSU Faculty: Peter Foltz

NMSU Post Doc: Brian Bell

NMSU Graduate Students: Janie DeJoode, Jamie Gorman, Preston Kiekel, Rebecca Keith, Melanie Martin, Harry Pedersen

US Positioning, LLC: Steven Shope

UCF: Eduardo Salas, Clint Bowers

Sponsors: Air Force Office of Scientific Research, Office of Naval Research, NASA Ames Research Center, Army Research Laboratory

Page 4: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 33

OverviewOverview

What is Collaborative Cognition? A Focus on Measurement Assessing Collaborative

Performance & Cognition Toward Diagnosis of Collaborative

Performance Conclusions

Page 5: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 44

What is Collaborative

Cognition?

Page 6: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 55

Collaborative Cognition in Collaborative Cognition in PracticePractice

Page 7: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 66

Experimental ContextExperimental ContextCERTT (Cognitive Engineering Research on

Team Tasks) LabA Synthetic Task Environment for the Study of Collaborative

Cognition

Five Participant Consoles Experimenter Console

Page 8: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 77

Individual knowledgeTeam Process

Behaviors

Team Knowledge

Team Performance

Collaborative Cognition FrameworkCollaborative Cognition Framework

Holistic Level

Collective level

Taskwork & Teamwork Knowledge

Long-term

Fleeting

Page 9: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 88

Defining Collaborative Defining Collaborative CognitionCognition

It is more than the sum of the cognition of individual team members.

It emerges from the interplay of the individual cognition of each team member and team process behaviors

Page 10: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 99

A Focus on Measurement

Page 11: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1010

Why Measurement?Why Measurement?Assessment of collaborative performance

or effectiveness (criterion or dependent measures) often taken for granted.Outcome measures of collaborative

performance do not reveal why performance is effective or ineffective.Process measures of collaborative

behavior are often subjective and lack reliability and validity.

Page 12: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1111

Why Measurement?Why Measurement?(continued)(continued)

Collaborative cognition is assumed to contribute to collaborative performance, and especially for growing number of cognitive tasks.Understanding the team cognition behind

team performance should inform interventions (design, training, selection) to improve that performance.

Page 13: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1212

Measurement LimitationsMeasurement Limitations Measures tend to assume homogeneous groups Measures tend to target collective level Aggregation methods are limited Measures are needed that target the more dynamic

and fleeting knowledge Measures are needed that target different types of

long-term collaborative knowledge A broader range of knowledge elicitation methods is

needed A need for streamlined and embedded measures Newly developed measures require validation

Page 14: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1313

Assessing Collaborative Performance and Cognition

Page 15: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1414

The The ““Apples and OrangesApples and Oranges””ProblemProblem

Shared knowledge = similar knowledge

Accuracy is relative to single referent

Person A Person B

Referent

Measures to assess collaborative knowledge often assume knowledge homogeneity among group members.

Page 16: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1515

The Groups We Study The Groups We Study Consist of Consist of

““Apples and OrangesApples and Oranges””

Airport Incident Command Center Telemedicine

Page 17: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1616

““SharedShared”” KnowledgeKnowledge

Shared = Common and Complementary

Page 18: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1717

An Approach to the An Approach to the Apples Apples and Orangesand Oranges ProblemProblem

Measures of team knowledge with heterogeneous

accuracy metrics

Page 19: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1818

Experimental ContextExperimental Context

Five studies: Two different 3-person tasks: UAV (Uninhabited Air Vehicle) and Navy helicopter rescue-and-reliefProcedure: Training, several missions,

knowledge measurement sessionsManipulate: co-located vs. distributed

environments, training regime, knowledge sharing capabilities, workload

Page 20: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 1919

Experimental ContextExperimental Context

MEASURES Team performance: composite measure Team process: observer ratings and

critical incident checklist Other: Communication (flow and audio

records), video, computer events, leadership, demographic questions, working memory, situation awareness

Taskwork & Teamwork Knowledge

Page 21: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2020

Scores from completed missions for all teams

325

425

525

625

725

825

925

1 2 3 4 5 6 7 8 9 10

Mission

Team

Per

form

ance

Sco

re 907.96902.83898.21886.13812.09809.52802.37799.42790.93768.18660.31

Page 22: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2121

LongLong--term term TaskworkTaskworkKnowledgeKnowledge

Factual Tests

Psychological scaling

The camera settings are determined by a) altitude, b) airspeed, c) light conditions, d) all of the above.

How related is airspeed to restricted operating zone?

Page 23: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2222

LongLong--term Teamwork term Teamwork KnowledgeKnowledge

Given a specific task scenario, who passes what information to whom?

Teamwork Checklist___AVO gives airspeed info to PLO___DEMPC gives waypoint restrictions to AVO___PLO gives current position to AVO

AVO= Air Vehicle Operator

PLO = Payload Operator

DEMPC = Navigator

Page 24: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2323

Traditional Accuracy Traditional Accuracy MetricsMetrics

Team Referent

.50

50% ACCURACYTeam Member: Air Vehicle Operator

Page 25: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2424

Heterogeneous Accuracy Heterogeneous Accuracy MetricsMetrics

Team Referent

DEMPC Referent

PLO Referent

AVO Referent

.50 1.0.33

0

ACCURACY

Overall: 50%

Positional: 100%

Interpositional: 17%

Team Member: AVO

AVO= Air Vehicle OperatorPLO = Payload OperatorDEMPC = Navigator

Page 26: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2525

Results Across StudiesResults Across Studies

Taskwork knowledge is predictive of team performance

But…True for psychological scaling,

not factual tests Timing of knowledge test is

critical

Page 27: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2626

Knowledge Profiles of Two TasksKnowledge Profiles of Two Tasks

Knowledge profile characterizing effective teams depends on task (UAV vs. Navy)

Knowledge Profile

0+Interposit.accuracy

++Positionalaccuracy

0+Intrateam similarity

0+Overall accuracy

Distributed(Navy

helicopter)

Common(UAV)

Knowledge metric

Page 28: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2727

Knowledge Profiles of Two Knowledge Profiles of Two TasksTasks

UAV Task

Command-and-Control

Interdependent

Knowledge sharing

Navy Helicopter Task

Planning and Execution

Less interdependent

Face-to-Face

Common Complementary

Page 29: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2828

A CrossA Cross--Training Study in Training Study in RetrospectRetrospect

Examined effects cross-training vs. other training regimes on collaborative performance and cognition

Unlike previous studies, cross-training had no performance benefit

Cross-training, did increase interpositional taskworkand teamwork knowledge

Perhaps knowledge profile for that task (specialization) was at odds with cross-training

Demonstrates benefits of assessing collaborative cognition

Page 30: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 2929

HoweverHowever……The descriptive information associated

with cognitive assessment is not sufficiently diagnostic

Symptoms vs. Diagnoses?Symptoms vs. Diagnoses?

Expert chess players can remember many more meaningful chess positions than chess novices

Experienced fighter pilots and undergraduate pilot trainees organize flight maneuver concepts differently

Good UAV teams have interpositionalknowledge

Effective teams communicate less than ineffective ones

Page 31: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3030

Toward Diagnosis of CollaborativePerformance

Page 32: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3131

To Move From Assessment To Move From Assessment to Diagnosis to Diagnosis ……..

Inefficient communication and low teamwork knowledge poor team situation awareness

Poor positional taskwork knowledge and coordination failures faulty mental model

Need to connect clusters of symptoms to Need to connect clusters of symptoms to diagnosis of team dysfunction or excellencediagnosis of team dysfunction or excellence

For exampleFor example……

Page 33: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3232

To Move From Assessment To Move From Assessment to Diagnosis to Diagnosis ……..

Also, in operational environments diagnosis is valuable to the extent that it is…

Leading indicatorLeading indicator Resident pathogensResident pathogens NonNon--routine eventsroutine events

Conducted in real time with task Conducted in real time with task performance performance (i.e., task(i.e., task--embedded, embedded, automated measures)automated measures)

Or better yet Or better yet ……prior to task performance prior to task performance (based on performance precursors)(based on performance precursors)

Page 34: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3333

Communication as a Window Communication as a Window to Collaborative Cognitionto Collaborative Cognition

Observable; Think aloud “in the wild” Reflects collaborative cognition at the

holistic level; is collaborative cognition Embedded in the task But…labor intensive transcription, coding,

and interpretation Exploit its richness by automating

analyses

Page 35: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3434

Our Approach to Our Approach to Communication AnalysisCommunication Analysis

Communication Flow AnalysisContent Analysis Using Latent

Semantic Analysis

Page 36: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3535

Analyzing Flow: CERTT Lab Analyzing Flow: CERTT Lab ComLogComLog DataData

Team members use push-to-talk intercom buttons to communicate. At regular intervals speaker and listener identities are logged

Page 37: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3636

Analyzing Flow: Analyzing Flow: ProNetProNet----Procedural NetworksProcedural Networks

Nodes define events that occur in a sequence An Example from UAV study: 6 nodes: Abeg, Aend,

Pbeg, Pend, Dbeg, Dend ProNet: Find representative event sequences

Quantitative: Chain lengths-->PerformanceMission 2: R2 = .509, F(2, 8) = 4.144, p = .058Mission 3: R2 = .275, F(1, 9) = 3.415, p = .098Mission 5: R2 = .628, F(2, 8) = 5.074, p = .051

Page 38: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3737

Analyzing Flow: Analyzing Flow: ProNetProNet----Procedural NetworksProcedural NetworksQualitative: Communication patterns

predictive of performance

Abeg

AendPend

Dbeg

Dend

Pbeg

Team 2 before PLO-DEMPC’s fight

Abeg

Aend

Pbeg

Pend

Dbeg

Dend

Team 2 after PLO-DEMPC’s fight

AVO= Air Vehicle OperatorPLO = Payload OperatorDEMPC = Navigator

Page 39: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3838

Analyzing Flow: Other Analyzing Flow: Other ApproachesApproaches

Measure of speaker dominanceDeviations from ideal flowClustering model-based patterns

Page 40: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 3939

Content Analysis with Content Analysis with Latent Semantic Analysis (LSA)Latent Semantic Analysis (LSA)

A tool for measuring cognitive artifacts based on semantic information.

Provides measures of the semantic relatedness, quality, and quantity of information contained in discourse.

Automatic and fast. We can derive the meaning of words through

analyses of large corpora.

Landauer, Foltz, & Laham, 1998

Page 41: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4040

Content Analysis with Content Analysis with Latent Semantic Analysis (LSA)Latent Semantic Analysis (LSA)

(continued)(continued) Large constraint satisfaction problem of

estimating the meaning of many passages based on their contained words (like factor analysis)

Method represents units of text (words, sentences, discourse, essays) as vectors in a high dimensional semantic space based on correlations of usage across text contexts

Compute degree of semantic similarity between any two units of text

Page 42: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4141

Content Analysis with Content Analysis with Latent Semantic Analysis (LSA)Latent Semantic Analysis (LSA)

67 Transcripts from missions 1-7

XML tagged with speaker and listener information

~2700 minutes of spoken dialogue

20,545 separate utterances (turns)

232,000 words (660 k bytes of text)

Semantic Space: 22,802 documents

Utterances from dialogues

Training material

Interviews with domain experts Derived several statistical measures of the quality of

each transcript

An Example from UAV Study 1

Page 43: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4242

Content Analysis with Content Analysis with Latent Semantic Analysis (LSA)Latent Semantic Analysis (LSA)

Team 1 Mission 3Score: 750

Team 7 Mission 3Score 580

Team 3 Mission 4Score: 620

Team 5 Mission 4Score 460

Team 6 Mission 3Score 490

Team 8 Mission 3Score ????

Team 8 Mission 6Score 560

LSA-based communication score predicts performance (r =.79).

Page 44: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4343

Variance of scores of similar dialogues r=-.58

Vector length, r=-.35

Number of words, r=-.34

Zipf R2, r=-.47

Percent non-function words, r=.34

….

Five factor RMMR model: r=.76

Other Significant Variables

Conclusion: We can accurately predict team performance from dialogues as a whole.

Page 45: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4444

Analyzing Content: Analyzing Content: Other ApproachesOther Approaches

Automatic transcript codingCoherence in team dialogueMeasures of individual

contributions

Page 46: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4545

Implications of Communication Work

Collaborative cognition is revealed through discourse

Measurement of the team as a whole, as well as individuals

Techniques move toward automated analyses of the content of team dialogues

Avoids tedious hand coding, keeps high reliability

Automation will allow for task-embedded measures that assess and diagnose in real-time

Page 47: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4646

Conclusions

Page 48: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4747

SummarySummary

Understanding collaborative cognition is critical for diagnosis of team dysfunction or excellence and later interventionAssessment is only a first stepDiagnostic information is neededIn operational environments diagnosis

needs to be task-embedded, automatic, and forward-looking

Page 49: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4848

Implications & ApplicationsImplications & Applications

Suggestions for aiding collaborative cognition through training or technology.Selecting/composing teams for

optimal collaborationOn-line monitoring of collaborative

cognition in high-risk environments

Page 50: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 4949

Questions or Comments?Questions or Comments?

Page 51: Measuring Collaborative Measuring Collaborative Cognition

January 2003January 2003 CKM workshopCKM workshop 5050

ContactNancy J. Cooke

Arizona State University

[email protected]://psych.nmsu.edu/

CERTT/AZ

NM