Assessing novel C2 socio-technical AI enabled planning concepts: �What we learnt using the Wizard of Oz and other evaluation methods��Whether to WoZ or not to WoZ
Simon Attfield, Andrew Leggatt, Paddy Turner, Holly Roberts,
Rachel Asquith & Richard Ellis
Related Machine-Speed C2 briefs today
2
Outline of paper
3
Introduction
4
Needs
Opportunities
Innovation
Evaluation enables…
5
Key concepts
6
User-centred evaluation methods (a taxonomy)
Expert Review
Heuristic Evaluation
Experts review system or interface against established principles or heuristics (e.g. Nielsen, 1992).
Cognitive Walkthrough
Step-by-step task simulation with predefined questions. Evaluates ‘walk-up-and-use’ learnability (Wharton, C., Rieman, J., Lewis, C., & Polson, P., 1994)
Feature Inspection
Examination of specific features rather than the overall user interface or experience. Determines how features support user-tasks, integration and usability.
7
Lab-based methods
Pluralistic Walkthrough (participatory design review, storyboarding)
Facilitated walkthrough with stakeholders Based on a typical task flow.
User Testing (usability testing)
Users perform tasks using system. Interaction observed and recorded.
Post-task questionnaire or interview.
Wizard of Oz
Wizard of Oz is a user testing method proposed for AI because they are difficult to prototype. Responses created by human (Wizard).
Field based methods
Naturalistic Observation
Observing participants in natural environment without interference by researcher.
Interviews
Insights into people’s thoughts, experiences, perceptions, and motivations (not observable).
Contextual Interview
Researcher in natural working environment, observe and engage directly with participants as they perform their routine tasks.
Evaluation of two military planning concepts using �Wizard of OZ method.
8
Auto-piggery
Helps planners understanding of stakeholders within an operational environment.
Hey Socrates
Monitors ongoing planning process, and offers questions and prompts to highlight possible omissions or deviations from ‘good’ planning practice.
Auto-piggery
“Jane is a J5 planner and Secretary of the Joint Task Force Mission Analysis Group. The Joint Task Force Commander (JTFC) has been given a mission to deploy a force into <country x> as part of a coalition acting under an international mandate. The JTFC has a specified task to secure the city of <city y> and surrounding region and subsequently secure the southern flank. Internally, there are a number of threats and the JTFC needs to consider who the main players are to set up the Mission Analysis discussion.”
9
Auto-piggery wireframe
10
An interactive design representation for exploring the problem.
Brings to concept solution to life.
Each page represented an information display focusing on a given stakeholder group.
Interactivity simulated using linked objects to force page transitions.
Auto-piggery mock-up
11
Auto Piggery evaluation (Wizard of Oz)
12
Hey Socrates
13
Hey Socrates evaluation (Wizard of Oz)
14
Guidelines
Data Recording and Analysis
15
Discussion
16
| Auto Piggery | Hey Socrates |
Interaction modality | linguistic and visual (agent plus artefact) | linguistic (agent only) |
Anthropomorphic? | less so | yes |
Naturalistic interaction? | less so | yes |
Critical path? | critical path | optional |
Materials preparation | mock-ups with canned content | guidelines |
Preparation | high | medium |
Prior design commitment | high (evaluating more) | low |
Interaction constraints | high | low |
User experience | rigid (potential source of frustration) | flexible |
Recommendation | prior lower-fidelity evaluation strategy. | suitable for low-maturity evaluation. |
Conclusions
17