Each participant response captures, for one clip:
| selected_emotion | enum (9) | One of the locked 9-emotion taxonomy |
| secondary_emotions | multi-enum | Required when primary = mixed_more_than_one, ≥ 2 codes |
| selected_cues | multi-enum (9) | Which signals the reader used; multi-select |
| confidence_rating | 1–5 Likert | Self-reported confidence in the reading |
| free_text_reasoning | text ≤ 280 chars | Optional first-person reasoning |
Each session captures, per participant per wave, immutable demographic snapshots: age, country (ISO-2), primary spoken language, English confidence (1–5), gender, cultural background (categorical: East Asian, SE Asian, South Asian, White European, Black African, Hispanic/Latino, MENA, Mixed, Other), and self-rated emotion-reading difficulty (1–4).
Each clip carries curator-assigned metadata — social_complexity (simple/moderate/complex), verbal_dependency (low/medium/high), and a target person description identifying whose emotion is being read.
Each clip optionally has one approved AI annotation (Claude Sonnet 4.6 in current configuration) with the AI's primary emotion, secondary emotions, cue selections, public-facing rationale, and ambiguity caveat.
9-emotion taxonomy
- happy_amused
- sad_disappointed
- angry_frustrated
- anxious_nervous
- embarrassed_awkward
- surprised
- confused_uncertain
- neutral_hard_to_tell
- mixed_more_than_one
9-cue taxonomy
- facial_expression
- tone_of_voice
- verbal_content
- body_language
- situation_context
- timing_pacing
- others_reaction
- something_else
- not_sure