Multimodal Interactive Deception Assessment
- MIDA is a benchmark framework for detecting deception in complex, interactive social settings using synchronized multimodal data.
- It employs advanced fusion strategies and epistemic reasoning to integrate audio, video, and textual cues, addressing challenges like neutral bias.
- Empirical results demonstrate improved accuracy and explainability, paving the way for innovative real-time deception analysis.
Multimodal Interactive Deception Assessment (MIDA) is a benchmark framework and methodology for automatic detection and analysis of deceptive behaviors in complex, interactive social settings using synchronized multimodal data. MIDA tasks the model with inferring the veracity of utterances within dynamic, multi-party or dyadic conversations by leveraging coordinated video, audio, and textual streams, and with modeling the social-epistemic context required for effective “reading of the room” (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025, Miah et al., 11 Jun 2025, Rugolon et al., 26 Jun 2025).
1. Formal Problem Definition
MIDA operationalizes deception assessment as follows. Let be a set of multi-party conversational videos (e.g., from social deduction games), and be the corresponding dialogue transcripts. Each transcript is decomposed into a sequence of utterances . The ground-truth veracity label is captured by a function
where each utterance is labeled as factually correct, a lie, or non-verifiable. The model’s objective is, for each , to predict
where is the preceding utterance history, and compare against (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025). MIDA further incorporates, in dyadic or group settings, joint feature extraction from all participants, potentially fusing sender and receiver signals, and can be implemented in both multi-turn and real-time interactive regimes (Rugolon et al., 26 Jun 2025).
2. Datasets and Annotation Protocols
The core MIDA datasets reflect ecologically valid deception contexts:
- MIDA-Ego4D: 40 game-based sessions, 5–8 players, 819 utterances/session, comprehensively annotated (Kang et al., 20 Nov 2025).
- MIDA-YouTube: 151 publicly uploaded game videos, 1,541 labeled utterances (Kang et al., 20 Nov 2025).
- MU3D: 1,200 face-to-face interviews, video/audio/pose annotations, 5 crowd labels per clip (Cohen’s ) (Miah et al., 11 Jun 2025).
- RLTD: 600 courtroom clips with expert-verified binary deception labels (Miah et al., 11 Jun 2025).
- Dyadic Cohorts (e.g., Swedish): 22 dyads, 44 participants, sender/receiver recording, with post-hoc trustworthiness/interaction ratings (Rugolon et al., 26 Jun 2025).
Annotation steps (for high-stakes games):
- Manual extraction of latent game state (who did what), enabling inference of available private knowledge.
- LLM pipelines (Gemini-2.5-Pro) generate preliminary veracity labels using game rules and dialogue context.
- Human validation audits yield 95% agreement with LLM for ground-truthing (Kang et al., 20 Nov 2025).
All video and transcript data are tightly synchronized via forced alignment. Facial crops and body bounding boxes are extracted (MTCNN, OpenPose). Audio is segmented per utterance, with low-level spectral/prosodic features (pitch, energy, formants) obtained (Parselmouth, Librosa, GeMAPS) (Kang et al., 20 Nov 2025, Rugolon et al., 26 Jun 2025).
3. Model Architectures and Fusion Strategies
MIDA benchmarks span a broad spectrum of architectures:
- Multimodal LLMs (MLLMs): e.g., GPT-4o, Gemini-2.5-pro, Llama-3-8B, InternVL3.5-8B (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
- Vision-LLMs (LMMs): CLIP, BLIP-2; fine-tuned on frame-text and full game interactions (Miah et al., 11 Jun 2025).
- Hybrid Mixture-of-Experts: Modality Interactive Mixture-of-Experts (MIMoE) for explicit gating based on unimodal agreement and semantic alignment, dynamically activating expert fusion blocks (Liu et al., 21 Jan 2025).
- Late Fusion: Per-modality classifiers (audio, video, text) feeding a meta-classifier (decision tree) (Rugolon et al., 26 Jun 2025).
- Early Fusion: Frame-aligned concatenation of audio/video features (Rugolon et al., 26 Jun 2025).
- Interactive Components: Modality routers adapt channel weighting by real-time signal quality; user-feedback loops enable continuous adaptation via LoRA or online fine-tuning (Miah et al., 11 Jun 2025).
SoCoT (Social Chain-of-Thought) pipeline decomposes inference into:
- Low-level perception: extraction of face, body, and voice primitives ().
- High-level social inference: theory-of-mind simulation .
- Decision and rationale: aggregation and explanation .
Dynamic Social Epistemic Memory (DSEM) module maintains, for each player, a persistent board storing observed/felt/known states, updated each turn via multimodal signals (Kang et al., 20 Nov 2025):
4. Performance Benchmarks and Metrics
MIDA utilizes strict classification metrics reflecting the multi-class nature (TRUE, FALSE, NEUTRAL) and strong class imbalance:
Let , , be the per-class (c) true positives, false positives, and false negatives.
- Precision:
- Recall:
- F1:
Macro-averaged:
Overall accuracy:
Binary accuracy (restricting to TRUE/FALSE in the denominator) is also commonly reported (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
Empirical highlights (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025, Rugolon et al., 26 Jun 2025):
- GPT-4o achieves Acc , Macro-F1 on Ego4D.
- GPT-4o-mini achieves best Binary Acc (39.4%) and F1 (36.1%) on Ego4D-MIDA.
- Late fusion in dyadic settings combining both modalities/participants yields (Rugolon et al., 26 Jun 2025).
- Open-source models lag behind: Qwen2.5-VL Macro-F1 ; DeepSeek-R1-8B Macro-F1 (Kang et al., 20 Nov 2025).
5. Failure Analyses and Cognitive Bottlenecks
MIDA experiments consistently expose core limitations in current multimodal models:
- Conservative NEUTRAL Bias: Models overpredict NEUTRAL (85% F1) at the expense of FALSE, due to risk-averse alignment and class imbalance in training distributions (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
- Lack of Theory-of-Mind: Models lack explicit epistemic modeling of what each participant knows or believes, critical for distinguishing deliberate lies from honest mistakes or unverifiable statements (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
- Weak Multimodal Grounding: Increasing the number of visual frames, or ablating text/temporal context, does not confer improvement; models generally fail to exploit facial or gestural information reliably—visual descriptions often do not impact final deception verdicts (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
- Limited Integration of Receiver Cues (Dyadic): Inclusion of synchrony and receiver features substantially improves accuracy, highlighting the inadequacy of sender-only approaches (Rugolon et al., 26 Jun 2025).
6. Enhanced Reasoning Pipelines and Theoretical Extensions
To address these bottlenecks, MIDA research proposes:
- Social Chain-of-Thought (SoCoT): Decomposition of multimodal reasoning into symbolic perception, high-level epistemic inference, and explanatory decision-making. SoCoT yields measurable gains over direct MLLM prompting (Acc +8%, Macro-F1 +2.6 points) (Kang et al., 20 Nov 2025).
- Dynamic Social Epistemic Memory (DSEM): Persistent, player-specific state tracking for belief modeling; shown to improve Macro-F1 and binary accuracy by 1.6–3.3 points in open-source and closed-source models (Kang et al., 20 Nov 2025).
- Mixture-of-Experts Fusion: Gating ensembles based on measured modality agreement/alignment route examples to specialized fusion experts, effectively reconciling concordant and discordant modality signals (Liu et al., 21 Jan 2025).
- Chain-of-Thought Prompting: Step-by-step inference over linguistic, prosodic, and gestural evidence, shown to increase explainability (but not always faithfulness) of verdicts (Miah et al., 11 Jun 2025).
- Dyadic Synchrony and Interpersonal Features: Quantifying audio-visual synchrony (e.g., over feature windows), decision-tree meta-fusion, and inclusion of both sender and receiver modalities for robust deception cue extraction (Rugolon et al., 26 Jun 2025).
7. Open Challenges and Future Directions
Key directions for advancing MIDA’s efficacy include:
- Innate Theory-of-Mind Integration: Moving beyond post-hoc prompts to architectures with built-in ToM modules or graph-structured epistemic reasoning (Kang et al., 20 Nov 2025, Kang et al., 31 Oct 2025).
- Context-Adaptive Alignment: Allowing calibrated, context-sensitive risk-taking in model outputs, enabling more accurate judgments under uncertainty (Kang et al., 31 Oct 2025).
- Robust Multimodal Embedding: Developing representation learning that filters noise and isolates salient behavioral cues for deception inference (Kang et al., 20 Nov 2025).
- Cross-Domain, Cross-Cultural Transfer: Addressing domain shift and cultural variation in gesture/speech cues; supporting multilingual, multicultural adaptation with Bayesian/uncertainty-aware fusion (Miah et al., 11 Jun 2025).
- Human-in-the-Loop Adaptation: Leveraging active learning, user feedback, and online fine-tuning for continual improvement (Miah et al., 11 Jun 2025).
- Deployment: Efficient streaming architectures, on-device model distillation, and interface dashboards presenting verdicts with explanatory rationales and modifiable feedback (Miah et al., 11 Jun 2025).
A plausible implication is that genuine “reading the room” for deception in open domains will require models capable of real-time, contextually grounded, and cognitively inspired multimodal social reasoning, integrating both sender and receiver, and robust to shifts in domain, channel, and social context.
References:
- (Kang et al., 20 Nov 2025)
- (Kang et al., 31 Oct 2025)
- (Miah et al., 11 Jun 2025)
- (Rugolon et al., 26 Jun 2025)
- (Liu et al., 21 Jan 2025)
- (Kopev et al., 2019)