Papers
Topics
Authors
Recent
2000 character limit reached

More AI Assistance Reduces Cognitive Engagement: Examining the AI Assistance Dilemma in AI-Supported Note-Taking (2509.03392v1)

Published 3 Sep 2025 in cs.HC

Abstract: As AI tools become increasingly embedded in cognitively demanding tasks such as note-taking, questions remain about whether they enhance or undermine cognitive engagement. This paper examines the "AI Assistance Dilemma" in note-taking, investigating how varying levels of AI support affect user engagement and comprehension. In a within-subject experiment, we asked participants (N=30) to take notes during lecture videos under three conditions: Automated AI (high assistance with structured notes), Intermediate AI (moderate assistance with real-time summary, and Minimal AI (low assistance with transcript). Results reveal that Intermediate AI yields the highest post-test scores and Automated AI the lowest. Participants, however, preferred the automated setup due to its perceived ease of use and lower cognitive effort, suggesting a discrepancy between preferred convenience and cognitive benefits. Our study provides insights into designing AI assistance that preserves cognitive engagement, offering implications for designing moderate AI support in cognitive tasks.

Summary

  • The paper demonstrates that intermediate AI assistance yields the highest post-test scores and optimal cognitive load compared to other levels.
  • It employs mixed-effects linear regression and repeated-measures ANOVA to show that excessive automation reduces active engagement in note-taking.
  • The findings imply that adaptive, scaffolded AI support can balance structure and user control to enhance learning outcomes.

More AI Assistance Reduces Cognitive Engagement: The AI Assistance Dilemma in Note-Taking

Introduction and Motivation

This paper investigates the "AI Assistance Dilemma" in the context of note-taking, specifically examining how varying levels of AI support impact cognitive engagement and comprehension during real-time lecture consumption. The study is motivated by the increasing integration of AI tools in cognitively demanding tasks and the open question of whether such assistance enhances or undermines meaningful engagement. Drawing on the encoding-storage paradigm and cognitive load theory, the authors hypothesize that excessive automation may reduce the cognitive effort required for active learning, potentially leading to suboptimal outcomes.

System Design and Experimental Setup

The authors developed NoteCopilot, a custom AI-assisted note-taking system with three distinct levels of AI support:

  • Automated AI (high assistance): Provides structured notes every 1–2 minutes, resembling commercial auto-note services.
  • Intermediate AI (moderate assistance): Delivers real-time summary blocks after each speaking turn.
  • Transcript AI (low assistance): Offers real-time transcript blocks with minimal abstraction.

All variants feature a real-time AI-generated note panel, a search and synthesis panel, and a rich text editor for manual note-taking and integration of AI-generated content. Figure 1

Figure 1: System overview showing the real-time AI-generated note panel, search and synthesis panel, and text editor for user interaction across all assistance levels.

A within-subject experiment was conducted with 30 participants, each experiencing all three conditions in counterbalanced order. Participants watched lecture videos, took notes using the system, completed post-tests (both closed-book and open-note revision), and participated in surveys and interviews. Figure 2

Figure 2: Study design flow, detailing the sequence of introduction, note-taking, testing, revision, and interviews for each AI assistance condition.

Granularity and Abstraction of AI Notes

The three assistance levels differ in the granularity and abstraction of AI-generated content:

  • Transcript AI: Presents raw, unfiltered transcript blocks.
  • Intermediate AI: Offers concise summaries for each transcript chunk, facilitating comprehension.
  • Automated AI: Consolidates summaries into structured notes with definitions and advantages, providing detailed, organized information. Figure 3

    Figure 3: Comparison of AI note granularity across conditions, illustrating increasing abstraction and organization from transcript to structured notes.

Quantitative Findings

Learning Outcomes

Mixed-effects linear regression and post-hoc Tukey HSD analyses revealed:

  • Intermediate AI yielded the highest post-test and revision scores.
  • Automated AI resulted in the lowest scores, despite being the most preferred for ease of use.
  • Transcript AI performed better than Automated AI but worse than Intermediate AI. Figure 4

    Figure 4: Post-test and revision scores across conditions, showing significant superiority of Intermediate AI over Automated and Transcript AI.

Cognitive Load

Repeated-measures ANOVA showed:

  • Extraneous cognitive load was lowest in Automated AI, but germane cognitive load (beneficial for learning) was highest in Intermediate AI.
  • Manual note-taking (Transcript AI) imposed the highest extraneous load.
  • Reading AI-generated notes in Automated AI required more effort than in other conditions.

Note-Taking Behaviors

  • Automated AI led to the highest total note count, but with minimal user modification.
  • Intermediate AI encouraged more active engagement, with users frequently modifying and reorganizing AI-generated summaries.
  • Transcript AI users predominantly relied on manual note-taking.

Qualitative Insights

Interviews revealed nuanced perceptions:

  • Intermediate AI was valued for fostering active engagement and providing digestible, synchronous information.
  • Automated AI was appreciated for its structure and completeness but led to disengagement and over-reliance.
  • Transcript AI, while cognitively demanding, was trusted for its fidelity to the original content.
  • Participants preferred maintaining control over their notes, using AI as augmentation rather than replacement.
  • Users desired more contextual and intent-aware AI assistance, adaptable to their evolving needs and learning goals.

Theoretical and Practical Implications

The Assistance Dilemma

The findings empirically validate the assistance dilemma: excessive automation reduces cognitive engagement and comprehension, while moderate, scaffolded AI support optimizes learning outcomes. This aligns with cognitive load theory and the encoding-storage paradigm, suggesting that AI should offload mechanical effort but preserve opportunities for active reasoning and reflection.

Design of Intermediate AI

Intermediate AI, characterized by real-time, granular summaries and user agency in note composition, emerges as an effective paradigm for supporting cognitive tasks. The system's middleware layer enables users to evaluate, modify, and integrate AI-generated content, promoting desirable difficulty and scaffolding metacognitive processes. Figure 5

Figure 5: Interfaces for Intermediate and Transcript AI, highlighting user interaction with AI-generated summaries and transcript blocks.

Adaptivity and Personalization

Future AI systems should dynamically adjust content granularity, abstraction, and interaction modalities based on user intent, topic complexity, and behavioral signals. This requires moving beyond static scaffolding to engagement-sensitive, goal-oriented adaptation.

Limitations

  • The study's sample size and short video duration limit generalizability.
  • Note quality was not assessed; only quantity and modification behaviors were measured.
  • Dynamic editing processes during note-taking were not fully captured.

Conclusion

This work demonstrates that more AI assistance does not necessarily yield better cognitive outcomes in note-taking. Moderate, scaffolded AI support (Intermediate AI) maximizes comprehension and engagement, while excessive automation (Automated AI) risks disengagement and over-reliance. The results highlight the need for careful calibration of AI support in cognitive tasks, advocating for adaptive, intermediate AI paradigms that preserve human agency and foster meaningful learning.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 1 like about this paper.