Papers
Topics
Authors
Recent
2000 character limit reached

CPADP: Adaptive Dropout Prediction for Online Courses

Updated 18 November 2025
  • CPADP is a framework that adaptively predicts dropout in LLM-driven online courses by analyzing chapter-level student interactions.
  • It employs a three-stage prediction architecture (Zero-Shot, Few-Shot, Fine-Tuned) to optimize performance as more progress data accrues.
  • Personalized interventions, such as LLM-generated recall emails, effectively re-engage at-risk learners and boost course completion rates.

Course-Progress-Adaptive Dropout Prediction (CPADP) is a predictive and intervention framework devised for LLM-driven interactive online courses, particularly within the context of Massive AI-empowered Courses (MAIC). Leveraging multi-agent LLM platforms to create highly dynamic, text-intensive educational environments, CPADP predicts and reduces learner dropout by adaptively analyzing students’ interaction histories and intervening with personalized, content-aware recalls. The framework achieves high predictive accuracy using chapter-level, progress-gated predictors and downstream re-engagement agents that integrate deeply with the course’s dialogic structure (Wang et al., 24 Aug 2025).

1. Formalization of Dropout and Student Interaction

CPADP operates on granular, chapter-level representations of learner progression through a K-chapter MAIC. For student ii:

  • C={1,...,K}C = \{1, ..., K\} is the set of course chapters.
  • SinCS^\mathrm{n}_i \subseteq C are chapters completed by ii (i.e., the instructor agent has presented all slides).
  • Pi=Sin/KP_i = |S^\mathrm{n}_i| / K is normalized course progress.
  • Dropout yi{0,1}y_i \in \{0,1\} is defined as yi=1y_i = 1 if Sin<K|S^\mathrm{n}_i| < K (failure to complete all chapters); yi=0y_i = 0 otherwise.
  • Ii,h=((tj,j))j=1Ni,hI_{i,h} = ((t_j, \ell_j))_{j=1}^{N_{i,h}} is the ordered sequence of Ni,hN_{i,h} interaction messages (j)(\ell_j) and timestamps (tj)(t_j) up to the start of chapter ChC_h.

A single student’s record may yield multiple (Ch,Cp)(C_h, C_p) training instances: ChC_h demarcates the start of the interaction history window, CpC_p the end of the dropout prediction window (1ChCpK1 \leq C_h \leq C_p \leq K), enabling temporally localized predictions as learners progress.

2. Feature Engineering and Representation

CPADP transforms interaction logs into structured features grouped into:

  • Textual Interaction Metrics
    • fi,hfreq=Ni,h/(Ch1)f_{i, h}^\mathrm{freq} = N_{i,h} / (C_h-1): average messages per completed chapter.
    • fi,hlen=(1/Ni,h)jjf_{i, h}^\mathrm{len} = (1/N_{i,h}) \sum_j |\ell_j|: average token-length per message.
    • fi,hlex=TFIDF(1...Ni,h)RDf_{i, h}^\mathrm{lex} = \mathrm{TFIDF}(\ell_1 ... \ell_{N_{i,h}}) \in \mathbb{R}^D: TF–IDF vectorized content embedding.
  • Time-Based Features
    • fi,hidlef_{i, h}^\mathrm{idle}: mean inter-message interval within Ii,hI_{i,h} or time elapsed since the last message at chapter start.
  • Progress Indicators
    • fi,hprog=Pi(Ch)=Sin(Ch)/Kf_{i, h}^\mathrm{prog} = P_i(C_h) = |S^\mathrm{n}_i(C_h)| / K: normalized chapters finished before ChC_h.

These are concatenated into Xi,h=[fi,hfreq;fi,hlen;fi,hlex;fi,hidle;fi,hprog]R2+D+1+1X_{i,h} = [f_{i,h}^\mathrm{freq}; f_{i,h}^\mathrm{len}; f_{i,h}^\mathrm{lex}; f_{i,h}^\mathrm{idle}; f_{i,h}^\mathrm{prog}] \in \mathbb{R}^{2+D+1+1}. For variants using pretrained LLM (PLM) encodings, fi,hlexf_{i, h}^\mathrm{lex} is replaced by hi=PLMenc(1,...,Ni,h)h_i = \mathrm{PLM}_\mathrm{enc}(\ell_1, ..., \ell_{N_{i,h}}), yielding Xi,h=[fi,hfreq;fi,hlen;fi,hidle;fi,hprog;hi]R4+dX_{i,h} = [f_{i,h}^\mathrm{freq}; f_{i,h}^\mathrm{len}; f_{i,h}^\mathrm{idle}; f_{i,h}^\mathrm{prog}; h_i] \in \mathbb{R}^{4+d}.

CPADP’s architecture does not leverage demographics or self-reported learner traits, as empirical evidence indicates interaction features are more predictive of dropout (Wang et al., 24 Aug 2025).

3. Three-Stage Progress-Adaptive Prediction Architecture

Prediction in CPADP is modulated by a gating function based on course progress Pi(Ch)P_i(C_h), partitioned into intervals I1,...,IKI_1, ..., I_K:

Model Stage Input Regime Description
Zero-Shot (ZS) Ch=1C_h = 1 LLM-based prediction with unprimed prompt
Few-Shot (FS) 1<Ch<τ1 < C_h < \tau LLM-based prediction with prompt including a small, curated set of labeled cases; τ\tau is the chapter where \sim100 labels accrue
Fine-Tuned (FT) ChτC_h \geq \tau PLM encoder with MLP, trained on accumulated labels

Formally, for a given Xi,hX_{i,h}, prediction is:

y^i,hp=g(Pi(Ch))(Xi,h)\hat{y}_{i, h \to p} = g(P_i(C_h))(X_{i,h})

where gg selects gZSg_\mathrm{ZS}, gFSg_\mathrm{FS}, or gFTg_\mathrm{FT} based on ChC_h.

Within FT, the MLP classifier operates as: hi=PLMenc(Ii,h);zi=W2ReLU(W1hi+b1)+b2;[pi,1pi]=Softmax(zi)h_i = \mathrm{PLM}_\mathrm{enc}(I_{i, h}); \quad z_i = W_2 \cdot \mathrm{ReLU}(W_1 h_i + b_1) + b_2; \quad [p_i, 1-p_i] = \mathrm{Softmax}(z_i) or, in the simplest logistic form,

y^i=pi=σ(wThi+b)\hat{y}_i = p_i = \sigma(w^T h_i + b)

For ZS/FS, y^i=gLLM(prompt(Ii,h,examples))\hat{y}_i = g_\mathrm{LLM}(\mathrm{prompt}(I_{i,h}, \text{examples})).

A plausible implication is that the staged approach balances label scarcity at earlier course stages with model specificity at later ones.

4. Training Objective, Optimization, and Experimental Results

The FT (PLM+MLP) stage employs a weighted cross-entropy objective: L(θ)=1Ni,h,p[w+yi,h,plogy^i,h,p+w(1yi,h,p)log(1y^i,h,p)]L(\theta) = -\frac{1}{N} \sum_{i, h, p} [ w^+\, y_{i, h, p} \log \hat{y}_{i, h, p} + w^- (1 - y_{i, h, p}) \log(1-\hat{y}_{i,h,p}) ] where class weights w+w^+ and ww^- address a \sim40% dropout imbalance.

Key experimental details:

  • Dataset: 186 students, 1,201 labeled (Ch,Cp)(C_h, C_p) instances.
  • Split: 80% train (961 instances), 20% test (240 instances); cross-validation inside training.
  • PLM+MLP training: learning rate 2×1052 \times 10^{-5}, batch size 16, 3–5 epochs.
  • No explicit validation set; hyperparameters determined by cross-validation on train.
  • Metrics: Precision, Recall, F1, Accuracy.

Performance summary:

Model Precision Recall F1 Accuracy
PLM+MLP (FT) 0.966 0.906 0.935 0.954
GPT-4 (FS) 0.779
GPT-4 (ZS) 0.716

95.4% held-out accuracy confirms substantial improvement over zero-shot/few-shot baselines (Wang et al., 24 Aug 2025).

5. Personalized Recall Agent for Intervention

A downstream intervention mechanism, the Email-Agent, is invoked upon high-risk prediction. Built on an LLM (e.g., GPT-4), the recall agent:

  • Incorporates the student’s name, last completed topic/slide, and relevant interaction snippets
  • Adopts a motivational, supportive tone with previews of upcoming course content

Emails are dispatched mid-semester. Empirical post-intervention analysis shows that of 17 post-email logins, 8 students (“recalled” by the agent) had been offline for an average of ~52 days and were at ~0.75 progress, while the remaining 9 (self-initiated logins) had been offline ~7.7 days and reached 2.56 progress on average. This suggests the recall agent efficiently targets dormant, at-risk learners rather than active students.

6. Empirical Insights, Generalization, and Future Directions

Empirical analysis demonstrates that:

  • Interaction activity and text-derived embeddings are more predictive for dropout than demographic or psychographic data.
  • A staged, progress-adaptive approach—ZS, FS, then FT—yields optimal cost–accuracy balance in practical deployments.
  • PLM-based embeddings systematically outperform sparse TF–IDF features once sufficient label data accrues.
  • Content-personalized, LLM-generated recalls can drive re-engagement among students most likely to drop out.

Although validated in a single MAIC course, the CPADP pipeline is described as modular and transferable to any interactive, conversation-logging online course. Future research directions outlined include broadening multi-course validation, adding richer temporal/social signal extraction, and rigorously A/B testing interventions across modalities (emails, in-platform prompts, SMS).

7. Significance within LLM-Enhanced Online Education

CPADP stands as a comprehensive, LLM-native framework integrating granular interaction analysis, dynamically staged predictive modeling, and automated, just-in-time personalized interventions. Its canonical formulation and empirical results set a functional template for dropout risk management in next-generation LLM-augmented education platforms, providing a technically robust and extensible methodology for both prediction and active retention management (Wang et al., 24 Aug 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Course-Progress-Adaptive Dropout Prediction (CPADP).