Alignment between LLM perplexity and human neurophysiological responses to confusing code
Determine whether large language models assign higher perplexity to code snippets that contain atoms of confusion compared to clean, functionally equivalent snippets, and ascertain whether token-level perplexity from large language models aligns with human neurophysiological responses during program comprehension.
References
While atoms of confusion are known to induce confusion, and perplexity has been linked to comprehension difficulty in isolated studies, it is unclear whether LLMs assign higher perplexity to these known confusing constructs or whether LLM perplexity aligns with human neurophysiological responses during program comprehension.
                — How do Humans and LLMs Process Confusing Code?
                
                (2508.18547 - Abdelsalam et al., 25 Aug 2025) in Section 2.4, Research Gap