Overview of "Large Models of What? Mistaking Engineering Achievements for Human Linguistic Agency"
The paper "Large Models of What? Mistaking Engineering Achievements for Human Linguistic Agency", authored by Abeba Birhane and Marek McGann, provides a critical evaluation of the linguistic capabilities attributed to LLMs. The authors argue that such capabilities are frequently overstated and often based on misguided assumptions about language and data completeness. From an enactive cognitive science standpoint, the paper asserts that language, as practiced by humans, cannot be comprehensively modeled by current LLM architectures due to their lack of embodiment, participation, and precarity.
Key Assumptions Challenged
Birhane and McGann identify two foundational but flawed assumptions underpinning the exaggerated claims about LLMs:
- Language Completeness: This assumption posits that natural language is a stable entity that can be entirely captured and modeled.
- Data Completeness: This assumption holds that sufficient data can comprehensively represent a language for effective modeling by LLMs.
Enactive Perspective on Language
The paper contrasts these assumptions with the enactive approach to cognitive science, which views language as dynamic, embodied, and participatory. According to this view, language is a living activity—termed 'languaging'—which cannot be encapsulated by static datasets. This framework highlights three key aspects of human linguistic interaction that are absent in LLMs:
- Embodiment: Language arises from our embodied experiences in the world.
- Participation: Linguistic acts are inherently social and collaborative.
- Precariousness: Human language use involves continual negotiation and management of context-specific tensions and risks.
Contrasting LLMs and Human Linguistic Agency
Embodiment
Human language is deeply tied to our bodily experiences and the contexts within which interactions occur. In contrast, LLMs operate by generating text based on statistical relationships in pre-existing data. They lack the physical and experiential grounding that provides depth and nuance to human communication. This disconnect is apparent in situations where context and shared experiences significantly shape the meaning of interaction.
Participation
Human linguistic interactions are collaborative and adaptive. People frequently seek clarification and adjust their communication based on immediate feedback, a coordination that LLMs are not inherently equipped to handle. The participatory nature of human language defies static classification and cannot be fully captured by text corpora used to train LLMs.
Precarity
Human linguistic actions are driven by the necessity to manage social and contextual tensions. The stakes in our linguistic interactions are intrinsically tied to our social identities and lived experiences. LLMs, however, lack any form of self-driven motivation or risk, as their output is devoid of intentionality, ethical considerations, and the subtleties of human social structures and power dynamics.
Algospeak: A Case Study in Human Linguistic Innovation
The phenomenon of 'algospeak'—where users modify language to bypass automated content moderation—is highlighted as a clear example of human linguistic agency. This dynamic adaptation underscores the embodied, participatory, and precarious nature of human language, all elements that current LLMs cannot replicate or understand. Algospeak emerges as a form of resistance and innovation, pointing to the creative and context-sensitive aspects of human communication that LLMs lack.
Practical and Theoretical Implications
The misconceptions about LLM capabilities have significant implications. Practically, they could misinform public policy and social practices, leading to inappropriate reliance on these tools in sensitive domains. Theoretically, they obscure the deeper understanding of what constitutes human linguistic activity. LLMs should be seen as advanced tools aiding human language users rather than as entities capable of true linguistic agency.
Future Directions
Future developments in AI should account for the nuanced, dynamic, and embodied nature of human language. Integrating insights from enactive cognitive science could guide more robust understandings of linguistic agency and inform responsible AI design and deployment. This multidisciplinary approach is essential to address ethical concerns and potential societal impacts, particularly regarding marginalized communities.
Conclusion
In summary, Birhane and McGann's paper emphasizes the need to recalibrate our understanding of LLMs in light of their inherent limitations. Recognizing the distinct nature of human linguistic agency can guide the development of more ethically and socially aware AI technologies. The paper serves as a crucial reminder that while LLMs represent significant engineering advancements, they do not mirror the complex, embodied, and participative nature of human language.