Papers
Topics
Authors
Recent
2000 character limit reached

On the Compatibility of Generative AI and Generative Linguistics (2411.10533v2)

Published 15 Nov 2024 in cs.CL

Abstract: In mid-20th century, the linguist Noam Chomsky established generative linguistics, and made significant contributions to linguistics, computer science, and cognitive science by developing the computational and philosophical foundations for a theory that defined language as a formal system, instantiated in human minds or artificial machines. These developments in turn ushered a wave of research on symbolic AI. More recently, a new wave of non-symbolic AI has emerged with neural LLMs (LMs) that exhibit impressive linguistic performance, leading many to question the older approach and wonder about the the compatibility of generative AI and generative linguistics. In this paper, we argue that generative AI is compatible with generative linguistics and reinforces its basic tenets in at least three ways. First, we argue that LMs are formal generative models as intended originally in Chomsky's work on formal language theory. Second, LMs can help develop a program for discovery procedures as defined by Chomsky's "Syntactic Structures". Third, LMs can be a major asset for Chomsky's minimalist approach to Universal Grammar and language acquisition. In turn, generative linguistics can provide the foundation for evaluating and improving LMs as well as other generative computational models of language.

Summary

  • The paper demonstrates that current AI models exhibit strong observational adequacy but struggle with descriptive and explanatory linguistic clarity.
  • It compares traditional neural language models with emerging neural grammar induction models that explicitly derive syntactic rules.
  • The study suggests integrating neural grammar induction techniques to refine language learnability and advance the understanding of Universal Grammar.

Revisiting Chomskyan Theoretical Frameworks in Light of AI Linguistic Models

The paper "On the goals of linguistic theory: Revisiting Chomskyan theories in the era of AI" by Eva Portelance and Masoud Jasbi revisits foundational principles of generative linguistics within the context of advancements in AI LLMs. By examining these principles, the paper aims to evaluate whether AI models can effectively contribute to the goals of theoretical linguistics, a field long driven by the tenets laid out by Noam Chomsky. It specifically focuses on three critical areas: levels of theoretical adequacy, procedures for linguistic theory development, and language learnability and Universal Grammar.

Generative linguistics aims to describe human linguistic capability through formal grammars and explore how such grammars are acquired and processed. Chomsky's framework postulates several levels of adequacy for theoretical models, namely observational, descriptive, and explanatory adequacy. Observational adequacy pertains to generating grammatical sentences from a corpus, descriptive adequacy extends to explaining structural intuitions among native speakers, and explanatory adequacy addresses why a particular grammar is acquired over others. The paper draws parallels between these levels and Marr's computational, algorithmic, and implementational levels of analysis.

The authors argue that neural LLMs such as BERT and GPT-2 exhibit significant observational adequacy. However, these models often lack the capacity to provide a clear, human-readable grammatical structure, thus lagging behind in both descriptive and explanatory adequacy. Despite their robust performance in language production, they are frequently seen as "black boxes," where the underlying linguistic representations remain opaque.

In contrast, neural grammar induction models are presented as a promising alternative. These models do not merely replicate surface linguistic patterns but aim to infer explicit sets of syntactic rules akin to the grammars outlined in generative linguistics. Contemporary models, such as C-PCFG and L-PCFG, have demonstrated capabilities in deriving formal grammars with minimal prior linguistic assumptions. This positions them favorably as they can aspire to both descriptive and explanatory adequacy, making them valuable assets for revisiting the procedures for linguistic theory development.

Portelance and Jasbi also revisit the question of language learnability and Universal Grammar (UG) in light of AI models. Chomsky's UG hypothesis suggests that an innate linguistic structure constrains the space of possible grammars, a direct response to the learnability challenge. The paper posits that AI models, particularly those that implement probabilistic and neural network approaches, could provide insights into alternative frameworks, potentially refuting or refining long-standing theories like UG. These models present a learning paradigm that diverges from the traditional symbolic representation of grammatical knowledge, offering a more distributed and less biased approach to modeling language acquisition.

Despite the comprehensive evaluation of AI models in theoretical linguistics, the paper emphasizes the need for cautious integration. Neural network models, while robust, must be judged against criteria for model adequacy pertinent to linguistic theories before displacing or redefining existing frameworks. The authors argue for a balanced exploration of both neural LLMs and neural grammar induction models, particularly where they complement the theoretical goals of generative linguistics.

In conclusion, Portelance and Jasbi provide a nuanced examination of the synergies and divergences between AI models and traditional linguistic theories. They advocate for the potential of neural grammar induction models in evaluating and possibly reshaping existing linguistic tenets, particularly when extending research into multimodal and multichannel inputs. Future investigations might capitalize on these AI models to uncover deeper insights into language processing and acquisition, ultimately driving towards the overarching goal of comprehending the essence and structure of human language.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

Sign up for free to view the 2 tweets with 42 likes about this paper.