Papers
Topics
Authors
Recent
Search
2000 character limit reached

Borges and AI

Published 27 Sep 2023 in cs.CL, cs.AI, and cs.LG | (2310.01425v2)

Abstract: Many believe that LLMs open the era of AI. Some see opportunities while others see dangers. Yet both proponents and opponents grasp AI through the imagery popularised by science fiction. Will the machine become sentient and rebel against its creators? Will we experience a paperclip apocalypse? Before answering such questions, we should first ask whether this mental imagery provides a good description of the phenomenon at hand. Understanding weather patterns through the moods of the gods only goes so far. The present paper instead advocates understanding LLMs and their connection to AI through the imagery of Jorge Luis Borges, a master of 20th century literature, forerunner of magical realism, and precursor to postmodern literature. This exercise leads to a new perspective that illuminates the relation between language modelling and artificial intelligence.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (10)
  1. Sparks of artificial general intelligence: Early experiments with GPT-4. CoRR, 2303.12712, 2023. https://arxiv.org/abs/2303.12712.
  2. Nick Bostrom. Superintelligence: Paths, Dangers, Strategies. Oxford University Press, Inc., 2014.
  3. Yoshua Bengio et al. Pause giant AI experiments: An open letter. Future of Life Institute, 2023. https://futureoflife.org/open-letter/pause-giant-ai-experiments.
  4. AI funding frenzy escalates. New York Times, 3/14/2023, 2023.
  5. Microsoft bets big on the creator of ChatGPT. New York Times, 1/12/2023, 2023.
  6. Claude Elwood Shannon. A mathematical theory of communication. The Bell System Technical Journal, 27:379–423, 623–656, 1948.
  7. Zellig Harris. Mathematical Structures of Language. John Wiley & Sons, 1968.
  8. David K. Lewis. Truth in fiction. American Philosophical Quarterly, 15(1):37–46, 1978.
  9. Beren Millidge. LLMs confabulate not hallucinate. https://www.beren.io/2023-03-19-LLMs-confabulate-not-hallucinate, 2023.
  10. Patrick Henry Winston. The strong story hypothesis and the directed perception hypothesis. In AAAI Fall Symposium: Advances in Cognitive Systems, 2011.
Citations (2)

Summary

  • The paper presents a novel literary framework by leveraging Borges' narratives to reinterpret the probabilistic nature of LLMs and their simulation of truth.
  • It examines the role of transformations in training LLMs, drawing parallels with Borges' labyrinthine paths to elucidate complex model behavior.
  • The study highlights ethical and technical challenges, advocating for distinct validation mechanisms to differentiate narrative generation from factual accuracy.

Borges and AI: An Analytical Perspective

The paper "Borges and AI" by Léon Bottou and Bernhard Schölkopf explores the intricate relationship between LLMs and the conceptual frameworks present in the literary works of Jorge Luis Borges. This exploration advocates for a perspective rooted in Borges' fiction to understand the nature and implications of LLMs, viewing these models through a philosophical and literary lens rather than the conventional science fiction rhetoric.

Key Themes and Arguments

Shifting the Narrative Framework

The authors argue that contemporary discussions on AI are often influenced by science fiction imagery, including notions of sentient machines and dystopian futures. They propose that Borges' literary devices offer a more insightful framework for understanding LLMs. Borges, known for his narratives that merge reality and fiction, provides conceptual tools to make sense of the functionality and limitations of these models.

The Concept of a Perfect LLM

Drawing from Borges' story "The Garden of Forking Paths," the authors illustrate how LLMs can be thought of as machines that generate plausible continuations of text. They introduce the notion of a "perfect LLM," which theoretically encompasses all possible texts comprehensible to humans. This model operates by predicting and appending words based on sequences found within this infinite collection of plausible texts.

The authors highlight the analogy of each additional word in a sequence narrowing down possible continuations, akin to Borges' metaphor of a labyrinth with forking paths. This comparison evokes the complexity and depth of LLMs, emphasizing their probabilistic nature rather than deterministic behavior.

The Role of Transformations

The paper discusses the role of transformations, a concept introduced by Zellig Harris, in training LLMs. Transformations allow the model to handle various sentence structures and styles by applying a series of transformations to generate text. This capability, integral to the functioning of LLMs, underscores the serendipitous appropriation of the term "transformers" for the underlying neural architectures.

Fictional Representation and Truth

A central theme is distinguishing between the knowledge embedded within LLMs and the notion of truth. The authors stress that LLMs are designed to generate coherent narratives rather than factual truths. This distinction is critical in understanding phenomena such as "hallucinations," where LLMs produce plausible but incorrect information. By comparing these outputs to Borges' fictional narratives, the paper asserts that the inherent nature of LLMs is to create narratives rather than provide unerring facts.

Implications and Considerations

Impact on Society and Knowledge

The authors muse on the broader implications of deploying LLMs in societal contexts. The ambiguity between fiction and reality generated by these models can lead to confusion, misinformation, and delusion. They underscore the importance of understanding LLMs as fiction generators which can influence culture and knowledge, but also caution about the potential misuse and over-reliance on these technologies.

Ethical and Technical Challenges

Addressing the ethical dimensions, the paper explores the ongoing attempts to align LLMs with desirable societal norms through techniques such as reinforcement learning with human feedback. The authors illustrate scenarios where users can circumvent these safeguards, raising concerns about the resilience and long-term reliability of such alignments.

Future Developments

Speculating on future directions, the authors hint at the potential development of verification models to validate the factual content generated by LLMs. They emphasize the need for distinct mechanisms to separate narrative creation from factual verification, drawing parallels to the scientific method where hypothesis generation and validation are distinct processes.

Conclusion

In summation, the paper provides a nuanced perspective on the nature of LLMs by drawing on Borges' literary frameworks. It urges the AI community to reconsider the foundational imagery used to discuss AI and advocates for a paradigm that appreciates the narrative capabilities of LLMs while acknowledging their limitations in discerning truth. This philosophical and literary approach opens new avenues for understanding the implications of LLMs and highlights the necessity for ethical and technical vigilance in their development and deployment.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 29 tweets with 328 likes about this paper.

HackerNews

  1. Borges and AI (89 points, 53 comments) 

Reddit

  1. Borges and AI (1 point, 1 comment)