Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Geotechnical Parrot Tales (GPT): Harnessing Large Language Models in geotechnical engineering (2304.02138v3)

Published 4 Apr 2023 in cs.CL and physics.geo-ph

Abstract: The widespread adoption of LLMs, such as OpenAI's ChatGPT, could revolutionize various industries, including geotechnical engineering. However, GPT models can sometimes generate plausible-sounding but false outputs, leading to hallucinations. In this article, we discuss the importance of prompt engineering in mitigating these risks and harnessing the full potential of GPT for geotechnical applications. We explore the challenges and pitfalls associated with LLMs and highlight the role of context in ensuring accurate and valuable responses. Furthermore, we examine the development of context-specific search engines and the potential of LLMs to become a natural interface for complex tasks, such as data analysis and design. We also develop a unified interface using natural language to handle complex geotechnical engineering tasks and data analysis. By integrating GPT into geotechnical engineering workflows, professionals can streamline their work and develop sustainable and resilient infrastructure systems for the future.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (7)
  1. “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, 2021, pp. 610–623
  2. “Sparks of artificial general intelligence: Early experiments with gpt-4” In arXiv preprint arXiv:2303.12712, 2023
  3. “How context affects language models’ factual predictions” In arXiv preprint arXiv:2005.04611, 2020
  4. “Llama: Open and efficient foundation language models” In arXiv preprint arXiv:2302.13971, 2023
  5. “Attention is all you need” In Advances in neural information processing systems 30, 2017
  6. “Chain of thought prompting elicits reasoning in large language models” In arXiv preprint arXiv:2201.11903, 2022
  7. “React: Synergizing reasoning and acting in language models” In arXiv preprint arXiv:2210.03629, 2022
Citations (5)

Summary

We haven't generated a summary for this paper yet.