Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

K2: A Foundation Language Model for Geoscience Knowledge Understanding and Utilization (2306.05064v2)

Published 8 Jun 2023 in cs.CL and cs.AI

Abstract: LLMs have achieved great success in general domains of natural language processing. In this paper, we bring LLMs to the realm of geoscience with the objective of advancing research and applications in this field. To this end, we present the first-ever LLM in geoscience, K2, alongside a suite of resources developed to further promote LLM research within geoscience. For instance, we have curated the first geoscience instruction tuning dataset, GeoSignal, which aims to align LLM responses to geoscience-related user queries. Additionally, we have established the first geoscience benchmark, GeoBench, to evaluate LLMs in the context of geoscience. In this work, we experiment with a complete recipe to adapt a pre-trained general-domain LLM to the geoscience domain. Specifically, we further train the LLaMA-7B model on 5.5B tokens of geoscience text corpus, including over 1 million pieces of geoscience literature, and utilize GeoSignal's supervised data to fine-tune the model. Moreover, we share a protocol that can efficiently gather domain-specific data and construct domain-supervised data, even in situations where manpower is scarce. Meanwhile, we equip K2 with the abilities of using tools to be a naive geoscience aide. Experiments conducted on the GeoBench demonstrate the effectiveness of our approach and datasets on geoscience knowledge understanding and utilization.We open-source all the training data and K2 model checkpoints at https://github.com/davendw49/k2.

An Expert Overview of the K2 LLM for Geoscience Knowledge

The paper presents K2, a specialized LLM (LM) calibrated specifically for the geoscience domain. K2 emerges as the first geoscience-focused LLM, constructed by further training an existing general-domain model (LLaMA-7B) on a substantial corpus of geoscience-specific texts, involving approximately 5.5 billion tokens collected from academic literature, Wikipedia pages, and other pertinent sources.

Key Contributions

  1. Specialized Tuning Data: The authors introduce a novel instruction tuning dataset named GeoSignal, crafted to enhance LMs' alignment with geoscience-related queries. Additionally, they propose GeoBench, a benchmark explicitly designed to evaluate LMs' proficiency in geoscience tasks, comprising objective and subjective questions derived from academic and standardized tests.
  2. Training and Fine-Tuning: K2's training pipeline includes two principal stages—further pre-training on domain-specific texts and instruction tuning using GeoSignal and other datasets. This approach intends to align K2 with both general human instructions and expert-level geoscience knowledge.
  3. Tool Augmentation: The model is endowed with tool-use capabilities, enabling it to support tasks typical for geoscientists, such as academic literature searches through a tool named GeoSearch. This feature allows K2 to autonomously draw upon external resources, enhancing its value as a research and knowledge assistant.

Evaluation and Performance

The effectiveness of K2 is demonstrated through comprehensive evaluations outlined in GeoBench. The model outperforms similar-sized models, leveraging both objective metrics and human evaluators. In particular, K2 showcases superior capability in knowledge understanding, and reasoning, significantly distinguishing itself in subjective evaluations where understanding and expertise in geoscience are crucial.

Theoretical and Practical Implications

Practically, K2 stands to serve as a substantial asset for geoscientists, functioning as an intelligent assistant capable of engaging with complex data queries, conducting searches, and generating coherent, informed responses that align with expert knowledge. Theoretically, this work suggests potential advancements in domain-specific LLM adaptation. The structured approach to data collection, preprocessing, and instruction tuning reveals a scalable blueprint for similar endeavors in other specialized fields. The integration of tools further emphasizes the expanding role of LMs in incorporating dynamic datasets into ongoing interactions within specialized domains.

Future Developments

Looking ahead, this research sets a foundational platform for progressively larger and more sophisticated models within the geoscience sector. Potential developments may include expanding K2's capacity to incorporate real-time data flows, enhancing its adaptability with more sophisticated tool integrations, and exploring its application to other interdisciplinary areas within natural and life sciences.

Conclusion

The introduction of K2 aligns with a growing trajectory towards specialized LMs that address particular academic domains. This model not only enriches the resources available to geoscientists but also illustrates the broader applicability of specialized LMs across various complex domains requiring a granular understanding of the subject matter. The open-sourced data and tools pave the way for further developments and collaborations within the geoscience community and beyond.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Cheng Deng (67 papers)
  2. Tianhang Zhang (16 papers)
  3. Zhongmou He (5 papers)
  4. Yi Xu (304 papers)
  5. Qiyuan Chen (22 papers)
  6. Yuanyuan Shi (62 papers)
  7. Luoyi Fu (41 papers)
  8. Weinan Zhang (322 papers)
  9. Xinbing Wang (98 papers)
  10. Chenghu Zhou (55 papers)
  11. Zhouhan Lin (57 papers)
  12. Junxian He (66 papers)
Citations (43)