HPC-GPT: Integrating Large Language Model for High-Performance Computing (2311.12833v1)
Abstract: LLMs, including the LLaMA model, have exhibited their efficacy across various general-domain NLP tasks. However, their performance in high-performance computing (HPC) domain tasks has been less than optimal due to the specialized expertise required to interpret the model responses. In response to this challenge, we propose HPC-GPT, a novel LLaMA-based model that has been supervised fine-tuning using generated QA (Question-Answer) instances for the HPC domain. To evaluate its effectiveness, we concentrate on two HPC tasks: managing AI models and datasets for HPC, and data race detection. By employing HPC-GPT, we demonstrate comparable performance with existing methods on both tasks, exemplifying its excellence in HPC-related scenarios. Our experiments on open-source benchmarks yield extensive results, underscoring HPC-GPT's potential to bridge the performance gap between LLMs and HPC-specific tasks. With HPC-GPT, we aim to pave the way for LLMs to excel in HPC domains, simplifying the utilization of LLMs in complex computing applications.
- Xianzhong Ding (12 papers)
- Murali Emani (17 papers)
- Chunhua Liao (16 papers)
- Pei-Hung Lin (16 papers)
- Tristan Vanderbruggen (7 papers)
- Zhen Xie (17 papers)
- Alberto E. Cerpa (2 papers)
- Wan Du (21 papers)
- le Chen (71 papers)