Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HPC-GPT: Integrating Large Language Model for High-Performance Computing (2311.12833v1)

Published 3 Oct 2023 in cs.DC, cs.AI, and cs.CL

Abstract: LLMs, including the LLaMA model, have exhibited their efficacy across various general-domain NLP tasks. However, their performance in high-performance computing (HPC) domain tasks has been less than optimal due to the specialized expertise required to interpret the model responses. In response to this challenge, we propose HPC-GPT, a novel LLaMA-based model that has been supervised fine-tuning using generated QA (Question-Answer) instances for the HPC domain. To evaluate its effectiveness, we concentrate on two HPC tasks: managing AI models and datasets for HPC, and data race detection. By employing HPC-GPT, we demonstrate comparable performance with existing methods on both tasks, exemplifying its excellence in HPC-related scenarios. Our experiments on open-source benchmarks yield extensive results, underscoring HPC-GPT's potential to bridge the performance gap between LLMs and HPC-specific tasks. With HPC-GPT, we aim to pave the way for LLMs to excel in HPC domains, simplifying the utilization of LLMs in complex computing applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Xianzhong Ding (12 papers)
  2. Murali Emani (17 papers)
  3. Chunhua Liao (16 papers)
  4. Pei-Hung Lin (16 papers)
  5. Tristan Vanderbruggen (7 papers)
  6. Zhen Xie (17 papers)
  7. Alberto E. Cerpa (2 papers)
  8. Wan Du (21 papers)
  9. le Chen (71 papers)
Citations (21)

Summary

We haven't generated a summary for this paper yet.