Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Panda LLM: Training Data and Evaluation for Open-Sourced Chinese Instruction-Following Large Language Models (2305.03025v1)

Published 4 May 2023 in cs.CL and cs.AI

Abstract: This project focuses on enhancing open-source LLMs through instruction-tuning and providing comprehensive evaluations of their performance. We explore how various training data factors, such as quantity, quality, and linguistic distribution, influence the performance of instruction-tuned models trained on publicly accessible high-quality instruction datasets for both English and Chinese languages. Our goal is to supplement evaluation with quantitative analyses, providing valuable insights for the continued advancement of open-source chat models. Our model, data, and code are publicly available for others to use and build upon.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Fangkai Jiao (19 papers)
  2. Bosheng Ding (16 papers)
  3. Tianze Luo (11 papers)
  4. Zhanfeng Mo (3 papers)
Citations (3)