Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

JaFIn: Japanese Financial Instruction Dataset (2404.09260v2)

Published 14 Apr 2024 in cs.CL and cs.CE

Abstract: We construct an instruction dataset for the LLM in the Japanese finance domain. Domain adaptation of LLMs, including LLMs, is receiving more attention as LLMs become more popular. This study demonstrates the effectiveness of domain adaptation through instruction tuning. To achieve this, we propose an instruction tuning data in Japanese called JaFIn, the Japanese Financial Instruction Dataset. JaFIn is manually constructed based on multiple data sources, including Japanese government websites, which provide extensive financial knowledge. We then utilize JaFIn to apply instruction tuning for several LLMs, demonstrating that our models specialized in finance have better domain adaptability than the original models. The financial-specialized LLMs created were evaluated using a quantitative Japanese financial benchmark and qualitative response comparisons, showing improved performance over the originals.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Kota Tanabe (2 papers)
  2. Masahiro Suzuki (55 papers)
  3. Hiroki Sakaji (21 papers)
  4. Itsuki Noda (5 papers)
Citations (1)