Papers
Topics
Authors
Recent
Search
2000 character limit reached

Alignment at Pre-training! Towards Native Alignment for Arabic LLMs

Published 4 Dec 2024 in cs.CL | (2412.03253v1)

Abstract: The alignment of LLMs is critical for developing effective and safe LLMs. Traditional approaches focus on aligning models during the instruction tuning or reinforcement learning stages, referred to in this paper as post alignment'. We argue that alignment during the pre-training phase, which we termnative alignment', warrants investigation. Native alignment aims to prevent unaligned content from the beginning, rather than relying on post-hoc processing. This approach leverages extensively aligned pre-training data to enhance the effectiveness and usability of pre-trained models. Our study specifically explores the application of native alignment in the context of Arabic LLMs. We conduct comprehensive experiments and ablation studies to evaluate the impact of native alignment on model performance and alignment stability. Additionally, we release open-source Arabic LLMs that demonstrate state-of-the-art performance on various benchmarks, providing significant benefits to the Arabic LLM community.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 3 likes about this paper.