Conjecture: Scaling In-Context Learning to the Pretraining-Corpus Level
Establish whether scaling in-context learning so that models can access and utilize the entire pretraining corpus as context yields stronger and more robust knowledge capabilities than conventional parametric storage.
References
To summarize the viewpoints discussed in this blog post, we propose two conjectures regarding potential improvements to the LLM knowledge paradigm: (1) In-context learning demonstrates certain advantages over traditional LLM knowledge modeling paradigm, and could potentially be scaled up to the pre-training corpus level to enable models to acquire stronger and more robust knowledge capabilities; (2) The hidden states of sequence models may offer a highly generalizable mechanism for knowledge encoding and utilizing, and could potentially serve as a major knowledge storage module.