Unknown training token count for Mistral 7B
Ascertain the number of training tokens used to train Mistral 7B to enable precise comparisons with Llama-2 and Solar and to evaluate the role of training data volume in preference learning and brittleness.
References
The number of training tokens is unknown.
                — Do Large Language Models Learn Human-Like Strategic Preferences?
                
                (2404.08710 - Roberts et al., 11 Apr 2024) in Section 3.4, Why are Solar and Mistral Not Brittle?