Has Your Pretrained Model Improved? A Multi-head Posterior Based Approach (2401.02987v4)
Abstract: The emergence of pre-trained models has significantly impacted NLP and Computer Vision to relational datasets. Traditionally, these models are assessed through fine-tuned downstream tasks. However, this raises the question of how to evaluate these models more efficiently and more effectively. In this study, we explore a novel approach where we leverage the meta-features associated with each entity as a source of worldly knowledge and employ entity representations from the models. We propose using the consistency between these representations and the meta-features as a metric for evaluating pre-trained models. Our method's effectiveness is demonstrated across various domains, including models with relational datasets, LLMs and image models.
- Don’t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors. Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, 1:238–247, 2014.
- Gulp: a prediction-based metric between representations, 2022.
- Leo Breiman. Random forests. Machine learning, 45:5–32, 2001.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Xnli: Evaluating cross-lingual sentence representations. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2475–2485, 2018.
- Deconfounded representation similarity for comparison of neural networks, 2022.
- Pdt: Pretrained dual transformers for time-aware bipartite graphs. arXiv preprint arXiv:2306.01913, 2023.
- Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255, 2009. doi: 10.1109/CVPR.2009.5206848.
- Grounding representation similarity with statistical testing, 2021.
- Robustness (python library), 2019a. URL https://github.com/MadryLab/robustness.
- Adversarial robustness as a prior for learned representations, 2019b.
- Language models represent space and time. arXiv preprint arXiv:2310.02207, 2023.
- The movielens datasets: History and context. Acm transactions on interactive intelligent systems (tiis), 5(4):1–19, 2015.
- Self-attentive sequential recommendation. In 2018 IEEE international conference on data mining (ICDM), pp. 197–206. IEEE, 2018.
- Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of naacL-HLT, volume 1, pp.  2, 2019.
- Efficient estimation of word representations in vector space. In International Conference on Learning Representations, 2013. URL https://api.semanticscholar.org/CorpusID:5959482.
- George A. Miller. WordNet: A lexical database for English. In Human Language Technology: Proceedings of a Workshop held at Plainsboro, New Jersey, March 8-11, 1994, 1994. URL https://aclanthology.org/H94-1111.
- Probing neural network comprehension of natural language arguments. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 4658–4664, 2019.
- OpenAI. Gpt-4 technical report, 2023.
- Learning transferable visual models from natural language supervision, 2021.
- Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551, 2020.
- Image synthesis with a single (robust) classifier, 2019.
- Breeds: Benchmarks for subpopulation shift, 2020.
- Evaluation methods for unsupervised word embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 298–307, 2015.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100, 2022.
- Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
- Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations, 2020.
- Embeddingtree: Hierarchical exploration of entity features in embedding. In 2023 IEEE 16th Pacific Visualization Symposium (PacificVis), pp. 217–221. IEEE, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.