Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 130 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 29 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 76 tok/s Pro
Kimi K2 196 tok/s Pro
GPT OSS 120B 434 tok/s Pro
Claude Sonnet 4.5 39 tok/s Pro
2000 character limit reached

Efficient Knowledge Probing of Large Language Models by Adapting Pre-trained Embeddings (2508.06030v1)

Published 8 Aug 2025 in cs.CL and cs.LG

Abstract: LLMs acquire knowledge across diverse domains such as science, history, and geography encountered during generative pre-training. However, due to their stochasticity, it is difficult to predict what LLMs have acquired. Prior work has developed different ways to probe this knowledge by investigating the hidden representations, crafting specific task prompts, curating representative samples, and estimating their uncertainty. However, these methods require making forward passes through the underlying model to probe the LLM's knowledge about a specific fact, making them computationally expensive and time-consuming. To bridge this gap, we propose $\textbf{PEEK}$ or $\textbf{P}$roxy $\textbf{E}$mbeddings to $\textbf{E}$stimate $\textbf{K}$nowledge of LLMs, by leveraging the pre-trained embedding models that effectively encode factual knowledge as text or graphs as proxies for LLMs. First, we identify a training set of facts known by LLMs through various probing strategies and then adapt embedding models to predict the LLM outputs with a linear decoder layer. Comprehensive evaluation on $3$ Wikipedia-derived datasets, $4$ LLMs, and $7$ embedding models shows that embeddings can predict LLM knowledge on a held-out set with up to 90 % accuracy. Furthermore, we find that sentence embedding models are more suitable than graph embeddings to predict LLM knowledge, shedding light on the underlying representation of the factual landscape. Thus, we believe that knowledge-adapted embeddings can be used to identify knowledge gaps in LLMs at scale and can provide deeper insights into LLMs' internal inductive bias. The code and data are made available at https://github.com/claws-lab/peek.

Summary

  • The paper introduces PEEK, a method that efficiently probes LLM knowledge by adapting pre-trained embeddings instead of using costly direct queries.
  • It leverages techniques like sentence and graph embeddings with proxy tuning (e.g., low-rank and linear tuning) to predict factual accuracy.
  • Experimental results show high performance with up to 91% accuracy and 88% AUC, demonstrating the approach's scalability and efficiency.

Efficient Knowledge Probing of LLMs by Adapting Pre-trained Embeddings

Introduction

The paper "Efficient Knowledge Probing of LLMs by Adapting Pre-trained Embeddings" proposes a novel approach to probing knowledge in LLMs by using Proxy Embeddings to Estimate Knowledge (PEEK). This method leverages pre-trained embedding models to determine which facts are known by LLMs without requiring multiple forward passes through the model. LLMs have emerged as general-purpose knowledge bases, yet their stochastic learning objectives obscure their learned knowledge landscapes. Existing knowledge probing techniques often necessitate computationally expensive operations and access to the model's internal states, limiting their scalability.

PEEK addresses this challenge by adapting existing embedding representations to predict LLM knowledge, offering a scalable and efficient alternative to traditional probing methods. Figure 1

Figure 1: Comparison of our proposed approach, Proxy Embeddings to Estimate Knowledge (PEEK) with other knowledge probing approaches.

Methodology

PEEK capitalizes on the capabilities of embedding models to predict LLM knowledge through several key steps:

  1. Knowledge Probing Functions: The approach employs four distinct probing functions to determine LLM knowledge:
    • Binary Generation: Simply asks whether a statement is true or false.
    • Binary Logits Generation: Extracts the logits of expected tokens for a given prompt.
    • Binary Activation Prediction: Utilizes hidden representations to predict if a fact is true.
    • Fact Generation: Leverages existing datasets to determine if the LLM-supported facts are accurate.
  2. Proxy Embeddings: The foundation lies in embedding models, including:
  3. Proxy Tuning: This involves refining embedding models using techniques such as Low-rank tuning and Linear tuning to align them with LLM predictions. The adapted embeddings then serve as proxies for estimating LLM knowledge without querying the LLM directly. Figure 2

    Figure 2: Proxy Embeddings to Estimate Knowledge (PEEK): In this framework, pre-trained embedding models are adapted to match the LLM knowledge for a training set of facts identified using different probing mechanisms. On a held-out set, we can then predict whether an LLM knows a fact or not by using the fact's embedding.

Experimental Setup

The experiments evaluate PEEK on various datasets, including DBP100k and YAGO310, downsampled to control complexity and focus on positive facts. A range of embedding models was tested for knowledge estimation, and several LLMs, such as GPT-4o, GPT-4o-mini, and Llama models, were used to benchmark the efficacy of PEEK.

Metrics such as Accuracy (ACC) and Area Under the ROC Curve (AUC) for binary tasks, and Mean Absolute Error (MAE) for continuous predictions, quantify performance.

Results and Analysis

  • Binary Generation: High performance with up to 91% accuracy and 88% AUC was achieved using sentence embeddings, verifying the effectiveness of PEEK compared to more traditional hidden representation methods. Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Effect of changing the number of negative samples in GPT models for knowledge graphs.

  • Negative Sampling & Fine Tuning: Results indicate that models generally perform better with increased negative samples, reflecting natural class imbalances. Furthermore, linear tuning was competitive with more complex LoRA tuning, suggesting minimal fine-tuning is sufficient. Figure 4

Figure 4

Figure 4

Figure 4

Figure 4

Figure 4: Effect of changing the number of negative samples in Llama3.1-8B for knowledge graphs.

Conclusion

This research introduces PEEK as a scalable method to probe LLM knowledge efficiently. By adapting pre-trained embeddings, PEEK provides an avenue to assess knowledge without requiring costly computations linked with direct LLM querying. Future directions could explore how these proxy embeddings dynamically adapt as LLMs evolve and expand their knowledge bases. Additionally, understanding the implications of embedding-based knowledge probing on the continual learning and updating of LLMs could be a significant area of further research.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 tweets and received 10 likes.

Upgrade to Pro to view all of the tweets about this paper: