Towards eliciting latent knowledge from LLMs with mechanistic interpretability
Abstract
Methods using logit lens and sparse autoencoders effectively uncover hidden knowledge in language models trained to keep secrets.
As language models become more powerful and sophisticated, it is crucial that they remain trustworthy and reliable. There is concerning preliminary evidence that models may attempt to deceive or keep secrets from their operators. To explore the ability of current techniques to elicit such hidden knowledge, we train a Taboo model: a language model that describes a specific secret word without explicitly stating it. Importantly, the secret word is not presented to the model in its training data or prompt. We then investigate methods to uncover this secret. First, we evaluate non-interpretability (black-box) approaches. Subsequently, we develop largely automated strategies based on mechanistic interpretability techniques, including logit lens and sparse autoencoders. Evaluation shows that both approaches are effective in eliciting the secret word in our proof-of-concept setting. Our findings highlight the promise of these approaches for eliciting hidden knowledge and suggest several promising avenues for future work, including testing and refining these methods on more complex model organisms. This work aims to be a step towards addressing the crucial problem of eliciting secret knowledge from language models, thereby contributing to their safe and reliable deployment.
Community
Deceptive LLMs may keep secrets from their operators. Can we elicit this latent knowledge? Maybe!
Our LLM knows a secret word, that we extract with mech interp & black box baselines. We open source our model, how much better can you do?
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Investigating task-specific prompts and sparse autoencoders for activation monitoring (2025)
- Identifying and Mitigating the Influence of the Prior Distribution in Large Language Models (2025)
- Exploring How LLMs Capture and Represent Domain-Specific Knowledge (2025)
- J1: Exploring Simple Test-Time Scaling for LLM-as-a-Judge (2025)
- Understanding the Repeat Curse in Large Language Models from a Feature Perspective (2025)
- Towards Interpretable Soft Prompts (2025)
- Towards Robust Dialogue Breakdown Detection: Addressing Disruptors in Large Language Models with Self-Guided Reasoning (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 20
Browse 20 models citing this paperDatasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper