How to Steer LLM Latents for Hallucination Detection? Paper • 2503.01917 • Published 12 days ago • 10
Rethinking Mixture-of-Agents: Is Mixing Different Large Language Models Beneficial? Paper • 2502.00674 • Published Feb 2 • 13
Language Models Prefer What They Know: Relative Confidence Estimation via Confidence Preferences Paper • 2502.01126 • Published Feb 3 • 4
Concept Steerers: Leveraging K-Sparse Autoencoders for Controllable Generations Paper • 2501.19066 • Published Jan 31 • 12