AAAI 2026  ·  Oral Presentation
Oral Presentation  ·  Top ~4.5% of accepted papers
|
23,680 submissions (record high)
|
17.6% acceptance rate
|
~188 oral papers total

Efficient Hallucination Detection: Adaptive Bayesian Estimation
of Semantic Entropy with Guided Semantic Exploration

Qiyao Sun1Xingming Li1Xixiang He1Ao Cheng1Xuanyu Ji1
Hailun Lu2Runke Huang3Qingyong Hu†2

Corresponding author

1 National University of Defense Technology 2 Intelligent Game and Decision Lab, Beijing 3 The Chinese University of Hong Kong, Shenzhen

Abstract

Large language models (LLMs) have achieved remarkable success in various natural language processing tasks, yet they remain prone to generating factually incorrect outputs—known as "hallucinations". While recent approaches have shown promise for hallucination detection by repeatedly sampling from LLMs and quantifying the semantic inconsistency among the generated responses, they rely on fixed sampling budgets that fail to adapt to query complexity, resulting in computational inefficiency.

We propose an Adaptive Bayesian Estimation framework for Semantic Entropy with Guided Semantic Exploration, which dynamically adjusts sampling requirements based on observed uncertainty. Our approach employs a hierarchical Bayesian framework to model the semantic distribution, enabling dynamic control of sampling iterations through variance-based thresholds that terminate generation once sufficient certainty is achieved. We also develop a perturbation-based importance sampling strategy to systematically explore the semantic space.

Extensive experiments on four QA datasets demonstrate that our method achieves superior hallucination detection performance with significant efficiency gains. In low-budget scenarios, our approach requires about 50% fewer samples to achieve comparable detection performance to existing methods, while delivers an average AUROC improvement of 12.6% under the same sampling budget.
Teaser: comparison of fixed sampling vs. adaptive Bayesian approach

Figure 1. Comparison of fixed sampling (a) versus our adaptive Bayesian approach (b). Fixed sampling wastes computational resources on simple queries while failing to discover semantic diversity in complex cases. Our method dynamically adjusts sampling based on variance thresholds, enabling efficient and accurate hallucination detection.

Algorithm Highlights

Key results on hallucination detection benchmarks

~0%
Fewer LLM Queries
Low-budget hallucination detection with same accuracy
0%
AUROC Gain (N=2)
Average improvement over strongest baseline
23/24
State-of-the-Art
Best AUROC across all model × dataset settings
4
QA Benchmarks
CoQA · TriviaQA · TruthfulQA · SimpleQA

Method Overview

Adaptive Bayesian Estimation with Guided Semantic Exploration

Adaptive Sampling Loop
Input Query
LLM Sampling
Initial N₀ responses
Bayesian Update
Posterior on K, Var[h|D]
Var[h|D] < γ?
No →
Guided Exploration
Perturb critical tokens
Yes →
Semantic Entropy
Final estimate
1

Hierarchical Bayesian Framework

Models the semantic distribution via a Dirichlet prior over semantic categories. Marginalizes over the unknown number of meanings K to compute E[h|D] and Var[h|D] with tighter posterior bounds through generation probability constraints.

2

Variance-Based Adaptive Sampling

Dynamically terminates sampling once posterior variance falls below threshold γ. Simple queries converge quickly; complex queries receive additional exploration— achieving ≈50% sample reduction in low-budget settings.

3

Guided Semantic Exploration

Identifies semantically critical tokens via importance weights and systematically perturbs them to discover diverse interpretations. Importance sampling maintains unbiased estimates while accelerating variance convergence.

Performance vs. Sampling Budget

AUROC as a function of sampling budget N — Llama-3.1-8B

Our method (solid blue) consistently outperforms baselines at every sampling budget N=1–10. At N=2, we achieve an average 12.6% higher AUROC than the strongest baseline.

Benchmark Results

AUROC on four QA datasets across three LLMs (higher is better)

Dataset P(True) SAR SE SE-SDLG Ours ↑

Bold values indicate state-of-the-art. Our method achieves the highest AUROC in 23 out of 24 experimental settings.

Authors

Qiyao Sun
Qiyao Sun
National University of Defense Technology
Xingming Li
Xingming Li
National University of Defense Technology
Xixiang He
Xixiang He
National University of Defense Technology
Ao Cheng
Ao Cheng
National University of Defense Technology
Xuanyu Ji
Xuanyu Ji
National University of Defense Technology
Hailun Lu
Intelligent Game and Decision Lab, Beijing
Runke Huang
Runke Huang
The Chinese University of Hong Kong, Shenzhen
Qingyong Hu
Qingyong Hu
Intelligent Game and Decision Lab, Beijing
Corresponding

BibTeX

@inproceedings{sun2026efficienthallucination, title = {Efficient Hallucination Detection: Adaptive {Bayesian} Estimation of Semantic Entropy with Guided Semantic Exploration}, author = {Sun, Qiyao and Li, Xingming and He, Xixiang and Cheng, Ao and Ji, Xuanyu and Lu, Hailun and Huang, Runke and Hu, Qingyong}, booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence}, year = {2026}, note = {Oral Presentation} }

Get in Touch

For questions about the paper, method, or potential collaborations, please contact the corresponding author.

Qingyong Hu

Qingyong Hu

Corresponding Author  ·  Intelligent Game and Decision Lab, Beijing

huqingyong15@outlook.com

We welcome questions about the adaptive Bayesian sampling framework, the guided semantic exploration strategy, and potential collaboration opportunities in LLM hallucination detection and uncertainty quantification. For code/implementation questions, you may also reach out to the first authors at sunqiyao18@nudt.edu.cn or lixingming@nudt.edu.cn.

Visitors Around the World