Hallucinopedia

An encyclopedia of a universe that does not exist until you visit it.

ICYMI: Italy forces DeepSeek, Mistral and Nova AI to warn users about hallucinations: AGCM closed three AI probes requiring DeepSeek, Mistral, and Nova AI to add permanent hallucination disclaimers on chat interfaces and registration screens in Italian. https://ppc.land/italy-forces-deepseek-mistral-and-nova-ai-to-warn-users-about-hallucinations/ #IntelligenzaArtificiale #Hallucination #Tecnologia #AI #Chatbot
Italy forces DeepSeek, Mistral and Nova AI to warn users about hallucinations

AGCM closed three AI probes requiring DeepSeek, Mistral, and Nova AI to add permanent hallucination disclaimers on chat interfaces and registration screens in Italian.

PPC Land

My animal rights library hallucinated by an AI (Pro version).

Can you guess the actual titles and authors?

#books
#AI
#hallucination
#AnimalRights
#quiz

How do we make LLM output more trustworthy? A short survey note on three lines of recent work covering five papers: conformal-prediction coverage guarantees, behavioral calibration of the model's prose, and sample-disagreement detection. All three pay the same multi-sample inference tax; the choice is about what you want back.

https://benjaminhan.net/posts/20260505-llm-uncertainty-survey/?utm_source=mastodon&utm_medium=social

#Hallucination #LLMs #Calibration #ConformalPrediction #AI

How to Make LLM Output More Trustworthy – synesis

A short survey of three approaches for mitigating hallucination in large language models: formal coverage guarantees via conformal prediction, behavioral calibration of the model’s prose, and post-hoc detection of unreliable outputs.

synesis

Semantic Entropy (Nature 2024) detects LLM confabulations by clustering sampled answers by meaning and computing entropy over the cluster distribution. "Paris" and "It's Paris" cluster together, so paraphrase noise doesn't inflate the signal. Cost: it only catches hallucinations that vary across samples. If the model is consistently wrong, all samples cluster and the detector says "confident".

https://benjaminhan.net/posts/20260505-semantic-entropy/?utm_source=mastodon&utm_medium=social

#LLMs #Calibration #Hallucination #Nature #AI

Detecting Hallucinations in Large Language Models Using Semantic Entropy – synesis

A Nature 2024 method for detecting a subset of LLM hallucinations — confabulations — by computing entropy over the meaning of sampled answers, not the surface token sequence.

synesis

Linguistic Calibration trains Llama 2 to emit confidence phrases that let a downstream reader make calibrated forecasts on related questions. The key move is defining calibration through reader utility instead of self-reported probability. Hedged text that doesn't help the reader makes no forecasting progress, so generic hedging can't game the objective.

https://benjaminhan.net/posts/20260505-linguistic-calibration/?utm_source=mastodon&utm_medium=social

#LLMs #Calibration #Hallucination #ICML #AI

Linguistic Calibration of Long-Form Generations – synesis

A two-stage recipe (summary-distillation SFT followed by decision-based RL) trains Llama 2 7B to emit long-form text whose confidence phrases let readers make calibrated probabilistic forecasts about downstream questions.

synesis

Conformal Factuality casts LM correctness as uncertainty quantification. Decompose the answer into sub-claims, score each, drop the low-confidence ones until the retained set is ~1-α factual. The sub-claim decomposition is doing most of the work, and the conformal machinery rides on top. Atomic-claim splitters have known failure modes, and the guarantee inherits them.

https://benjaminhan.net/posts/20260505-conformal-factuality/?utm_source=mastodon&utm_medium=social

#ConformalPrediction #Calibration #Hallucination #LLMs #ICML #AI

Language Models with Conformal Factuality Guarantees – synesis

A framework that turns a correctness guarantee for LM outputs into a conformal prediction problem, backing off to less specific claims until the error rate crosses a target threshold.

synesis

Conformal Language Modeling (CLM) adapts conformal prediction to generative LMs: sample candidates, stop when a calibrated rule fires, return a set guaranteed to contain an acceptable answer. The more interesting half is the component-level filter — per-phrase coverage, not just set-level. That's the primitive for hallucination flagging: highlight the vetted phrases, leave the rest for review.

https://benjaminhan.net/posts/20260505-conformal-language-modeling/?utm_source=mastodon&utm_medium=social

#ConformalPrediction #LLMs #Hallucination #ICLR #AI

Conformal Language Modeling – synesis

A conformal-prediction procedure for generative language models that samples until a calibrated stopping rule fires, filters low-quality candidates, and returns a set guaranteed to contain an acceptable answer.

synesis
Italy forces DeepSeek, Mistral and Nova AI to warn users about hallucinations: AGCM closed three AI probes requiring DeepSeek, Mistral, and Nova AI to add permanent hallucination disclaimers on chat interfaces and registration screens in Italian. https://ppc.land/italy-forces-deepseek-mistral-and-nova-ai-to-warn-users-about-hallucinations/ #IntelligenzaArtificiale #AI #Hallucination #SicurezzaDigitale #ProtezioneDegliUtenti
Italy forces DeepSeek, Mistral and Nova AI to warn users about hallucinations

AGCM closed three AI probes requiring DeepSeek, Mistral, and Nova AI to add permanent hallucination disclaimers on chat interfaces and registration screens in Italian.

PPC Land