|国家预印本平台
首页|UQLM: A Python Package for Uncertainty Quantification in Large Language Models

UQLM: A Python Package for Uncertainty Quantification in Large Language Models

UQLM: A Python Package for Uncertainty Quantification in Large Language Models

来源:Arxiv_logoArxiv
英文摘要

Hallucinations, defined as instances where Large Language Models (LLMs) generate false or misleading content, pose a significant challenge that impacts the safety and trust of downstream applications. We introduce UQLM, a Python package for LLM hallucination detection using state-of-the-art uncertainty quantification (UQ) techniques. This toolkit offers a suite of UQ-based scorers that compute response-level confidence scores ranging from 0 to 1. This library provides an off-the-shelf solution for UQ-based hallucination detection that can be easily integrated to enhance the reliability of LLM outputs.

Dylan Bouchard、Mohit Singh Chauhan、David Skarbrevik、Ho-Kyeong Ra、Viren Bajaj、Zeya Ahmad

计算技术、计算机技术

Dylan Bouchard,Mohit Singh Chauhan,David Skarbrevik,Ho-Kyeong Ra,Viren Bajaj,Zeya Ahmad.UQLM: A Python Package for Uncertainty Quantification in Large Language Models[EB/OL].(2025-07-08)[2025-08-02].https://arxiv.org/abs/2507.06196.点此复制

评论