Evaluating and Enhancing Language Model Factuality
Автор: Ai2
Загружено: 2025-03-07
Просмотров: 227
Abstract: Language models (LMs) are increasingly adopted in real-world applications, yet their tendency to generate factual errors remains a major concern. In this talk, I will describe my work on LM factuality, i.e., its consistency with established facts. I address factuality challenges across two key dimensions: evaluation and enhancement. On the evaluation front, I will present a factuality evaluation framework comprising an updatable benchmark curated from real-world LM usage and a fine-grained evaluation technique that robustly identifies LM inaccuracies. For factuality enhancement, I will propose two complementary approaches: (1) a post-processing framework that verifies and refines LM outputs against external knowledge sources; and (2) learnable intervention systems that leverages LMs' internal representations of truth to adjust generations at inference time. Together, these methods advance our understanding of factuality challenges and offer practical pathways to improve LM reliability.
Bio: Farima Fatahi Bayat is a Ph.D. candidate in the Computer Science and Engineering Department at University of Michigan, advised by Prof. H. Jagadish and Prof. Lu Wang. Her research focuses on advancing responsible AI, with a particular emphasis on enhancing the factuality of Language Models (LMs). Her recent works include creating evaluation benchmarks to assess LMs’ factuality, designing adaptive intervention frameworks that enable uncertainty expression, and building correction mechanisms to increase the quality of LM output.

Доступные форматы для скачивания:
Скачать видео mp4
-
Информация по загрузке: