Are LLMs (Really) Ideological? An IRT-based Analysis and Alignment Tool for Perceived Socio-Economic Bias in LLMs
Article Status
Published
Authors/contributors
- Wachter, Jasmin (Author)
- Radloff, Michael (Author)
- Smolej, Maja (Author)
- Kinder-Kurlanda, Katharina (Author)
Title
Are LLMs (Really) Ideological? An IRT-based Analysis and Alignment Tool for Perceived Socio-Economic Bias in LLMs
Abstract
We introduce an Item Response Theory (IRT)-based framework to detect and quantify socioeconomic bias in large language models (LLMs) without relying on subjective human judgments. Unlike traditional methods, IRT accounts for item difficulty, improving ideological bias estimation. We fine-tune two LLM families (Meta-LLaMa 3.2-1B-Instruct and Chat- GPT 3.5) to represent distinct ideological positions and introduce a two-stage approach: (1) modeling response avoidance and (2) estimating perceived bias in answered responses. Our results show that off-the-shelf LLMs often avoid ideological engagement rather than exhibit bias, challenging prior claims of partisanship. This empirically validated framework enhances AI alignment research and promotes fairer AI governance.
Repository
arXiv
Archive ID
arXiv:2503.13149
Date
2025-03-17
Accessed
08/07/2025, 21:28
Short Title
Are LLMs (Really) Ideological?
Library Catalogue
Extra
arXiv:2503.13149 [cs]
Citation Key: wachter2025
<标题>: 大型语言模型(真的)有意识形态倾向吗?一种基于项目反应理论的分析与校准工具,用于感知大型语言模型中的社会经济偏见
<AI Smry>: An Item Response Theory (IRT)-based framework to detect and quantify socioeconomic bias in large language models (LLMs) without relying on subjective human judgments is introduced, challenging prior claims of partisanship.
Citation
Wachter, J., Radloff, M., Smolej, M., & Kinder-Kurlanda, K. (2025). Are LLMs (Really) Ideological? An IRT-based Analysis and Alignment Tool for Perceived Socio-Economic Bias in LLMs (arXiv:2503.13149). arXiv. https://doi.org/10.48550/arXiv.2503.13149
Link to this record