In authors or contributors
Publication year

1 resource

  • Jasmin Wachter, Michael Radloff, Maja Sm...
    |
    Mar 17th, 2025
    |
    preprint
    Jasmin Wachter, Michael Radloff, Maja Sm...
    Mar 17th, 2025

    We introduce an Item Response Theory (IRT)-based framework to detect and quantify socioeconomic bias in large language models (LLMs) without relying on subjective human judgments. Unlike traditional methods, IRT accounts for item difficulty, improving ideological bias estimation. We fine-tune two LLM families (Meta-LLaMa 3.2-1B-Instruct and Chat- GPT 3.5) to represent distinct ideological positions and introduce a two-stage approach: (1) modeling response avoidance and (2) estimating...

Last update from database: 27/10/2025, 21:15 (UTC)
Powered by Zotero and Kerko.