• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Detecting Linguistic Indicators for Stereotype Assessment with Large Language Models
 
  • Details
  • Full
Options
June 23, 2025
Conference Paper
Title

Detecting Linguistic Indicators for Stereotype Assessment with Large Language Models

Abstract
Social categories and stereotypes embedded in language can introduce data bias into the training of Large Language Models (LLMs). Despite safeguards, these biases often persist in model behavior, potentially leading to representational harm in outputs. While sociolinguistic research provides valuable insights into the formation and spread of stereotypes, NLP approaches for bias evaluation rarely draw on this foundation and often lack objectivity, precision, and interpretability. To fill this gap, we propose a new approach to assess stereotypes by detecting and quantifying the linguistic indication of a stereotype. We derive linguistic indicators from the Social Category and Stereotype Communication (SCSC) framework indicating strong social category formulation and stereotyping in language, and use them to build a categorization scheme. We use in-context learning to instruct LLMs to examine the linguistic properties of a sentence containing stereotypes, providing a basis for a fine-grained stereotype assessment. We develop a scoring function to measure linguistic indicators of stereotypes based on empirical evaluation. Our annotations of stereotyped sentences reveal that these linguistic indicators explain the strength of a stereotype. The models perform well in detecting and classifying linguistic indicators used to denote a category, but sometimes struggle with accurately evaluating the described associations. The use of more few-shot examples significantly improves the performance. Model performance increases with size, as Llama-3.3-70B-Instruct and GPT-4 achieve comparable results that surpass those of Mixtral-8x7B-Instruct, GPT-4-mini and Llama-3.1-8B-Instruct_4bit. Code and annotations can be found in https://github.com/r-goerge/Detecting-Linguistic-Indicators-for-Stereotype-Assessment-with-LLMs.
Author(s)
Görge, Rebekka
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Mock, Michael  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Allende-Cid, Héctor  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Mainwork
FAccT 2025, ACM Conference on Fairness, Accountability, and Transparency. Proceedings  
Conference
Conference on Fairness, Accountability, and Transparency 2025  
Open Access
File(s)
Download (971.79 KB)
Rights
CC BY 4.0: Creative Commons Attribution
DOI
10.1145/3715275.3732181
10.24406/publica-5120
Additional link
Full text
Language
English
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Keyword(s)
  • Large language models

  • Fairness

  • Stereotype Detection

  • Linguistics

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024