SESGO: Spanish Evaluation of Stereotypical Generative Outputs

dc.contributor.authorMelissa Robles
dc.contributor.authorC. Bernal Bellido
dc.contributor.authorDenniss Raigoso
dc.contributor.authorMateo Dulce Rubio
dc.coverage.spatialBolivia
dc.date.accessioned2026-03-22T19:47:02Z
dc.date.available2026-03-22T19:47:02Z
dc.date.issued2025
dc.description.abstractThis paper addresses the critical gap in evaluating bias in multilingual Large Language Models (LLMs), with a specific focus on Spanish language within culturally-aware Latin American contexts. Despite widespread global deployment, current evaluations remain predominantly US-English-centric, leaving potential harms in other linguistic and cultural contexts largely underexamined. We introduce a novel, culturally-grounded framework for detecting social biases in instruction-tuned LLMs. Our approach adapts the underspecified question methodology from the BBQ dataset by incorporating culturally-specific expressions and sayings that encode regional stereotypes across four social categories: gender, race, socioeconomic class, and national origin. Using more than 4,000 prompts, we propose a new metric that combines accuracy with the direction of error to effectively balance model performance and bias alignment in both ambiguous and disambiguated contexts. To our knowledge, our work presents the first systematic evaluation examining how leading commercial LLMs respond to culturally specific bias in the Spanish language, revealing varying patterns of bias manifestation across state-of-the-art models. We also contribute evidence that bias mitigation techniques optimized for English do not effectively transfer to Spanish tasks, and that bias patterns remain largely consistent across different sampling temperatures. Our modular framework offers a natural extension to new stereotypes, bias categories, or languages and cultural contexts, representing a significant step toward more equitable and culturally-aware evaluation of AI systems in the diverse linguistic environments where they operate.
dc.identifier.doi10.1609/aies.v8i3.36707
dc.identifier.urihttps://doi.org/10.1609/aies.v8i3.36707
dc.identifier.urihttps://andeanlibrary.org/handle/123456789/78094
dc.language.isoen
dc.relation.ispartofProceedings of the AAAI/ACM Conference on AI Ethics and Society
dc.sourceUniversidad de Los Andes
dc.subjectComputer science
dc.subjectConverse
dc.subjectFocus (optics)
dc.subjectArtificial intelligence
dc.subjectNatural language processing
dc.subjectGenerative grammar
dc.subjectCognitive psychology
dc.subjectCultural bias
dc.subjectSocioeconomic status
dc.subjectLinguistics
dc.titleSESGO: Spanish Evaluation of Stereotypical Generative Outputs
dc.typearticle

Files