MarIA and BETO are sexist: evaluating gender bias in large language models for Spanish
dc.contributor.author | Garrido-Muñoz, Ismael | |
dc.contributor.author | Martínez-Santiago, Fernando | |
dc.contributor.author | Montejo-Ráez, Arturo | |
dc.date.accessioned | 2024-02-02T13:32:08Z | |
dc.date.available | 2024-02-02T13:32:08Z | |
dc.date.issued | 2023-07-23 | |
dc.description.abstract | The study of bias in language models is a growing area of work, however, both research and resources are focused on English. In this paper, we make a first approach focusing on gender bias in some freely available Spanish language models trained using popular deep neural networks, like BERT or RoBERTa. Some of these models are known for achieving state-of-the-art results on downstream tasks. These promising results have promoted such models’ integration in many real-world applications and production environments, which could be detrimental to people affected for those systems. This work proposes an evaluation framework to identify gender bias in masked language models, with explainability in mind to ease the interpretation of the evaluation results. We have evaluated 20 different models for Spanish, including some of the most popular pretrained ones in the research community. Our findings state that varying levels of gender bias are present across these models.This approach compares the adjectives proposed by the model for a set of templates. We classify the given adjectives into understandable categories and compute two new metrics from model predictions, one based on the internal state (probability) and the other one on the external state (rank). Those metrics are used to reveal biased models according to the given categories and quantify the degree of bias of the models under study. | es_ES |
dc.description.sponsorship | Funding for open access publishing: Universidad de Jaén/CBUA. This work has been partially supported by WeLee project (1380939, FEDER Andalucía 2014-2020) funded by the Andalusian Regional Government, and projects CONSENSO (PID2021-122263OB-C21), MODERATES (TED2021-130145B-I00), SocialTOX (PDC2022-133146-C21) funded by Plan Nacional I+D+i from the Spanish Government, and project PRECOM (SUBV-00016) funded by the Ministry of Consumer Affairs of the Spanish Government. | es_ES |
dc.identifier.citation | Garrido-Muñoz, I., Martínez-Santiago, F. & Montejo-Ráez, A. MarIA and BETO are sexist: evaluating gender bias in large language models for Spanish. Lang Resources & Evaluation (2023). https://doi.org/10.1007/s10579-023-09670-3 | es_ES |
dc.identifier.issn | 1574-0218 | es_ES |
dc.identifier.other | 10.1007/s10579-023-09670-3 | es_ES |
dc.identifier.uri | https://link.springer.com/article/10.1007/s10579-023-09670-3 | es_ES |
dc.identifier.uri | https://hdl.handle.net/10953/1911 | |
dc.language.iso | eng | es_ES |
dc.publisher | Springer | es_ES |
dc.relation.ispartof | Language Resources and Evaluation | es_ES |
dc.rights | Atribución-NoComercial-SinDerivadas 3.0 España | * |
dc.rights.accessRights | info:eu-repo/semantics/openAccess | es_ES |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/es/ | * |
dc.subject | BERT | es_ES |
dc.subject | Bias evaluation | es_ES |
dc.subject | Deep learning | es_ES |
dc.subject | Gender bias | es_ES |
dc.subject | Language model | es_ES |
dc.subject | RoBERTa | es_ES |
dc.title | MarIA and BETO are sexist: evaluating gender bias in large language models for Spanish | es_ES |
dc.type | info:eu-repo/semantics/article | es_ES |
dc.type.version | info:eu-repo/semantics/publishedVersion | es_ES |
Archivos
Bloque original
1 - 1 de 1
Cargando...
- Nombre:
- 2023_MarIAandBETO.pdf
- Tamaño:
- 1.78 MB
- Formato:
- Adobe Portable Document Format
- Descripción:
- Published version
Bloque de licencias
1 - 1 de 1
No hay miniatura disponible
- Nombre:
- license.txt
- Tamaño:
- 1.98 KB
- Formato:
- Item-specific license agreed upon to submission
- Descripción: