Tècnic de PLN a la Universitat de Barcelona, CLiC Group
Soc Físic Teòric especialitzat en Aprenentatge Automàtic i Intel·ligència Artificial. Actualment treballo com a Enginyer de Recerca en Processament del Llenguatge Natural al CLiC (UB). Soc un apassionat de la ciència i estic cursant una segon grau en lingüística general.
Sexism is generally defined as prejudice and discrimination based on sex or gender, affecting every sector of society, from social institutions to relationships and individual behavior. Social media platforms amplify the impact of sexism by conveying discriminatory content not only through text but also across multiple modalities, highlighting the critical need for a multimodal approach to the analysis of sexism online. With the rise of social media platforms where users share short videos, sexism is increasingly spreading through video content. Automatically detecting sexism in videos is a challenging task, as it requires analyzing the combination of verbal, audio, and visual elements to identify sexist content. In this study, (1) we introduce MuSeD, a new Multimodal Spanish dataset for Sexism Detection consisting of 11 hours of videos extracted from TikTok and BitChute; (2) we propose an innovative annotation framework for analyzing the contribution of textual and multimodal labels in the classification of sexist and non-sexist content; and (3) we evaluate a range of large language models (LLMs) and multimodal LLMs on the task of sexism detection. We find that visual information plays a key role in labeling sexist content for both humans and models. Models effectively detect explicit sexism; however, they struggle with implicit cases, such as stereotypes, instances where annotators also show low agreement. This highlights the inherent difficulty of the task, as identifying implicit sexism depends on the social and cultural context.
@article{de2025mused,title={MuSeD: A Multimodal Spanish Dataset for Sexism Detection in Social Media Videos},author={De Grazia, Laura and Pastells, Pol and Chas, Mauro V{\'a}zquez and Elliott, Desmond and Villegas, Danae S{\'a}nchez and Farr{\'u}s, Mireia and Taul{\'e}, Mariona},journal={arXiv preprint arXiv:2504.11169},year={2025},}
Interspeech 2025
SCRIBAL: A Digital Transcription Tool in Higher Education
Javier Román, Pol Pastells, Mauro Vázquez, Clara Puigventós, Montserrat Nofre, Mariona Taulé, and Mireia Farrús
SCRIBAL is a digital transcription and translation tool for university teaching, covering Catalan transcription and translation to the main foreign languages in class. Based on Whisper, SCRIBAL is fine-tuned for specific Catalan dialectal varieties and specialized academic terminology. It becomes an essential tool for accessibility, as well as for breaking language barriers and preserving the national languages in higher education.
@inproceedings{roman25_interspeech,title={SCRIBAL: A Digital Transcription Tool in Higher Education},author={Román, Javier and Pastells, Pol and Vázquez, Mauro and Puigventós, Clara and Nofre, Montserrat and Taulé, Mariona and Farrús, Mireia},year={2025},booktitle={Interspeech 2025},pages={4958--4959},issn={2958-1796},}
LREC-COLING 2024
Human vs. Machine Perceptions on Immigration Stereotypes
Wolfgang S. Schmeisser-Nieto, Pol Pastells, Simona Frenda, and Mariona Taule
In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024) , May 2024
The increasing popularity of natural language processing has led to a race to improve machine learning models that often leaves aside the core study object, the language itself. In this study, we present classification models designed to detect stereotypes related to immigrants, along with both quantitative and qualitative analyses, shedding light on linguistic distinctions in how humans and various models perceive stereotypes. Given the subjective nature of this task, one of the models incorporates the judgments of all annotators by utilizing soft labels. Through a comparative analysis of BERT-based models using both hard and soft labels, along with predictions from GPT-4, we gain a clearer understanding of the linguistic challenges posed by texts containing stereotypes. Our dataset comprises Spanish Twitter posts collected as responses to immigrant-related hoaxes, annotated with binary values indicating the presence of stereotypes, implicitness, and the requirement for conversational context to understand the stereotype. Our findings suggest that both model prediction confidence and inter-annotator agreement are higher for explicit stereotypes, while stereotypes conveyed through irony and other figures of speech prove more challenging to detect than other implicit stereotypes.
@inproceedings{schmeisser-nieto-etal-2024-human-vs,title={Human vs. Machine Perceptions on Immigration Stereotypes},author={Schmeisser-Nieto, Wolfgang S. and Pastells, Pol and Frenda, Simona and Taule, Mariona},year={2024},month=may,booktitle={
Proceedings of the 2024 Joint International Conference on Computational
Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
},publisher={ELRA and ICCL},address={Torino, Italia},pages={8453--8463},url={https://aclanthology.org/2024.lrec-main.741},editor={Calzolari, Nicoletta and Kan, Min-Yen and Hoste, Veronique and Lenci, Alessandro and Sakti, Sakriani and Xue, Nianwen},}