Abstract
Data filtering strategies are a crucial component to develop safe Large Language Models (LLM), since they support the removal of harmful contents from pretraining datasets. There is a lack of research on the actual impact of these strategies on vulnerable groups to discrimination, though, and their effectiveness has not been yet systematically addressed. In this paper we present a benchmark study of data filtering strategies for harm reduction aimed at providing a systematic evaluation on these approaches. We provide an overview 55 technical reports of English LMs and LLMs to identify the existing filtering strategies in literature and implement an experimental setting to test their impact against vulnerable groups. Our results show that the positive impact that strategies have in reducing harmful contents from documents has the side effect of increasing the underrepresentation of vulnerable groups to discrimination in datasets.
| Originalsprog | Engelsk |
|---|---|
| Titel | Proceedings of the AAAI Conference on Artificial Intelligence : AAAI Special Track on AI for Social Impact II |
| Antal sider | 11 |
| Vol/bind | 40 |
| Forlag | AAAI Press |
| Publikationsdato | 2026 |
| Udgave | 46 |
| Sider | 39303-39313 |
| ISBN (Elektronisk) | 978-1-57735-906-7 |
| DOI | |
| Status | Udgivet - 2026 |
| Begivenhed | AAAI Conference on Artificial Intelligence - Singapore EXPO, Singapore Varighed: 20 jan. 2026 → 27 jan. 2026 Konferencens nummer: 40 |
Konference
| Konference | AAAI Conference on Artificial Intelligence |
|---|---|
| Nummer | 40 |
| Lokation | Singapore EXPO |
| Land/Område | Singapore |
| Periode | 20/01/2026 → 27/01/2026 |
Citationsformater
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver