"

Exploitative Labour to “Train” Datasets

Datasets are “trained” by exploitative, racialized labour practices, hiring workers in the global south to view traumatic images and flagged content for racist, sexist, and offensive content, for example.

Training AI datasets also has internal processes that perpetuate gendered harms. In their Excavating AI project, Kate Crawford and Trevor Paglen revealed, “We find an implicit assumption here: only ‘male’ and ‘female’ bodies are ‘natural’.”

 

Read more: 

Harms: Ableist; Gendered; Genocidal; Racist; Socioeconomic

License

Icon for the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License

Harm Considerations of Large Language Models (LLMs) Copyright © by Teaching and Learning, University Libraries is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License, except where otherwise noted.