"The development of AI relies on a large global network of data workers who have been purposely made invisible in order to make systems appear more autonomous, even sentient. OpenAI flaunts that GPT, its LLM, is trained on millions of data points scraped from the internet, but hides the fact that workers must make this data suitable for training. Workers perform a wide set of tasks including cleaning, labeling, keywording, and semantic segmentation." (p. 6)