and what's ADDITIONALLY infuriating is some engineer or product team at openai (or whatever) is going to read this paper and think they can "fix" the problem by applying human feedback alignment blalala to this particular situation (or even this particular corpus!), instead of recognizing that there are an infinite number of ways (both overt and subtle) that language can enact prejudice, and the system they've made necessarily amplifies that prejudice