“A grad student who fell asleep in 1982 and woke up in 2022 might see large language models as a triumph for cultural theory.” My contribution to the debate this week in the CI blog. #LLM #Foucault https://critinq.wordpress.com/2023/06/29/the-empirical-triumph-of-theory/
Conversation
Notices
-
Embed this notice
Ted Underwood (tedunderwood@sigmoid.social)'s status on Friday, 30-Jun-2023 21:19:37 JST Ted Underwood -
Embed this notice
Ted Underwood (tedunderwood@sigmoid.social)'s status on Saturday, 01-Jul-2023 21:30:57 JST Ted Underwood @UlrichJunker In that piece I think I’m talking about instruction tuning rather than RLHF as such? It was an earlier advance, although the purposes are similar. But I agree with you that this whole topic is under-discussed. One way to put it is that the models responded to / addressed the Stochastic Parrots critique that they weren’t grounded in a communicative situation. But it served no one’s polemical purpose to take note of that.
In conversation permalink -
Embed this notice
Ulrich Junker (ulrichjunker@fediscience.org)'s status on Saturday, 01-Jul-2023 21:30:58 JST Ulrich Junker @TedUnderwood you are referring to #RLHF (reinforcement learning by human feedback) as a way of correcting transformer output by human authors. But this technique also covers learning preferences from humans and this aspect hasn’t found much attention in the debate of #LLMs, but may rather be determining for ChatGPT’s success. What is your opinion about this? https://proceedings.neurips.cc//paper_files/paper/2022/hash/b1efde53be364a73914f58805a001731-Abstract-Conference.html
In conversation permalink Attachments
-
Embed this notice