This consistent trend of models getting smaller, faster and more capable on the same hardware is one of the reasons I'm not particularly concerned by the ongoing discourse about models hitting a plateau
https://simonwillison.net/2024/Dec/9/llama-33-70b/#is-performance-about-to-plateau-
Conversation
Notices
-
Embed this notice
Simon Willison (simon@fedi.simonwillison.net)'s status on Tuesday, 10-Dec-2024 04:53:10 JST Simon Willison -
Embed this notice
Simon Willison (simon@fedi.simonwillison.net)'s status on Tuesday, 10-Dec-2024 04:53:13 JST Simon Willison I can now run a GPT-4 class model on my laptop
(The exact same laptop that could just about run a GPT-3 class model 20 months ago)
The new Llama 3.3 70B is a striking example of the huge efficiency gains we've seen in the last two years
https://simonwillison.net/2024/Dec/9/llama-33-70b/In conversation permalink Attachments
Tim Chambers repeated this.
-
Embed this notice