GNU social JP
  • FAQ
  • Login
GNU social JPは日本のGNU socialサーバーです。
Usage/ToS/admin/test/Pleroma FE
  • Public

    • Public
    • Network
    • Groups
    • Featured
    • Popular
    • People

On one hand, it’s legitimately striking that xAI could stand up a data center with 200,000 GPUs and train a decent model in less than two years. On the other hand, it would have been more striking had DeepSeek not accomplished a similar feat at a tiny fraction of the cost, and with worse technology. It’s too early to say with any certainty how good Grok’s models are. Leading researcher Andrej Karpathy finds that it passes many of the tests he throws at it, but it also fails almost as many. Even if you accept, though, that this is a state-of-the-art model, not a single person working in AI believes it will stay there for long. Leading AI labs push out new models every few days, and any innovations are almost all quickly copied and absorbed by their rivals.

Download link

https://files.mastodon.social/media_attachments/files/114/027/932/875/874/457/original/6a747cd7d7fcf661.png

Notices where this attachment appears

  1. Embed this notice
    Casey Newton (caseynewton@mastodon.social)'s status on Wednesday, 19-Feb-2025 10:23:32 JST Casey Newton Casey Newton

    Training Grok 3 took Elon 200,000 GPUs and untold billions, and it's ... decent at best? I wrote about AI's commodity problem

    https://www.platformer.news/too-many-ai-companies/

    In conversation about 5 months ago from mastodon.social permalink
  • Help
  • About
  • FAQ
  • TOS
  • Privacy
  • Source
  • Version
  • Contact

GNU social JP is a social network, courtesy of GNU social JP管理人. It runs on GNU social, version 2.0.2-dev, available under the GNU Affero General Public License.

Creative Commons Attribution 3.0 All GNU social JP content and data are available under the Creative Commons Attribution 3.0 license.