On one hand, it’s legitimately striking that xAI could stand up a data center with 200,000 GPUs and train a decent model in less than two years. On the other hand, it would have been more striking had DeepSeek not accomplished a similar feat at a tiny fraction of the cost, and with worse technology. It’s too early to say with any certainty how good Grok’s models are. Leading researcher Andrej Karpathy finds that it passes many of the tests he throws at it, but it also fails almost as many. Even if you accept, though, that this is a state-of-the-art model, not a single person working in AI believes it will stay there for long. Leading AI labs push out new models every few days, and any innovations are almost all quickly copied and absorbed by their rivals.
https://files.mastodon.social/media_attachments/files/114/027/932/875/874/457/original/6a747cd7d7fcf661.png