Colossus 2 Data Center: Seven Models Training in Parallel
On April 8, 2026, Elon Musk announced that SpaceXAI Colossus 2 (xAI’s new 1-gigawatt-scale supercluster) is now simultaneously training seven distinct models:
- Grok Imagine V2 (next-generation image and video generation)
- 2 variants of 1T-parameter Grok models
- 2 variants of 1.5T-parameter Grok models
- 1 × 6T-parameter Grok model
- 1 × 10T-parameter Grok model
This appears to be a new hyper-scaling in action: xAI is running an entire portfolio of frontier-scale models at once on what is currently the world’s most powerful AI training cluster. The goal is explicit: to close the gap with (and eventually surpass) the very best systems from Anthropic and OpenAI.
Current Baseline: Grok 4.20 at 500B Parameters
For context, the model powering Grok today, Grok 4.20, has a total of approximately 500 billion parameters (0.5T).
Elon Musk put this in sharp perspective just days ago:
- Grok 4.20 is half the size of Anthropic’s Claude Sonnet (≈ 1T parameters).
- It is one-tenth the size of Anthropic’s Claude Opus (≈ 5T parameters).
Despite its relatively “small” size, Grok 4.20 is already delivering extremely strong real-world performance, recently topping the Chatbot Arena in Legal & Government categories and outperforming Opus 4.6 in several specialized benchmarks. It is, in Musk’s words, “a very strong model for its size.”
The Hyper-Scaling Leaping:
From 0.5T → 1T → 1.5T → 6T → 10T
What Colossus 2 represents is a deliberate, aggressive expansion phase: Six-trillion and ten-trillion-parameter models are now actively training. These models will "step into the boxing ring" with current frontier models. Their emergent properties could expand, and depending on their training, there might be heightened performance in specialized areas.
This hyper-scaling approach is strategic experimentation. At the absolute frontier, we are still doing AI research on how scaling affects reasoning, generalization, creativity, and agentic behavior at a high level. Training across the full spectrum of model sizes simultaneously will allow a lot of research to be collected on these new larger models for SpaceXAI.
Hyper-Scaling as Promised
Colossus 2 Data Center is an impressive crown jewel for Musk, and after the long wait from the build out, we finally have word about the advanced simultaneous training happening at the new facility. Musk also revealed that the first training step for the largest model, the 10T, takes about 2 months to complete the pre-training. In regards to the current Grok model, the jump from Grok 4.20’s 500B parameters to Grok 5's 10T is a 20× increase.
Colossus 2 training these seven models at once is the clearest signal yet that the frontier labs have fully embraced a hyper-scaling reality. OpenAI, Anthropic, and xAI have not and are not waiting for one model to finish before starting to train the next. They are now building entire next generations in parallel. This article is augmented by AI.

No comments:
Post a Comment