The $14.3 Billion Admission That Llama Was Not Enough
When Meta released Llama 4 in April 2025 to a lukewarm reception, Mark Zuckerberg did not opt for incremental improvements. Instead, he made one of the most expensive talent acquisitions in AI history: a $14.3 billion investment in Scale AI for a 49% stake, primarily to recruit co-founder Alexandr Wang as Meta's Chief AI Officer. Wang was tasked with building Meta Superintelligence Labs from the ground up -- not iterating on the existing Llama infrastructure, but starting over entirely.
The decision to rebuild rather than refine tells us something important about what went wrong with Llama. According to Wang's own account, MSL created new infrastructure, new architecture, and new data pipelines. The result -- Muse Spark matching Llama 4 Maverick capabilities with over 10x less compute -- suggests that Meta's previous approach was not just underperforming on benchmarks but was fundamentally inefficient at a deep technical level. The nine-month timeline from lab formation to model launch is aggressive by any standard and signals that the bottleneck was not raw compute or data (Meta has plenty of both) but rather the engineering and architectural decisions underlying the Llama lineage.
This is also a story about organizational design. Rather than reform FAIR or the existing Llama team, Zuckerberg created an entirely separate lab with a mandate to compete at the frontier. The implicit message to Meta's AI research community is stark: the open-source-first approach that defined Meta's AI identity since 2023 was not delivering results fast enough for a company spending tens of billions annually on AI infrastructure.
