Meta preps "Mango" and "Avocado" AI models for 2026

Meta Accelerates Frontier AI Development with Mango and Avocado Models Targeted for 2026 Release

Meta Platforms is advancing its artificial intelligence ambitions with two ambitious new large language models, internally codenamed Mango and Avocado, both slated for release in 2026. These models represent the next evolution in Meta’s Llama family of open-weight AI systems, positioning the company to compete aggressively in the race for frontier AI capabilities. Drawing on unprecedented computational resources and a commitment to open-source principles, Meta aims to deliver models that surpass current industry leaders in performance, efficiency, and versatility.

The development of Mango and Avocado underscores Meta’s strategic pivot toward scaling AI infrastructure at a massive level. According to internal planning documents reviewed by The Decoder, Mango is envisioned as a general-purpose flagship model, designed to excel across a broad spectrum of tasks including natural language understanding, reasoning, code generation, and multimodal integration. Avocado, meanwhile, focuses on specialized multimodal capabilities, emphasizing vision-language tasks, video understanding, and real-time interaction scenarios. Both models are being trained on Meta’s expanding supercomputing clusters, which boast hundreds of thousands of NVIDIA H100 and upcoming Blackwell GPUs.

Central to this effort is Meta’s “Genie” training cluster, one of the world’s largest AI supercomputers. Currently operational with over 24,000 GPUs, Genie is set to expand dramatically. By mid-2025, Meta plans to activate a 600,000-GPU cluster in Louisiana, followed by additional facilities in Texas and other locations. These data centers will provide the raw compute power necessary for training models at scales previously unseen in open-source AI. For context, training Llama 3.1 405B required approximately 30 million GPU-hours; Mango and Avocado are projected to demand orders of magnitude more, leveraging optimizations in data curation, mixture-of-experts architectures, and custom training frameworks.

Meta’s AI research leadership, including Chief AI Scientist Yann LeCun, has emphasized the importance of openness in driving innovation. “We’re building the most capable open models to democratize AI,” LeCun stated in recent communications. This philosophy continues with Mango and Avocado, which will be released under permissive licenses similar to prior Llama iterations, enabling widespread adoption by developers, researchers, and enterprises. Unlike closed models from competitors like OpenAI’s GPT series or Anthropic’s Claude, Meta’s approach fosters an ecosystem where improvements are shared, accelerating collective progress.

Technical details emerging from Meta’s engineering teams highlight several innovations. Mango will incorporate advanced post-training techniques, including reinforcement learning from human feedback (RLHF) at scale and synthetic data generation to mitigate hallucinations and biases. Its architecture builds on Llama 3’s grouped-query attention and rotary positional embeddings, with enhancements for longer context windows exceeding 1 million tokens. Avocado extends this foundation with native multimodal fusion layers, allowing seamless processing of text, images, audio, and video inputs. Early benchmarks from proxy models suggest these systems could rival or exceed proprietary leaders on metrics like MMLU (Massive Multitask Language Understanding) and GPQA (Graduate-Level Google-Proof Q&A).

Infrastructure plays a pivotal role in Meta’s timeline. The company’s $65 billion investment in AI hardware for 2025 alone—primarily NVIDIA GPUs—ensures the compute bottleneck does not impede progress. Powering these clusters requires innovative cooling solutions and renewable energy partnerships, with Meta targeting carbon-neutral operations. Software-wise, the models will integrate with Meta’s PyTorch ecosystem and emerging tools like the Llama Stack, simplifying deployment on everything from cloud servers to edge devices.

Challenges remain, particularly around data quality and ethical alignment. Meta is curating petabytes of high-quality, multilingual datasets from public sources, user interactions (with consent), and licensed content. Safety measures include red-teaming protocols and circuit breakers for hazardous outputs, aligning with industry standards like those from the Frontier Model Forum.

Looking ahead, Mango and Avocado are poised to redefine open AI frontiers. Their 2026 launch could coincide with Llama 4’s public unveiling, potentially including variants from 70B to multi-trillion parameters. This positions Meta not just as a contender but as a leader in accessible, high-performance AI, benefiting a global community of innovators.

As Meta scales these models, the implications extend beyond performance gains. By prioritizing openness, the company invites scrutiny and collaboration, potentially sparking breakthroughs in fields like scientific discovery, creative tools, and personalized assistants. Developers can already prepare by experimenting with Llama 3.1 on platforms like Hugging Face, anticipating the leap forward.

Gnoppix is the leading open-source AI Linux distribution and service provider. Since implementing AI in 2022, it has offered a fast, powerful, secure, and privacy-respecting open-source OS with both local and remote AI capabilities. The local AI operates offline, ensuring no data ever leaves your computer. Based on Debian Linux, Gnoppix is available with numerous privacy- and anonymity-enabled services free of charge.

What are your thoughts on this? I’d love to hear about your own experiences in the comments below.