Meta Bets Big On Mango And Avocado As The AI Image War Heats Up
The race to own the most-used AI images and videos is pulling Meta back into the centre of the fight.
After months of questions around its AI direction, Mark Zuckerberg is now steering the company toward a more closed, more competitive path, anchored by two new models designed to go head-to-head with Google and OpenAI.
At the heart of the plan are Mango, an image and video model, and Avocado, Meta’s next-generation text model.
Both are expected to launch in the first half of 2026, according to details shared internally by Chief AI Officer Alexandr Wang during a companywide Q&A with Chief Product Officer Chris Cox.
A Reset After Llama And A Shift Away From Open Models
Meta’s strategy marks a clear break from its Llama open-source lineage.
Internally, Llama 4 has been viewed as a disappointment, prompting leadership to rethink whether openness still offers an edge as rivals push faster and more polished systems into consumer apps.
Mango and Avocado are positioned as proprietary models, built to compete directly with Google’s Gemini line and OpenAI’s expanding image tools.
Mango is expected to focus on high-quality image and video generation, while Avocado is designed as a frontier text model with stronger reasoning and coding skills, areas where Meta has lagged in the past.
Inside Meta Superintelligence Labs
The new models are the first major outputs from Meta Superintelligence Labs, a division created during a major restructuring over the summer.
Zuckerberg personally recruited Alexandr Wang, founder of Scale AI, to lead the unit, following Meta’s $14 billion investment in Scale that brought key data and talent in-house.
Alexandr Wang is the founder of Scale AI and the current Chief AI Officer at Meta, recognised as the world’s youngest self-made billionaire for building the data infrastructure that powers modern artificial intelligence.
Since then, Meta has hired more than 20 researchers from OpenAI and assembled a team of over 50 specialists with deep experience in large models and generative media.
The focus is deliberate: image and video generation has become one of the most competitive battlegrounds in AI.
During the internal session, Wang also revealed that Meta has begun early work on world models, AI systems that learn by observing and understanding visual environments rather than just predicting text.
The effort signals a longer-term ambition to move beyond chat-based systems into models that can reason about the physical world.
Image Generation Becomes The Stickiest Feature
Meta’s push comes as rivals double down on visual AI.
In September, Meta released Vibes, a short-form video generator built with Midjourney.
Days later, OpenAI launched Sora, showing how quickly each player now reacts to the other.
Google had already raised the pressure earlier in the year with Nano Banana, a move that helped boost Gemini’s monthly users from 450 million in July to over 650 million by late October.
The competition intensified again in November when Google rolled out Gemini’s third generation.
Inside OpenAI, executives reportedly responded with a code red to reclaim top benchmark scores.
Soon after, the company released an updated version of ChatGPT Images.
Speaking to journalists later, Sam Altman said image creation is now one of the main reasons users keep coming back, calling it a sticky feature.
Google Pushes AI Into The Mass Market
Google is not slowing down.
On Wednesday, it announced Gemini 3 Flash, a faster and cheaper model designed for broad use.
While smaller than Gemini 3 Pro, it carries many of the same reasoning abilities and is aimed squarely at everyday apps rather than premium tiers.
Alphabet CEO Sundar Pichai said,
“With this release, Gemini 3’s next-generation intelligence is now rolling out to everyone across our products including Gemini app + AI Mode in Search. Devs can build with it in the Gemini API, Google AI Studio, Gemini CLI, and Google Antigravity and enterprises can get it in Vertex AI and Gemini Enterprise.”
With scale becoming essential, the strategy of keeping powerful tools behind enterprise paywalls may no longer succeed.
Internal Tension And High Stakes At Meta
The shift to closed models has not been seamless.
Reports of internal friction have emerged as teams move away from Llama and reallocate resources toward Avocado.
Some engineers see the pivot as necessary to stay competitive, while others worry about losing the goodwill and momentum built through open development.
Meta’s spending reflects the stakes.
Billions are being poured into compute, data and hiring, with Wang’s leadership now under close watch.
Avocado, in particular, is widely seen inside the company as a make-or-break test of whether Meta can truly match the best models on the market.
Can Meta Win The Image Arms Race?
Meta’s return to proprietary AI is a calculated risk, not a guaranteed comeback.
Mango enters a market where Google and OpenAI already move at speed, with massive user bases and tightly integrated products.
Avocado faces even tougher odds in text and reasoning, where benchmarks shift quickly and loyalty is thin.
From Coinlive’s perspective, Meta’s biggest challenge may not be talent or funding, but timing.
By 2026, image and video AI could already be commoditised, with success depending less on raw quality and more on distribution, cost and trust.
Mango and Avocado may be powerful, but survival in this market will hinge on whether Meta can turn technical strength into daily habit, not just headlines.