Meta launches first multimodal models from Llama 4 herd

Meta has launched Llama 4 Scout and Llama 4 Maverick, the first models of its Llama herd providing advanced multimodal capabilities.

The models integrate native multimodality, which refers to the ability of an AI model to process and integrate multiple types of data (modalities) such as text, images, audio, and video, within a single framework. The models are inherently designed to handle and combine different forms of data more easily, rather than relying on separate models for each modality.

The firm said the two new models provide better performances on cost across a broad range of widely reported benchmarks than Gemma 3, Gemini 2.0 Flash-Lite, Mistral 3.1, GPT-4o, and comparable results to the new DeepSeek v3 on reasoning and coding and better performances on cost.

Llama 4 Maverick seeks to offer improved performance in image and text understanding, enabling the creation of more sophisticated AI applications.

Llama 4 Scout provides better performance for long-context tasks like multi-document summarisation and reasoning over large codebases, providing understanding both text and images.

“We’re also previewing Llama 4 Behemoth, one of the smartest LLMs in the world and our most powerful yet to serve as a teacher for our new models,”said the company.

The Llama collection aims to provide improved intelligent systems capable of taking generalised actions, conversing naturally with humans, and working through complex problems, with multimodal AI capabilities allowing them to interact more naturally with users, understanding and responding to both visual and auditory cues.

“These models represent the best of Llama, offering multimodal intelligence at a compelling price while outperforming models of significantly larger sizes,” the firm continued.

Llama 4 Scout and Llama 4 Maverick models can be downloaded on Meta’s website, with users able to try Meta AI built with Llama 4 in WhatsApp, Messenger, Instagram Direct.

“As more people continue to use artificial intelligence to enhance their daily lives, it’s important that the leading models and systems are openly available so everyone can build the future of personalised experiences,” Meta added.

Earlier this month, Meta Platforms' head of artificial intelligence research, Joelle Pineau, announced her departure from the company after eight years, creating a significant vacancy as the tech giant aggressively invests in the AI space.

Pineau, who has served as vice president of AI research and led Meta's Fundamental AI Research group (FAIR) since 2023, announced her plans to leave in a LinkedIn post on Tuesday. Her last day will be 30 May.



Share Story:

Recent Stories


The future-ready CFO: Driving strategic growth and innovation
This National Technology News webinar sponsored by Sage will explore how CFOs can leverage their unique blend of financial acumen, technological savvy, and strategic mindset to foster cross-functional collaboration and shape overall company direction. Attendees will gain insights into breaking down operational silos, aligning goals across departments like IT, operations, HR, and marketing, and utilising technology to enable real-time data sharing and visibility.

The corporate roadmap to payment excellence: Keeping pace with emerging trends to maximise growth opportunities
In today's rapidly evolving finance and accounting landscape, one of the biggest challenges organisations face is attracting and retaining top talent. As automation and AI revolutionise the profession, finance teams require new skillsets centred on analysis, collaboration, and strategic thinking to drive sustainable competitive advantage.