L

Llama 4 Maverick 17B 128E Instruct FP8

Developed by RedHatAI
A native multi-modal AI model in the Llama 4 series, supporting text and image understanding, adopting a mixture-of-experts architecture, suitable for commercial and research scenarios.
Downloads 5,679
Release Time : 5/12/2025

Model Overview

Llama 4 Maverick is a multi-modal model with 17 billion parameters and 128 experts, supporting text and image input and output, and optimizing visual recognition and reasoning capabilities.

Model Features

Native multi-modal
Supports multi-modal input and output of text and images, enabling cross-modal understanding and generation.
Mixture-of-Experts architecture
Adopts a 128-expert MoE architecture, achieving a total parameter scale of 400B while maintaining 17B activated parameters.
Long context support
Llama 4 Maverick supports a context length of 1M tokens, suitable for processing long documents and complex tasks.
Multi-language optimization
Specifically optimized for 12 languages, with the potential to support 200 languages.

Model Capabilities

Text generation
Image understanding
Visual reasoning
Multi-language processing
Code generation
Long document processing

Use Cases

Business assistant
Intelligent customer service
Handles multi-language customer inquiries, understanding text and image content.
Achieved an accuracy of 80.5% in the MMLU Pro benchmark test
Education and research
Visual question-answering system
Answers questions about charts and document images.
Achieved a score of 94.4 ANLS on the DocVQA test set
Content creation
Multi-modal content generation
Generates descriptive text based on images or relevant image analysis based on text.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase