M

Meta Llama 3.1 70B FP8

Developed by RedHatAI
FP8 quantized version of Meta-Llama-3.1-70B, suitable for multilingual business and research applications, with both weights and activations quantized to FP8 format, reducing storage and memory requirements by approximately 50%.
Downloads 191
Release Time : 7/31/2024

Model Overview

This model is a quantized version of Meta-Llama-3.1-70B, optimized by quantizing weights and activations to FP8 format, suitable for text generation tasks and supporting multiple languages.

Model Features

FP8 quantization
Both weights and activations are quantized to FP8 format, reducing storage and memory requirements by approximately 50%.
Multilingual support
Supports text generation in multiple languages including English, German, French, and more.
High performance
Achieves an average score of 79.70 on the OpenLLM benchmark, close to the performance of the non-quantized model.

Model Capabilities

Text generation
Multilingual support
Business applications
Research purposes

Use Cases

Business applications
Multilingual customer service
Used for generating multilingual customer service responses, supporting instant replies in multiple languages.
Research purposes
Language model research
Used for studying the performance and behavior of large-scale language models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase