L

Llama 3.1 Minitron 4B Width Base

Developed by nvidia
Llama-3.1-Minitron-4B-Width-Base is a foundational text-to-text model obtained by pruning Llama-3.1-8B, suitable for various natural language generation tasks.
Downloads 10.15k
Release Time : 8/13/2024

Model Overview

This model was obtained by pruning Llama-3.1-8B, reducing the embedding size and MLP intermediate dimensions, and undergoing continuous distillation training with 94 billion tokens to produce the final model.

Model Features

Pruning Optimization
Optimized the model's embedding size and MLP intermediate dimensions by pruning Llama-3.1-8B, improving efficiency.
Distillation Training
Enhanced model performance through continuous distillation training with 94 billion tokens.
Commercial Use
The model is ready for commercial use and suitable for various natural language generation tasks.

Model Capabilities

Text Generation
Natural Language Understanding
Code Generation

Use Cases

Natural Language Processing
Text Completion
Used for automatically completing paragraphs or sentences to enhance writing efficiency.
Performs best within 8000 characters.
Question Answering System
Used to build question-answering systems that respond to user queries.
Performs well in zero-shot evaluations.
Code Generation
Code Completion
Used for automatically completing code snippets to improve development efficiency.
Scores 32.0 in the MBPP evaluation.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase