M

Mistral NeMo Minitron 8B Base

Developed by nvidia
Mistral-NeMo-Minitron-8B-Base is a basic text generation model obtained by pruning and distilling Mistral-NeMo 12B, suitable for various natural language generation tasks.
Downloads 7,924
Release Time : 8/19/2024

Model Overview

This model is a large language model (LLM) obtained by pruning and distilling Mistral-NeMo 12B. It can effectively handle various text tasks and provide strong support for natural language processing.

Model Features

Pruning and distillation optimization
Obtained by pruning and distilling Mistral-NeMo 12B, reducing the model size while maintaining performance
Efficient training
After pruning, use 380 billion tokens for distillation-based continuous training to optimize model performance
Advanced architecture
Adopt advanced technologies such as Grouped Query Attention (GQA) and Rotary Position Embedding (RoPE)

Model Capabilities

Text generation
Paragraph completion
Language understanding
Code generation

Use Cases

Content generation
Paragraph completion
Complete the paragraph content according to the prompt
Generate coherent subsequent text
Education
Science knowledge Q&A
Answer questions in the scientific field
Provide accurate scientific knowledge explanations
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase