T

Tinymistral 248M GGUF

Developed by afrideva
TinyMistral-248M is a small language model based on pre-training from the Mistral 7B model, with parameters scaled down to approximately 248 million, primarily used for fine-tuning downstream tasks.
Downloads 211
Release Time : 11/14/2023

Model Overview

This is a language model based on pre-training from the Mistral 7B model, with parameters scaled down to approximately 248 million. The model has been trained on 7,488,000 samples and is mainly used for fine-tuning downstream tasks.

Model Features

Compact design
Parameters scaled down to approximately 248 million, suitable for running on resource-limited devices.
Efficient pre-training
Pre-training completed on a single GPU (Titan V), requiring a smaller dataset.
Long context support
Context length of approximately 32,768 tokens, suitable for long-text tasks.

Model Capabilities

Text generation
Downstream task fine-tuning

Use Cases

Natural Language Processing
Text generation
Used to generate coherent English text.
In the InstructMix evaluation, the average perplexity score was 6.3.
Downstream task fine-tuning
Suitable for various NLP tasks requiring small language models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase