Y

Yarn Mistral 7B 128k AWQ

Developed by TheBloke
Yarn Mistral 7B 128K is an advanced language model optimized for long-context processing, further pre-trained on long-context data using the YaRN extension method, supporting a 128k token context window.
Downloads 483
Release Time : 11/2/2023

Model Overview

A language model based on Mistral-7B-v0.1, specifically optimized for handling long-context scenarios, suitable for various natural language processing tasks involving ultra-long texts.

Model Features

Ultra-long context support
Supports a 128k token context window, capable of processing ultra-long text content.
Efficient quantization
Provides AWQ-quantized versions to improve inference efficiency while maintaining quality.
Optimized pre-training
Undergoes 1500 additional pre-training steps on long-context data using the YaRN method.

Model Capabilities

Long text generation
Context understanding
Text continuation
Question answering systems

Use Cases

Document processing
Long document summarization
Summarizes and extracts key information from ultra-long documents.
Legal document analysis
Processes and analyzes complex legal contracts and clauses.
Code processing
Codebase analysis
Understands the structure and functionality of large codebases.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase