# Long context support

Qwen3 Embedding 4B GGUF
Apache-2.0
Qwen3-Embedding-4B is a text embedding model built on the Qwen3 series, designed specifically for text embedding and ranking tasks, and performs excellently in multilingual text processing and code retrieval.
Text Embedding
Q
Mungert
723
1
Medgemma 4b It GGUF
Other
MedGemma-4B-IT is a medical multimodal model based on Gemma 3, supporting the understanding of medical text and images, and suitable for the development of medical AI applications.
Image-to-Text Transformers
M
Mungert
637
2
Medgemma 4b Pt
Other
MedGemma is a medical multimodal model developed based on Gemma 3, focusing on medical text and image understanding and supporting the construction of healthcare AI applications.
Image-to-Text Transformers
M
unsloth
105
1
Medgemma 27b Text It
Other
MedGemma is a medical-specific model series optimized based on Gemma 3, offering a 27-billion-parameter pure-text version specifically optimized for medical text understanding and reasoning.
Large Language Model Transformers
M
unsloth
211
2
Qwen2 96M
Apache-2.0
Qwen2-96M is a miniature language model based on the Qwen2 architecture, containing 96 million parameters and supporting a context length of 8192 tokens, suitable for English text generation tasks.
Large Language Model English
Q
Felladrin
76
2
Qwen2.5 1.5B Instruct
Apache-2.0
A 1.5B parameter instruction fine-tuned model designed for Gensyn RL Swarm, supporting local fine-tuning via peer-to-peer reinforcement learning
Large Language Model Transformers English
Q
Gensyn
2.1M
4
ZYH LLM Qwen2.5 14B V4
Apache-2.0
ZYH-LLM-Qwen2.5-14B-V4 is a large language model improved based on Qwen2.5-14B. It enhances computational accuracy and reasoning ability through multi-stage model merging and distillation techniques.
Large Language Model Safetensors Supports Multiple Languages
Z
YOYO-AI
1,235
7
Jp ModernBERT Large Preview
MIT
Japanese BERT model trained by Algomatic team, supporting fill-mask tasks with a context length of up to 8192.
Large Language Model Japanese
J
makiart
20
1
Ganga 2 1B
Apache-2.0
Ganga-2-1b is an instruction fine-tuned model trained on a Hindi dataset and is part of Project Unity, aiming to handle the diversity and richness of Indian languages.
Large Language Model Transformers Supports Multiple Languages
G
LingoIITGN
598
3
Llama3.1 Typhoon2 8b Instruct
Llama3.1-Typhoon2-8B is a large Thai language model (instructional type) based on the Transformer architecture, capable of handling various language tasks and providing users with efficient and accurate language interaction services.
Large Language Model Safetensors
L
scb10x
2,831
7
Llama DNA 1.0 8B Instruct
A state-of-the-art bilingual language model based on the Llama architecture, specially optimized for Korean understanding and generation while maintaining strong English capabilities.
Large Language Model Transformers Supports Multiple Languages
L
dnotitia
661
58
Granite 8b Code Instruct 128k GGUF
Apache-2.0
IBM Granite 8B code instruction model, supporting a context length of 128k, focusing on code generation and instruction understanding tasks.
Large Language Model Transformers Other
G
tensorblock
186
1
Polish Reranker Bge V2
This is a reranking model based on BAAI/bge-reranker-v2-m3 and further fine-tuned on a large-scale Polish text pair dataset, supporting long-context processing.
Text Embedding Transformers Other
P
sdadas
549
1
Jais Family 1p3b
Apache-2.0
The Jais series is a bilingual large language model specialized in Arabic language processing, with strong English capabilities and 1.3 billion parameters.
Large Language Model Safetensors Supports Multiple Languages
J
inceptionai
318
9
H2o Danube3 4b Chat
Apache-2.0
A chat model fine-tuned by H2O.ai with 4 billion parameters, supporting text generation tasks and capable of offline operation on mobile phones.
Large Language Model Transformers English
H
h2oai
41.49k
67
Norwai Mixtral 8x7B Instruct
A large Norwegian language model fine-tuned on instructions based on NorwAI-Mixtral-8x7B, optimized using approximately 9000 high-quality Norwegian instructions.
Large Language Model Transformers
N
NorwAI
144
2
Gemma
Gemma is an advanced open-source model trained on high-quality datasets, supporting different context length requirements.
Large Language Model
G
cortexso
295
1
Phi 3 Small 8k Instruct Onnx Cuda
MIT
Phi-3 Small is a 7B-parameter lightweight cutting-edge open-source model, optimized for NVIDIA GPUs in ONNX format, supporting 8K context length with strong inference capabilities.
Large Language Model Transformers
P
microsoft
115
12
Deepseek V2 Chat GGUF
MIT
The GGUF quantized version of DeepSeek-V2-Chat, suitable for local deployment and operation.
Large Language Model Supports Multiple Languages
D
leafspark
1,388
27
H2o Danube2 1.8b Chat
Apache-2.0
An 1.8B parameter chat model fine-tuned by H2O.ai, based on adapted Llama 2 architecture with 8192 context length support
Large Language Model Transformers English
H
h2oai
948
61
Ruropebert E5 Base 2k
A Russian sentence encoder model based on the RoPEBert architecture, supporting a context length of 2048 tokens and excelling in the encodechka benchmark tests.
Text Embedding Transformers Other
R
Tochka-AI
2,422
11
Phogpt 4B Chat
Bsd-3-clause
PhoGPT is an open-source series of 4-billion-parameter Vietnamese language generation models, including the base pre-trained model PhoGPT-4B and its conversational variant PhoGPT-4B-Chat.
Large Language Model Transformers Other
P
vinai
3,647
34
Phogpt 4B
Bsd-3-clause
PhoGPT is currently the most advanced 4-billion-parameter Vietnamese language generation model series, including the base pre-trained monolingual model PhoGPT-4B and its conversational variant PhoGPT-4B-Chat.
Large Language Model Transformers Other
P
vinai
560
12
Speechless Coder Ds 6.7b
Apache-2.0
speechless-coder-ds-6.7b is a large language model fine-tuned based on deepseek-ai/deepseek-coder-6.7b, focusing on improving code generation and programming assistance capabilities.
Large Language Model Transformers Supports Multiple Languages
S
uukuguy
771
7
Geitje 7B Chat V2
Apache-2.0
GEITje-7B is a large open-source Dutch language model based on Mistral 7B, significantly enhancing Dutch language capabilities and local knowledge through additional training on 10 billion Dutch text tokens.
Large Language Model Transformers Other
G
Rijgersberg
78
17
Geitje 7B Chat
Apache-2.0
Large open-source Dutch conversational model based on Mistral 7B, significantly improving Dutch language capabilities through additional training on 10 billion Dutch text tokens
Large Language Model Transformers Other
G
Rijgersberg
38
12
Leo Hessianai 7b Chat
The first open commercial-use German base language model built on Llama-2, focusing on German language processing
Large Language Model Transformers Supports Multiple Languages
L
LeoLM
2,263
17
Genji Python 6B Split
Apache-2.0
GPT-J 6B fine-tuned model for Python code generation, specialized in Python programming assistance
Large Language Model Transformers English
G
baffo32
16
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase