# Retrieval-Augmented Generation
Treehop Rag
MIT
TreeHop is a lightweight embedding-level framework designed for efficient query embedding generation and filtering in multi-hop QA, significantly reducing computational overhead.
Question Answering System
T
allen-li1231
36
3
Qwen 2.5 7B Base RAG RL
Qwen-2.5-7B-base-RAG-RL is a large language model with 7B parameters trained from scratch on an unknown dataset, incorporating Retrieval-Augmented Generation (RAG) and Reinforcement Learning (RL) technologies.
Large Language Model
Transformers

Q
XXsongLALA
859
7
Jina Reranker V2 Base Multilingual Wiki Tr Rag Prefix
Apache-2.0
A fine-tuned cross-encoder model based on jina-reranker-v2-base-multilingual for text reranking and semantic search
Text Embedding
J
SMARTICT
173
0
Pleias RAG 350M
Apache-2.0
Pleias-RAG-350M is a 350-million-parameter compact reasoning model specifically trained for retrieval-augmented generation (RAG), search, and source summarization tasks.
Large Language Model
Transformers Supports Multiple Languages

P
PleIAs
292
13
Colnomic Embed Multimodal 3b
ColNomic Embed Multimodal 3B is a 3-billion-parameter multimodal embedding model specifically designed for visual document retrieval tasks, supporting unified encoding of multilingual text and images.
Multimodal Fusion Supports Multiple Languages
C
nomic-ai
4,636
17
Mxbai Rerank Large V2 GGUF
Apache-2.0
This repository contains GGUF format model files of mixedbread-ai/mxbai-rerank-large-v2, quantized by TensorBlock's machine, compatible with llama.cpp.
Text Embedding Supports Multiple Languages
M
tensorblock
150
1
Bge Reranker V2 M3 Q4 K M GGUF
Apache-2.0
This model is a GGUF format conversion of BAAI/bge-reranker-v2-m3, designed for text ranking tasks with multilingual support.
Text Embedding Other
B
sabafallah
49
0
Finseer
The first retriever specifically designed for financial time series forecasting, based on the Retrieval-Augmented Generation (RAG) framework
Large Language Model
Transformers English

F
TheFinAI
13
1
Stockllm
FinSeer StockLLM is an open-source 1-billion-parameter large language model specifically designed for financial time series forecasting, utilizing a Retrieval-Augmented Generation (RAG) framework.
Large Language Model
Transformers English

S
TheFinAI
29
1
Gen 8B R2
Apache-2.0
Gen-8B-R2 is a generation model focused on reducing hallucination issues in RAG systems, particularly suitable for handling retrieval noise and information overload.
Large Language Model
Transformers English

G
DISLab
34
2
Trendyol LLM 7B Chat V4.1.0
Apache-2.0
Trendyol LLM v4.1.0 is a generative model based on Trendyol LLM base v4.0 (a Qwen2.5 7B version further pre-trained on 13 billion tokens), specializing in e-commerce and Turkish language understanding.
Large Language Model
Safetensors Other
T
Trendyol
854
25
Kanana Nano 2.1b Embedding
Kanana is a bilingual (Korean/English) language model series developed by Kakao, excelling in Korean tasks while remaining competitive in English tasks, significantly reducing computational costs compared to models of similar scale.
Large Language Model
Transformers Supports Multiple Languages

K
kakaocorp
7,722
20
Bert Chunker 3
MIT
A text chunker based on BertForTokenClassification, suitable for structured and unstructured text, especially optimized for RAG scenarios.
Sequence Labeling
Safetensors Supports Multiple Languages
B
tim1900
1,226
1
Bge Reranker V2 M3 Q4 K M GGUF
Apache-2.0
This model is the GGUF quantized version of BAAI/bge-reranker-v2-m3, designed for text reranking tasks with multilingual support.
Text Embedding Other
B
sikreutz
46
1
Fluxi AI Small Vision
Apache-2.0
Fluxi AI is a multimodal intelligent assistant based on Qwen2-VL-7B-Instruct, capable of processing text, images, and videos, with special optimization for Portuguese language support.
Image-to-Text
Transformers Other

F
JJhooww
25
2
Lb Reranker 0.5B V1.0
Apache-2.0
The LB Reranker is a model for determining the relevance between queries and text snippets, supporting 95+ languages, suitable for ranking and reranking in retrieval tasks.
Large Language Model
Transformers Supports Multiple Languages

L
lightblue
917
66
Ruri Base V2
Apache-2.0
Ruri is a general-purpose text embedding model optimized for Japanese, based on the Sentence Transformers architecture, specifically designed for sentence similarity computation and feature extraction tasks.
Text Embedding
Safetensors Japanese
R
cl-nagoya
12.77k
4
GATE Reranker V1
Apache-2.0
A high-performance re-ranking model optimized for Arabic document retrieval and sorting
Text Embedding
Transformers Arabic

G
NAMAA-Space
615
7
Bge Reranker V2 M3 Q4 K M GGUF
Apache-2.0
This model is converted from BAAI/bge-reranker-v2-m3 into GGUF format for text reranking tasks, supporting multiple languages.
Text Embedding Other
B
puppyM
83
1
Vikhr Nemo 12B Instruct R 21 09 24
Apache-2.0
Vikhr-Nemo is a bilingual large language model optimized based on Mistral-Nemo-Instruct-2407, specifically designed for Russian and English, supporting various tasks such as logical reasoning, text summarization, and code generation.
Large Language Model
Transformers Supports Multiple Languages

V
Vikhrmodels
3,707
118
Bge Reranker V2 Gemma IQ4 NL GGUF
Apache-2.0
This is the GGUF format conversion version of the BAAI/bge-reranker-v2-gemma model, compatible with the llama.cpp framework, supporting multilingual text classification tasks.
Text Embedding Other
B
zimajava
20
1
Datagemma Rag 27b It
DataGemma is a series of models fine-tuned based on Gemma 2, specifically designed to help large language models access and integrate reliable public statistical data in Data Commons.
Large Language Model
Transformers

D
google
691
186
Phi3 Rag Relevance Judge Merge
A binary classification model for determining the relevance between reference text and questions, optimized for RAG systems
Large Language Model
Transformers

P
grounded-ai
21
1
Llm2vec Meta Llama 3 8B Instruct Mntp
MIT
LLM2Vec is a simple solution for converting decoder-only large language models into text encoders, achieved by enabling bidirectional attention mechanisms, masked next-token prediction, and unsupervised contrastive learning.
Text Embedding
Transformers English

L
McGill-NLP
3,885
16
Llama3 ChatQA 1.5 70B
Llama3-ChatQA-1.5 is a conversational QA and retrieval-augmented generation model developed based on the Llama-3 foundation model, excelling in conversational QA and RAG tasks.
Large Language Model
Transformers English

L
nvidia
205
333
Gte Base En V1.5
Apache-2.0
GTE-base-en-v1.5 is an English sentence transformer model focused on sentence similarity tasks, excelling in multiple text embedding benchmarks.
Text Embedding
Transformers Supports Multiple Languages

G
Alibaba-NLP
1.5M
63
Openthaigpt 1.0.0 70b Chat
OpenThaiGPT 70b 1.0.0 is a 70-billion-parameter Thai chat model based on LLaMA v2, specifically fine-tuned for Thai language instructions, significantly enhancing Thai language processing capabilities.
Large Language Model Supports Multiple Languages
O
openthaigpt
137
12
Llm2vec Sheared LLaMA Mntp
MIT
LLM2Vec is a simple solution for transforming decoder-only large language models into text encoders, achieved by enabling bidirectional attention, masked next-token prediction, and unsupervised contrastive learning.
Text Embedding
Transformers English

L
McGill-NLP
2,430
5
Openthaigpt 1.0.0 13b Chat
OpenThaiGPT 13b 1.0.0 is a 13-billion-parameter Thai chat model based on LLaMA v2, specifically optimized for Thai instructions
Large Language Model Supports Multiple Languages
O
openthaigpt
181
7
Slimplm Query Rewriting
A lightweight language model for query rewriting, capable of parsing user input into structured formats to optimize retrieval effectiveness.
Large Language Model
Transformers

S
zstanjj
53
9
Rank Zephyr 7b V1 Full GGUF
MIT
A text ranking model based on Mistral-7B, offering multiple quantized versions for efficient inference.
Large Language Model English
R
MaziyarPanahi
708
5
SFR Embedding Mistral
A text embedding model developed by Salesforce Research, trained on E5-mistral-7b-instruct and Mistral-7B-v0.1, primarily used for text retrieval tasks.
Text Embedding
Transformers English

S
Salesforce
34.75k
277
Guillaumetell 7b
Apache-2.0
A French large language model optimized from Mistral Open-Hermes 2.5, specifically designed for Retrieval-Augmented Generation (RAG) with source attribution and explainability
Large Language Model
Transformers French

G
AgentPublic
73
30
Bioinspiredllm
Specialized LLM fine-tuned from 13B-parameter Orca-2, focused on biological and bioinspired materials science
Large Language Model Supports Multiple Languages
B
lamm-mit
129
6
Wingpt2 14B Base
Apache-2.0
WiNGPT2 is a large-scale medical vertical domain model based on the GPT architecture, dedicated to integrating professional medical knowledge, healthcare information, and data to provide intelligent medical Q&A, diagnostic support, and medical knowledge services for the industry.
Large Language Model
Transformers Chinese

W
winninghealth
16
4
Climategpt 7b
Other
ClimateGPT is a 7-billion-parameter AI model specifically designed to integrate interdisciplinary research on climate change, developed based on the Llama-2 framework with a focus on climate science.
Large Language Model
Transformers English

C
eci-io
759
13
Climategpt 70b
Other
ClimateGPT is a 70-billion-parameter large language model specialized in climate science, based on continued pre-training and instruction fine-tuning of Llama-2, focusing on interdisciplinary research integration in climate change
Large Language Model
Transformers English

C
eci-io
74
11
Stella Base En V2
MIT
stella-base-en-v2 is an English text embedding model based on sentence-transformers, focusing on sentence similarity and feature extraction tasks.
Text Embedding English
S
infgrad
16.89k
15
Wingpt2 7B Base
Apache-2.0
WiNGPT2 is a large medical vertical domain model based on the GPT architecture, dedicated to integrating professional medical knowledge, healthcare information, and data to provide intelligent Q&A, diagnostic support, and medical knowledge services for the industry.
Large Language Model
Transformers Chinese

W
winninghealth
26
10
Sentence Transformers Multilingual E5 Base
This is a multilingual sentence transformer model that maps sentences and paragraphs into a 768-dimensional dense vector space, supporting multiple languages and suitable for tasks like clustering or semantic search.
Text Embedding
S
embaas
3,526
8
- 1
- 2
Featured Recommended AI Models