XYZ Embedding Zh
模型简介
该模型专门为中文文本设计,能够高效地将文本转换为高维向量表示,支持多种自然语言处理任务,如信息检索、重排序等。
模型特点
高维向量表示
将句子和段落映射到1792维的密集向量空间,捕捉丰富的语义信息。
多任务支持
支持多种任务,包括句子相似度计算、特征提取、重排序和信息检索。
中文优化
专门针对中文文本进行优化,能够更好地处理中文语义。
模型能力
句子相似度计算
特征提取
文本重排序
信息检索
使用案例
信息检索
医疗问答检索
在医疗问答数据集中进行信息检索,帮助用户快速找到相关答案。
在 MTEB Cmedqa检索 数据集上,map_at_10 达到 41.228。
电商产品检索
在电商平台上进行产品检索,提升用户搜索体验。
在 MTEB 电商检索 数据集上,ndcg_at_10 达到 69.719。
文本重排序
医疗问答重排序
对医疗问答结果进行重排序,提升答案的相关性。
在 MTEB CMedQAv1 数据集上,map 达到 89.618。
通用文本重排序
对通用文本检索结果进行重排序,优化搜索结果。
在 MTEB T2重排序 数据集上,map 达到 69.066。
🚀 XYZ-embedding-zh
XYZ-embedding-zh 是一个基于 sentence-transformers 的模型,它能够将句子和段落映射到 1792 维的密集向量空间,可用于聚类或语义搜索等任务。
🚀 快速开始
使用该模型前,你需要安装 sentence-transformers:
pip install -U sentence-transformers
然后,你可以按照以下方式使用该模型:
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('fangxq/XYZ-embedding-zh')
embeddings = model.encode(sentences)
print(embeddings)
✨ 主要特性
- 能够将句子和段落映射到 1792 维的密集向量空间。
- 可用于聚类或语义搜索等任务。
📦 安装指南
使用该模型前,你需要安装 sentence-transformers:
pip install -U sentence-transformers
💻 使用示例
基础用法
from sentence_transformers import SentenceTransformer
sentences = ["This is an example sentence", "Each sentence is converted"]
model = SentenceTransformer('fangxq/XYZ-embedding-zh')
embeddings = model.encode(sentences)
print(embeddings)
📚 详细文档
评估结果
若要对该模型进行自动评估,请参考 Sentence Embeddings Benchmark:https://seb.sbert.net
完整模型架构
SentenceTransformer(
(0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
(1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
(2): Dense({'in_features': 1024, 'out_features': 1792, 'bias': True, 'activation_function': 'torch.nn.modules.linear.Identity'})
)
引用与作者
原文档未提供相关内容。
模型指标详情
数据集名称 | 任务类型 | 指标类型 | 指标值 |
---|---|---|---|
MTEB CMedQAv1 | Reranking | map | 89.61792115239176 |
MTEB CMedQAv1 | Reranking | mrr | 91.46722222222222 |
MTEB CMedQAv1 | Reranking | main_score | 89.61792115239176 |
MTEB CMedQAv2 | Reranking | map | 89.22040591564271 |
MTEB CMedQAv2 | Reranking | mrr | 91.2995238095238 |
MTEB CMedQAv2 | Reranking | main_score | 89.22040591564271 |
MTEB CmedqaRetrieval | Retrieval | map_at_1 | 27.939000000000004 |
MTEB CmedqaRetrieval | Retrieval | map_at_10 | 41.227999999999994 |
MTEB CmedqaRetrieval | Retrieval | map_at_100 | 43.018 |
MTEB CmedqaRetrieval | Retrieval | map_at_1000 | 43.120000000000005 |
MTEB CmedqaRetrieval | Retrieval | map_at_3 | 36.895 |
MTEB CmedqaRetrieval | Retrieval | map_at_5 | 39.373999999999995 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_1 | 42.136 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_10 | 50.394000000000005 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_100 | 51.288 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_1000 | 51.324000000000005 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_3 | 47.887 |
MTEB CmedqaRetrieval | Retrieval | mrr_at_5 | 49.362 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_1 | 42.136 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_10 | 47.899 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_100 | 54.730999999999995 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_1000 | 56.462999999999994 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_3 | 42.66 |
MTEB CmedqaRetrieval | Retrieval | ndcg_at_5 | 44.913 |
MTEB CmedqaRetrieval | Retrieval | precision_at_1 | 42.136 |
MTEB CmedqaRetrieval | Retrieval | precision_at_10 | 10.52 |
MTEB CmedqaRetrieval | Retrieval | precision_at_100 | 1.6070000000000002 |
MTEB CmedqaRetrieval | Retrieval | precision_at_1000 | 0.183 |
MTEB CmedqaRetrieval | Retrieval | precision_at_3 | 24.064 |
MTEB CmedqaRetrieval | Retrieval | precision_at_5 | 17.374000000000002 |
MTEB CmedqaRetrieval | Retrieval | recall_at_1 | 27.939000000000004 |
MTEB CmedqaRetrieval | Retrieval | recall_at_10 | 58.29600000000001 |
MTEB CmedqaRetrieval | Retrieval | recall_at_100 | 86.504 |
MTEB CmedqaRetrieval | Retrieval | recall_at_1000 | 98.105 |
MTEB CmedqaRetrieval | Retrieval | recall_at_3 | 42.475 |
MTEB CmedqaRetrieval | Retrieval | recall_at_5 | 49.454 |
MTEB CmedqaRetrieval | Retrieval | main_score | 47.899 |
MTEB CovidRetrieval | Retrieval | map_at_1 | 77.371 |
MTEB CovidRetrieval | Retrieval | map_at_10 | 85.229 |
MTEB CovidRetrieval | Retrieval | map_at_100 | 85.358 |
MTEB CovidRetrieval | Retrieval | map_at_1000 | 85.36 |
MTEB CovidRetrieval | Retrieval | map_at_3 | 84.176 |
MTEB CovidRetrieval | Retrieval | map_at_5 | 84.79299999999999 |
MTEB CovidRetrieval | Retrieval | mrr_at_1 | 77.661 |
MTEB CovidRetrieval | Retrieval | mrr_at_10 | 85.207 |
MTEB CovidRetrieval | Retrieval | mrr_at_100 | 85.33699999999999 |
MTEB CovidRetrieval | Retrieval | mrr_at_1000 | 85.339 |
MTEB CovidRetrieval | Retrieval | mrr_at_3 | 84.229 |
MTEB CovidRetrieval | Retrieval | mrr_at_5 | 84.79299999999999 |
MTEB CovidRetrieval | Retrieval | ndcg_at_1 | 77.766 |
MTEB CovidRetrieval | Retrieval | ndcg_at_10 | 88.237 |
MTEB CovidRetrieval | Retrieval | ndcg_at_100 | 88.777 |
MTEB CovidRetrieval | Retrieval | ndcg_at_1000 | 88.818 |
MTEB CovidRetrieval | Retrieval | ndcg_at_3 | 86.16 |
MTEB CovidRetrieval | Retrieval | ndcg_at_5 | 87.22 |
MTEB CovidRetrieval | Retrieval | precision_at_1 | 77.766 |
MTEB CovidRetrieval | Retrieval | precision_at_10 | 9.841999999999999 |
MTEB CovidRetrieval | Retrieval | precision_at_100 | 1.0070000000000001 |
MTEB CovidRetrieval | Retrieval | precision_at_1000 | 0.101 |
MTEB CovidRetrieval | Retrieval | precision_at_3 | 30.875000000000004 |
MTEB CovidRetrieval | Retrieval | precision_at_5 | 19.073 |
MTEB CovidRetrieval | Retrieval | recall_at_1 | 77.371 |
MTEB CovidRetrieval | Retrieval | recall_at_10 | 97.366 |
MTEB CovidRetrieval | Retrieval | recall_at_100 | 99.684 |
MTEB CovidRetrieval | Retrieval | recall_at_1000 | 100.0 |
MTEB CovidRetrieval | Retrieval | recall_at_3 | 91.702 |
MTEB CovidRetrieval | Retrieval | recall_at_5 | 94.31 |
MTEB CovidRetrieval | Retrieval | main_score | 88.237 |
MTEB DuRetrieval | Retrieval | map_at_1 | 27.772000000000002 |
MTEB DuRetrieval | Retrieval | map_at_10 | 84.734 |
MTEB DuRetrieval | Retrieval | map_at_100 | 87.298 |
MTEB DuRetrieval | Retrieval | map_at_1000 | 87.32900000000001 |
MTEB DuRetrieval | Retrieval | map_at_3 | 59.431 |
MTEB DuRetrieval | Retrieval | map_at_5 | 74.82900000000001 |
MTEB DuRetrieval | Retrieval | mrr_at_1 | 93.65 |
MTEB DuRetrieval | Retrieval | mrr_at_10 | 95.568 |
MTEB DuRetrieval | Retrieval | mrr_at_100 | 95.608 |
MTEB DuRetrieval | Retrieval | mrr_at_1000 | 95.609 |
MTEB DuRetrieval | Retrieval | mrr_at_3 | 95.267 |
MTEB DuRetrieval | Retrieval | mrr_at_5 | 95.494 |
MTEB DuRetrieval | Retrieval | ndcg_at_1 | 93.65 |
MTEB DuRetrieval | Retrieval | ndcg_at_10 | 90.794 |
MTEB DuRetrieval | Retrieval | ndcg_at_100 | 92.88300000000001 |
MTEB DuRetrieval | Retrieval | ndcg_at_1000 | 93.144 |
MTEB DuRetrieval | Retrieval | ndcg_at_3 | 90.32 |
MTEB DuRetrieval | Retrieval | ndcg_at_5 | 89.242 |
MTEB DuRetrieval | Retrieval | precision_at_1 | 93.65 |
MTEB DuRetrieval | Retrieval | precision_at_10 | 42.9 |
MTEB DuRetrieval | Retrieval | precision_at_100 | 4.835 |
MTEB DuRetrieval | Retrieval | precision_at_1000 | 0.49 |
MTEB DuRetrieval | Retrieval | precision_at_3 | 80.85 |
MTEB DuRetrieval | Retrieval | precision_at_5 | 68.14 |
MTEB DuRetrieval | Retrieval | recall_at_1 | 27.772000000000002 |
MTEB DuRetrieval | Retrieval | recall_at_10 | 91.183 |
MTEB DuRetrieval | Retrieval | recall_at_100 | 98.219 |
MTEB DuRetrieval | Retrieval | recall_at_1000 | 99.55000000000001 |
MTEB DuRetrieval | Retrieval | recall_at_3 | 60.911 |
MTEB DuRetrieval | Retrieval | recall_at_5 | 78.31099999999999 |
MTEB DuRetrieval | Retrieval | main_score | 90.794 |
MTEB EcomRetrieval | Retrieval | map_at_1 | 54.6 |
MTEB EcomRetrieval | Retrieval | map_at_10 | 64.742 |
MTEB EcomRetrieval | Retrieval | map_at_100 | 65.289 |
MTEB EcomRetrieval | Retrieval | map_at_1000 | 65.29700000000001 |
MTEB EcomRetrieval | Retrieval | map_at_3 | 62.183 |
MTEB EcomRetrieval | Retrieval | map_at_5 | 63.883 |
MTEB EcomRetrieval | Retrieval | mrr_at_1 | 54.6 |
MTEB EcomRetrieval | Retrieval | mrr_at_10 | 64.742 |
MTEB EcomRetrieval | Retrieval | mrr_at_100 | 65.289 |
MTEB EcomRetrieval | Retrieval | mrr_at_1000 | 65.29700000000001 |
MTEB EcomRetrieval | Retrieval | mrr_at_3 | 62.183 |
MTEB EcomRetrieval | Retrieval | mrr_at_5 | 63.883 |
MTEB EcomRetrieval | Retrieval | ndcg_at_1 | 54.6 |
MTEB EcomRetrieval | Retrieval | ndcg_at_10 | 69.719 |
MTEB EcomRetrieval | Retrieval | ndcg_at_100 | 72.148 |
MTEB EcomRetrieval | Retrieval | ndcg_at_1000 | 72.393 |
MTEB EcomRetrieval | Retrieval | ndcg_at_3 | 64.606 |
MTEB EcomRetrieval | Retrieval | ndcg_at_5 | 67.682 |
MTEB EcomRetrieval | Retrieval | precision_at_1 | 54.6 |
MTEB EcomRetrieval | Retrieval | precision_at_10 | 8.53 |
MTEB EcomRetrieval | Retrieval | precision_at_100 | 0.962 |
MTEB EcomRetrieval | Retrieval | precision_at_1000 | 0.098 |
MTEB EcomRetrieval | Retrieval | precision_at_3 | 23.867 |
MTEB EcomRetrieval | Retrieval | precision_at_5 | 15.82 |
MTEB EcomRetrieval | Retrieval | recall_at_1 | 54.6 |
MTEB EcomRetrieval | Retrieval | recall_at_10 | 85.3 |
MTEB EcomRetrieval | Retrieval | recall_at_100 | 96.2 |
MTEB EcomRetrieval | Retrieval | recall_at_1000 | 98.2 |
MTEB EcomRetrieval | Retrieval | recall_at_3 | 71.6 |
MTEB EcomRetrieval | Retrieval | recall_at_5 | 79.10000000000001 |
MTEB EcomRetrieval | Retrieval | main_score | 69.719 |
MTEB MMarcoReranking | Reranking | map | 35.30260957061897 |
MTEB MMarcoReranking | Reranking | mrr | 34.098015873015875 |
MTEB MMarcoReranking | Reranking | main_score | 35.30260957061897 |
MTEB MMarcoRetrieval | Retrieval | map_at_1 | 69.51899999999999 |
MTEB MMarcoRetrieval | Retrieval | map_at_10 | 78.816 |
MTEB MMarcoRetrieval | Retrieval | map_at_100 | 79.08500000000001 |
MTEB MMarcoRetrieval | Retrieval | map_at_1000 | 79.091 |
MTEB MMarcoRetrieval | Retrieval | map_at_3 | 76.999 |
MTEB MMarcoRetrieval | Retrieval | map_at_5 | 78.194 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_1 | 71.80499999999999 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_10 | 79.29899999999999 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_100 | 79.532 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_1000 | 79.537 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_3 | 77.703 |
MTEB MMarcoRetrieval | Retrieval | mrr_at_5 | 78.75999999999999 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_1 | 71.80499999999999 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_10 | 82.479 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_100 | 83.611 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_1000 | 83.76400000000001 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_3 | 79.065 |
MTEB MMarcoRetrieval | Retrieval | ndcg_at_5 | 81.092 |
MTEB MMarcoRetrieval | Retrieval | precision_at_1 | 71.80499999999999 |
MTEB MMarcoRetrieval | Retrieval | precision_at_10 | 9.91 |
MTEB MMarcoRetrieval | Retrieval | precision_at_100 | 1.046 |
MTEB MMarcoRetrieval | Retrieval | precision_at_1000 | 0.106 |
MTEB MMarcoRetrieval | Retrieval | precision_at_3 | 29.727999999999998 |
MTEB MMarcoRetrieval | Retrieval | precision_at_5 | 18.908 |
MTEB MMarcoRetrieval | Retrieval | recall_at_1 | 69.51899999999999 |
MTEB MMarcoRetrieval | Retrieval | recall_at_10 | 93.24 |
MTEB MMarcoRetrieval | Retrieval | recall_at_100 | 98.19099999999999 |
MTEB MMarcoRetrieval | Retrieval | recall_at_1000 | 99.36500000000001 |
MTEB MMarcoRetrieval | Retrieval | recall_at_3 | 84.308 |
MTEB MMarcoRetrieval | Retrieval | recall_at_5 | 89.119 |
MTEB MMarcoRetrieval | Retrieval | main_score | 82.479 |
MTEB MedicalRetrieval | Retrieval | map_at_1 | 57.8 |
MTEB MedicalRetrieval | Retrieval | map_at_10 | 64.215 |
MTEB MedicalRetrieval | Retrieval | map_at_100 | 64.78 |
MTEB MedicalRetrieval | Retrieval | map_at_1000 | 64.81099999999999 |
MTEB MedicalRetrieval | Retrieval | map_at_3 | 62.64999999999999 |
MTEB MedicalRetrieval | Retrieval | map_at_5 | 63.57000000000001 |
MTEB MedicalRetrieval | Retrieval | mrr_at_1 | 58.099999999999994 |
MTEB MedicalRetrieval | Retrieval | mrr_at_10 | 64.371 |
MTEB MedicalRetrieval | Retrieval | mrr_at_100 | 64.936 |
MTEB MedicalRetrieval | Retrieval | mrr_at_1000 | 64.96600000000001 |
MTEB MedicalRetrieval | Retrieval | mrr_at_3 | 62.8 |
MTEB MedicalRetrieval | Retrieval | mrr_at_5 | 63.739999999999995 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_1 | 57.8 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_10 | 67.415 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_100 | 70.38799999999999 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_1000 | 71.229 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_3 | 64.206 |
MTEB MedicalRetrieval | Retrieval | ndcg_at_5 | 65.858 |
MTEB MedicalRetrieval | Retrieval | precision_at_1 | 57.8 |
MTEB MedicalRetrieval | Retrieval | precision_at_10 | 7.75 |
MTEB MedicalRetrieval | Retrieval | precision_at_100 | 0.919 |
MTEB MedicalRetrieval | Retrieval | precision_at_1000 | 0.099 |
MTEB MedicalRetrieval | Retrieval | precision_at_3 | 22.900000000000002 |
MTEB MedicalRetrieval | Retrieval | precision_at_5 | 14.540000000000001 |
MTEB MedicalRetrieval | Retrieval | recall_at_1 | 57.8 |
MTEB MedicalRetrieval | Retrieval | recall_at_10 | 77.5 |
MTEB MedicalRetrieval | Retrieval | recall_at_100 | 91.9 |
MTEB MedicalRetrieval | Retrieval | recall_at_1000 | 98.6 |
MTEB MedicalRetrieval | Retrieval | recall_at_3 | 68.7 |
MTEB MedicalRetrieval | Retrieval | recall_at_5 | 72.7 |
MTEB MedicalRetrieval | Retrieval | main_score | 67.415 |
MTEB T2Reranking | Reranking | map | 69.06615146698508 |
MTEB T2Reranking | Reranking | mrr | 79.7588755091294 |
MTEB T2Reranking | Reranking | main_score | 69.06615146698508 |
MTEB T2Retrieval | Retrieval | map_at_1 | 28.084999999999997 |
MTEB T2Retrieval | Retrieval | map_at_10 | 78.583 |
MTEB T2Retrieval | Retrieval | map_at_100 | 82.14399999999999 |
MTEB T2Retrieval | Retrieval | map_at_1000 | 82.204 |
MTEB T2Retrieval | Retrieval | map_at_3 | 55.422000000000004 |
MTEB T2Retrieval | Retrieval | map_at_5 | 67.973 |
MTEB T2Retrieval | Retrieval | mrr_at_1 | 91.014 |
MTEB T2Retrieval | Retrieval | mrr_at_10 | 93.381 |
MTEB T2Retrieval | Retrieval | mrr_at_100 | 93.45400000000001 |
MTEB T2Retrieval | Retrieval | mrr_at_1000 | 93.45599999999999 |
MTEB T2Retrieval | Retrieval | mrr_at_3 | 92.99300000000001 |
MTEB T2Retrieval | Retrieval | mrr_at_5 | 93.234 |
MTEB T2Retrieval | Retrieval | ndcg_at_1 | 91.014 |
MTEB T2Retrieval | Retrieval | ndcg_at_10 | 85.931 |
MTEB T2Retrieval | Retrieval | ndcg_at_100 | 89.31 |
MTEB T2Retrieval | Retrieval | ndcg_at_1000 | 89.869 |
MTEB T2Retrieval | Retrieval | ndcg_at_3 | 87.348 |
MTEB T2Retrieval | Retrieval | ndcg_at_5 | 85.929 |
MTEB T2Retrieval | Retrieval | precision_at_1 | 91.014 |
MTEB T2Retrieval | Retrieval | precision_at_10 | 42.495 |
MTEB T2Retrieval | Retrieval | precision_at_100 | 5.029999999999999 |
MTEB T2Retrieval | Retrieval | precision_at_1000 | 0.516 |
MTEB T2Retrieval | Retrieval | precision_at_3 | 76.248 |
MTEB T2Retrieval | Retrieval | precision_at_5 | 63.817 |
MTEB T2Retrieval | Retrieval | recall_at_1 | 28.084999999999997 |
MTEB T2Retrieval | Retrieval | recall_at_10 | 84.88 |
MTEB T2Retrieval | Retrieval | recall_at_100 | 95.902 |
MTEB T2Retrieval | Retrieval | recall_at_1000 | 98.699 |
MTEB T2Retrieval | Retrieval | recall_at_3 | 57.113 |
MTEB T2Retrieval | Retrieval | recall_at_5 | 71.251 |
MTEB T2Retrieval | Retrieval | main_score | 85.931 |
MTEB VideoRetrieval | Retrieval | map_at_1 | 66.4 |
MTEB VideoRetrieval | Retrieval | map_at_10 | 75.86 |
MTEB VideoRetrieval | Retrieval | map_at_100 | 76.185 |
MTEB VideoRetrieval | Retrieval | map_at_1000 | 76.188 |
MTEB VideoRetrieval | Retrieval | map_at_3 | 74.167 |
MTEB VideoRetrieval | Retrieval | map_at_5 | 75.187 |
MTEB VideoRetrieval | Retrieval | mrr_at_1 | 66.4 |
MTEB VideoRetrieval | Retrieval | mrr_at_10 | 75.86 |
MTEB VideoRetrieval | Retrieval | mrr_at_100 | 76.185 |
MTEB VideoRetrieval | Retrieval | mrr_at_1000 | 76.188 |
MTEB VideoRetrieval | Retrieval | mrr_at_3 | 74.167 |
MTEB VideoRetrieval | Retrieval | mrr_at_5 | 75.187 |
MTEB VideoRetrieval | Retrieval | ndcg_at_1 | 66.4 |
MTEB VideoRetrieval | Retrieval | ndcg_at_10 | 80.03099999999999 |
MTEB VideoRetrieval | Retrieval | ndcg_at_100 | 81.459 |
MTEB VideoRetrieval | Retrieval | ndcg_at_1000 | 81.527 |
MTEB VideoRetrieval | Retrieval | ndcg_at_3 | 76.621 |
MTEB VideoRetrieval | Retrieval | ndcg_at_5 | 78.446 |
MTEB VideoRetrieval | Retrieval | precision_at_1 | 66.4 |
MTEB VideoRetrieval | Retrieval | precision_at_10 | 9.29 |
MTEB VideoRetrieval | Retrieval | precision_at_100 | 0.992 |
MTEB VideoRetrieval | Retrieval | precision_at_1000 | 0.1 |
MTEB VideoRetrieval | Retrieval | precision_at_3 | 27.900000000000002 |
MTEB VideoRetrieval | Retrieval | precision_at_5 | 17.62 |
MTEB VideoRetrieval | Retrieval | recall_at_1 | 66.4 |
MTEB VideoRetrieval | Retrieval | recall_at_10 | 92.9 |
MTEB VideoRetrieval | Retrieval | recall_at_100 | 99.2 |
MTEB VideoRetrieval | Retrieval | recall_at_1000 | 99.7 |
MTEB VideoRetrieval | Retrieval | recall_at_3 | 83.7 |
MTEB VideoRetrieval | Retrieval | recall_at_5 | 88.1 |
MTEB VideoRetrieval | Retrieval | main_score | 80.03099999999999 |
Jina Embeddings V3
Jina Embeddings V3 是一个多语言句子嵌入模型,支持超过100种语言,专注于句子相似度和特征提取任务。
文本嵌入
Transformers 支持多种语言

J
jinaai
3.7M
911
Ms Marco MiniLM L6 V2
Apache-2.0
基于MS Marco段落排序任务训练的交叉编码器模型,用于信息检索中的查询-段落相关性评分
文本嵌入 英语
M
cross-encoder
2.5M
86
Opensearch Neural Sparse Encoding Doc V2 Distill
Apache-2.0
基于蒸馏技术的稀疏检索模型,专为OpenSearch优化,支持免推理文档编码,在搜索相关性和效率上优于V1版本
文本嵌入
Transformers 英语

O
opensearch-project
1.8M
7
Sapbert From PubMedBERT Fulltext
Apache-2.0
基于PubMedBERT的生物医学实体表征模型,通过自对齐预训练优化语义关系捕捉
文本嵌入 英语
S
cambridgeltl
1.7M
49
Gte Large
MIT
GTE-Large 是一个强大的句子转换器模型,专注于句子相似度和文本嵌入任务,在多个基准测试中表现出色。
文本嵌入 英语
G
thenlper
1.5M
278
Gte Base En V1.5
Apache-2.0
GTE-base-en-v1.5 是一个英文句子转换器模型,专注于句子相似度任务,在多个文本嵌入基准测试中表现优异。
文本嵌入
Transformers 支持多种语言

G
Alibaba-NLP
1.5M
63
Gte Multilingual Base
Apache-2.0
GTE Multilingual Base 是一个多语言的句子嵌入模型,支持超过50种语言,适用于句子相似度计算等任务。
文本嵌入
Transformers 支持多种语言

G
Alibaba-NLP
1.2M
246
Polybert
polyBERT是一个化学语言模型,旨在实现完全由机器驱动的超快聚合物信息学。它将PSMILES字符串映射为600维密集指纹,以数值形式表示聚合物化学结构。
文本嵌入
Transformers

P
kuelumbus
1.0M
5
Bert Base Turkish Cased Mean Nli Stsb Tr
Apache-2.0
基于土耳其语BERT的句子嵌入模型,专为语义相似度任务优化
文本嵌入
Transformers 其他

B
emrecan
1.0M
40
GIST Small Embedding V0
MIT
基于BAAI/bge-small-en-v1.5模型微调的文本嵌入模型,通过MEDI数据集与MTEB分类任务数据集训练,优化了检索任务的查询编码能力。
文本嵌入
Safetensors 英语
G
avsolatorio
945.68k
29
精选推荐AI模型
Llama 3 Typhoon V1.5x 8b Instruct
专为泰语设计的80亿参数指令模型,性能媲美GPT-3.5-turbo,优化了应用场景、检索增强生成、受限生成和推理任务
大型语言模型
Transformers 支持多种语言

L
scb10x
3,269
16
Cadet Tiny
Openrail
Cadet-Tiny是一个基于SODA数据集训练的超小型对话模型,专为边缘设备推理设计,体积仅为Cosmo-3B模型的2%左右。
对话系统
Transformers 英语

C
ToddGoldfarb
2,691
6
Roberta Base Chinese Extractive Qa
基于RoBERTa架构的中文抽取式问答模型,适用于从给定文本中提取答案的任务。
问答系统 中文
R
uer
2,694
98