# GPU Optimization
Gte Multilingual Reranker Base Onnx Op14 Opt Gpu Int8
MIT
This is the quantized ONNX version of Alibaba-NLP/gte-multilingual-reranker-base, utilizing INT8 quantization, optimized for GPU, and suitable for text classification tasks.
Text Embedding Other
G
JustJaro
91
1
Yoso 4096
YOSO is an efficient Transformer variant that reduces the self-attention complexity from quadratic to linear through Bernoulli sampling attention mechanism, supporting sequence lengths up to 4096.
Large Language Model
Transformers

Y
uw-madison
2,072
0
Featured Recommended AI Models