X

Xtremedistil L12 H384 Uncased

Developed by microsoft
XtremeDistilTransformers is a task-agnostic Transformer model distilled through task transfer learning, creating a small universal model applicable to any task and language.
Downloads 471
Release Time : 3/2/2022

Model Overview

This model combines multi-task distillation techniques to achieve a small universal model through task transfer learning, suitable for various natural language processing tasks and multilingual scenarios.

Model Features

Efficient distillation
Extracts knowledge from large models through multi-stage distillation techniques, significantly reducing model size while maintaining performance.
Task-agnostic
Designed as a task-agnostic model that can be flexibly applied to various natural language processing tasks.
Multilingual support
The model design supports multilingual application scenarios.
Efficient inference
5.3 times faster than BERT-base while maintaining high performance.

Model Capabilities

Text classification
Natural language understanding
Question answering systems

Use Cases

Natural language processing
Text classification
Can be used for various text classification tasks such as sentiment analysis, topic classification, etc.
Excellent performance on GLUE benchmark
Question answering system
Suitable for building question answering systems that can understand questions and provide accurate answers.
Achieved 76.6 F1 score on SQuAD-v2 dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase