Xtremedistil L6 H384 Uncased
XtremeDistilTransformers is a knowledge-distilled lightweight Transformer model with task-agnostic properties, applicable to various natural language processing tasks.
Downloads 1,854
Release Time : 3/2/2022
Model Overview
This model employs task transfer techniques for knowledge distillation, learning a small general-purpose model suitable for any task and language. Compared to BERT-base, it achieves a 5.3x speedup.
Model Features
Task-Agnostic Distillation
Learns a small general-purpose model through task transfer techniques, applicable to any task and language.
Efficient Performance
Achieves a 5.3x speedup compared to BERT-base while maintaining high performance.
Multi-task Distillation Technique
Incorporates multi-task distillation techniques like XtremeDistil and MiniLM to optimize model performance.
Model Capabilities
Text Classification
Question Answering System
Natural Language Understanding
Use Cases
Natural Language Processing
Text Classification
Can be used for sentiment analysis, topic classification, and other text classification tasks.
Achieves 85.4% accuracy on the MNLI task of the GLUE development set.
Question Answering System
Suitable for building efficient question answering systems.
Achieves 76.6% F1 score on the SQuAD-v2 task.
Featured Recommended AI Models
Š 2025AIbase