Minilm L12 H384 Uncased
MiniLM is a compact and efficient pre-trained language model, compressed through deep self-attention distillation technology, suitable for language understanding and generation tasks.
Downloads 10.19k
Release Time : 3/2/2022
Model Overview
MiniLM is a small pre-trained model based on the Transformer architecture, refined through task-agnostic compression and deep self-attention distillation techniques. It can directly replace BERT models but requires fine-tuning first.
Model Features
Efficient Compression
Achieves model compression through deep self-attention distillation technology, with a parameter size of only 33 million, much smaller than BERT-Base.
High Performance
Performs excellently on multiple NLP tasks, such as SQuAD 2.0 and GLUE benchmarks, with performance close to or exceeding BERT-Base.
Fast Inference
Inference speed is 2.7 times faster than BERT-Base, making it suitable for scenarios requiring efficient deployment.
Model Capabilities
Natural Language Understanding
Text Classification
Question Answering System
Use Cases
Text Analysis
Sentiment Analysis
Classifies the sentiment tendency of text
Achieves 93.0% accuracy on the SST-2 dataset
Natural Language Inference
Determines the logical relationship between two pieces of text
Achieves 85.7% accuracy on the MNLI dataset
Question Answering System
Open-domain Question Answering
Answers questions based on text content
Achieves 81.7% accuracy on the SQuAD 2.0 dataset
Featured Recommended AI Models
Š 2025AIbase