M

Minilmv2 L6 H384 Distilled From BERT Base

Developed by nreimers
MiniLMv2 is a lightweight pre-trained language model introduced by Microsoft, achieving efficient inference through knowledge distillation technology.
Downloads 179
Release Time : 3/2/2022

Model Overview

MiniLMv2 is a lightweight pre-trained language model based on the Transformer architecture. It extracts knowledge from large models via knowledge distillation, significantly reducing model size while maintaining high performance, making it suitable for resource-constrained environments.

Model Features

Lightweight and Efficient
Significantly reduces model size through knowledge distillation, suitable for deployment in resource-constrained environments.
High Performance
Achieves performance close to or on par with large models on multiple NLP tasks.
Strong Generalization
Applicable to various natural language processing tasks without requiring significant task-specific adjustments.

Model Capabilities

Text Classification
Question Answering Systems
Text Generation
Semantic Similarity Calculation
Information Extraction

Use Cases

Intelligent Customer Service
Automated Q&A
Used to build lightweight customer service Q&A systems
Efficiently and accurately answers common questions
Mobile Applications
Mobile Text Processing
Enables localized text analysis on mobile devices like smartphones
Low latency and enhanced user privacy
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase