Bert Base Uncased Sst2 From Bert Large Uncased Sst2
Using bert-large-uncased as the teacher model, fine-tuning the SST-2 sentiment analysis model of bert-base-uncased through knowledge distillation technology
Downloads 242
Release Time : 3/2/2022
Model Overview
This model is a BERT-base model optimized through knowledge distillation technology on the SST-2 sentiment analysis dataset. It uses a larger BERT-large as the teacher model for guided training, aiming to improve the sentiment classification accuracy while maintaining a smaller model size.
Model Features
Knowledge distillation optimization
Implement knowledge distillation from BERT-large to BERT-base using the torchdistill framework, improving performance while maintaining a smaller model size
GLUE benchmark verification
The model results have been submitted to the GLUE benchmark test, with an overall score of 78.9 points
Reproducible research
Provide complete training configurations and hyperparameters, supporting code-free experiment reproduction
Model Capabilities
English text sentiment analysis
Binary sentiment judgment (positive/negative)
Knowledge distillation model compression
Use Cases
Sentiment analysis
Movie review sentiment analysis
Analyze the sentiment tendency (positive/negative) of movie reviews
Achieved high accuracy on the SST-2 test set
Product review classification
Perform sentiment classification on user reviews on e-commerce platforms
Educational research
Knowledge distillation case study
Serves as a teaching example of knowledge distillation technology in the NLP field
Featured Recommended AI Models
Š 2025AIbase