B

Bert Base Uncased Sst2 From Bert Large Uncased Sst2

Developed by yoshitomo-matsubara
Using bert-large-uncased as the teacher model, fine-tuning the SST-2 sentiment analysis model of bert-base-uncased through knowledge distillation technology
Downloads 242
Release Time : 3/2/2022

Model Overview

This model is a BERT-base model optimized through knowledge distillation technology on the SST-2 sentiment analysis dataset. It uses a larger BERT-large as the teacher model for guided training, aiming to improve the sentiment classification accuracy while maintaining a smaller model size.

Model Features

Knowledge distillation optimization
Implement knowledge distillation from BERT-large to BERT-base using the torchdistill framework, improving performance while maintaining a smaller model size
GLUE benchmark verification
The model results have been submitted to the GLUE benchmark test, with an overall score of 78.9 points
Reproducible research
Provide complete training configurations and hyperparameters, supporting code-free experiment reproduction

Model Capabilities

English text sentiment analysis
Binary sentiment judgment (positive/negative)
Knowledge distillation model compression

Use Cases

Sentiment analysis
Movie review sentiment analysis
Analyze the sentiment tendency (positive/negative) of movie reviews
Achieved high accuracy on the SST-2 test set
Product review classification
Perform sentiment classification on user reviews on e-commerce platforms
Educational research
Knowledge distillation case study
Serves as a teaching example of knowledge distillation technology in the NLP field
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase