Prunedbert L12 H256 A4 Finetuned
P
Prunedbert L12 H256 A4 Finetuned
Developed by eli4s
A lightweight model based on the BERT architecture, pre-trained using knowledge distillation techniques, with a hidden layer dimension of 256 and 4 attention heads.
Downloads 16
Release Time : 3/2/2022
Model Overview
This model is a lightweight version of bert-base-uncased after pruning and knowledge distillation fine-tuning, suitable for masked language modeling tasks.
Model Features
Lightweight Design
Hidden layer dimension is 256, only one-third of BERT, making the model more lightweight.
Knowledge Distillation
Knowledge is transferred from the bert-base-uncased model using knowledge distillation techniques, reducing model size while maintaining performance.
Pruning Initialization
Model weights are initialized by pruning the weights of the bert-base-uncased model, optimizing the model structure.
Multi-Loss Fine-Tuning
Knowledge distillation fine-tuning is performed using multiple loss functions to enhance model performance.
Model Capabilities
Masked Language Prediction
Text Completion
Semantic Understanding
Use Cases
Natural Language Processing
Text Completion
Predict masked vocabulary in sentences for automatic text completion.
Accurately predicts masked vocabulary, improving text processing efficiency.
Semantic Analysis
Understand sentence semantics through masked language modeling tasks.
Effectively captures semantic information of sentences, suitable for downstream NLP tasks.
Featured Recommended AI Models