1

1epochv3

Developed by NDugar
DeBERTa is an enhanced BERT model based on the disentangled attention mechanism, surpassing BERT and RoBERTa in multiple natural language understanding tasks
Downloads 28
Release Time : 3/2/2022

Model Overview

DeBERTa improves the BERT architecture through a disentangled attention mechanism and enhanced masked decoder, particularly suitable for natural language understanding tasks. This version is XXLarge scale with 1.5 billion parameters

Model Features

Disentangled Attention Mechanism
Enhances the model's understanding of complex linguistic structures by separating content and position attention calculations
Enhanced Masked Decoder
Improved masked language modeling objective function to better capture contextual dependencies
Large-scale Pretraining
Trained on 160GB of raw text, achieving SOTA performance on multiple benchmarks

Model Capabilities

Text Classification
Natural Language Inference
Question Answering
Semantic Similarity Calculation
Sentence Pair Classification

Use Cases

Text Understanding
Sentiment Analysis
Determining text sentiment (e.g., SST-2 dataset)
97.2% accuracy
Natural Language Inference
Determining logical relationships between sentence pairs (MNLI task)
91.7%/91.9% accuracy (matched/mismatched sets)
Question Answering
SQuAD Question Answering
Passage-based question answering task
SQuAD 2.0 achieved 92.2/89.7 (F1/EM)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase