D

Deberta Large Mnli

Developed by microsoft
DeBERTa-V2-XXLarge is an improved BERT model based on the disentangled attention mechanism and enhanced masked decoder, excelling in multiple natural language understanding tasks.
Downloads 1.4M
Release Time : 3/2/2022

Model Overview

DeBERTa improves upon BERT and RoBERTa models through its disentangled attention mechanism and enhanced masked decoder. Trained on 80GB of data, it surpasses the performance of BERT and RoBERTa in most natural language understanding tasks.

Model Features

Disentangled Attention Mechanism
Improves traditional self-attention mechanisms through disentangled attention, enhancing model performance.
Enhanced Masked Decoder
Utilizes an enhanced masked decoder to further improve performance in natural language understanding tasks.
Large-scale Training Data
Trained on 80GB of data, covering a wide range of natural language understanding scenarios.

Model Capabilities

Text Classification
Question Answering Systems
Natural Language Inference
Semantic Similarity Calculation

Use Cases

Natural Language Understanding
Text Classification
Can be used for sentiment analysis, topic classification, and other text classification tasks.
Achieves 97.2% accuracy on the SST-2 dataset.
Question Answering Systems
Can be used to build question answering systems to respond to user queries.
Achieves 96.1/91.4 F1/EM scores on the SQuAD 1.1 dataset.
Natural Language Inference
Can be used to determine the logical relationship between two sentences.
Achieves 91.7/91.9 accuracy on the MNLI dataset.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase