Covidbert Nli
A BERT model trained on the CORD19 coronavirus research paper dataset, fine-tuned for natural language inference tasks to generate universal sentence embeddings
Downloads 26
Release Time : 3/2/2022
Model Overview
This model is pre-trained on the CORD19 dataset and fine-tuned on SNLI and MultiNLI datasets, specifically designed for generating semantic representations of COVID-19-related texts, suitable for tasks such as scientific literature retrieval and semantic similarity calculation.
Model Features
Coronavirus Domain Adaptation
Pre-trained on the CORD19 coronavirus research paper dataset, providing better representation capabilities for COVID-19-related texts
Natural Language Inference Fine-tuning
Fine-tuned on SNLI and MultiNLI datasets, optimizing sentence-level semantic representation capabilities
Efficient Training Configuration
Optimized with a batch size of 64, 23,000 training steps, and 1,450 warm-up steps, completing training in just 6 hours on a P100 GPU
Model Capabilities
Text Semantic Representation
Sentence Similarity Calculation
Scientific Literature Retrieval
Natural Language Inference
Use Cases
Scientific Literature Processing
COVID-19 Literature Semantic Search
COVID-19 research paper retrieval system based on semantic similarity
Applied in the COVID Semantic Browser project
Scientific Literature Classification
Automatic classification of coronavirus-related research papers
Featured Recommended AI Models
Š 2025AIbase