C

Covidbert Nli

Developed by gsarti
A BERT model trained on the CORD19 coronavirus research paper dataset, fine-tuned for natural language inference tasks to generate universal sentence embeddings
Downloads 26
Release Time : 3/2/2022

Model Overview

This model is pre-trained on the CORD19 dataset and fine-tuned on SNLI and MultiNLI datasets, specifically designed for generating semantic representations of COVID-19-related texts, suitable for tasks such as scientific literature retrieval and semantic similarity calculation.

Model Features

Coronavirus Domain Adaptation
Pre-trained on the CORD19 coronavirus research paper dataset, providing better representation capabilities for COVID-19-related texts
Natural Language Inference Fine-tuning
Fine-tuned on SNLI and MultiNLI datasets, optimizing sentence-level semantic representation capabilities
Efficient Training Configuration
Optimized with a batch size of 64, 23,000 training steps, and 1,450 warm-up steps, completing training in just 6 hours on a P100 GPU

Model Capabilities

Text Semantic Representation
Sentence Similarity Calculation
Scientific Literature Retrieval
Natural Language Inference

Use Cases

Scientific Literature Processing
COVID-19 Literature Semantic Search
COVID-19 research paper retrieval system based on semantic similarity
Applied in the COVID Semantic Browser project
Scientific Literature Classification
Automatic classification of coronavirus-related research papers
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase