S

Sportsbert

Developed by microsoft
SportsBERT is a BERT model specialized in the sports domain, trained on a corpus of sports news, suitable for sports-related natural language processing tasks.
Downloads 3,361
Release Time : 3/2/2022

Model Overview

SportsBERT is a transformer model based on the BERT architecture, specifically trained for the sports domain. Its training corpus includes sports news articles from the past four years, covering multiple sports, with approximately 8 million training samples. The model's core functionality is predicting masked words (masked language modeling task) and can be further fine-tuned for downstream tasks such as text classification and entity extraction.

Model Features

Sports domain specific
The model is specifically trained for the sports domain, including a tokenizer with more sports-related vocabulary, suitable for sports-related natural language processing tasks.
Large-scale training data
The training corpus includes sports news articles scraped from the web over the past four years, with approximately 8 million training samples, covering multiple sports.
Based on BERT architecture
The model adopts the BERT base (uncased) architecture, with strong language understanding and generation capabilities.

Model Capabilities

Fill-mask
Text classification
Entity extraction

Use Cases

Sports news analysis
Sports news classification
Classify sports news, such as football, basketball, tennis, etc.
Sports entity recognition
Identify entities in sports news, such as athletes, teams, matches, etc.
Sports content generation
Sports news summarization
Generate summaries or headlines for sports news.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase