Sportsbert
S
Sportsbert
Developed by microsoft
SportsBERT is a BERT model specialized in the sports domain, trained on a corpus of sports news, suitable for sports-related natural language processing tasks.
Downloads 3,361
Release Time : 3/2/2022
Model Overview
SportsBERT is a transformer model based on the BERT architecture, specifically trained for the sports domain. Its training corpus includes sports news articles from the past four years, covering multiple sports, with approximately 8 million training samples. The model's core functionality is predicting masked words (masked language modeling task) and can be further fine-tuned for downstream tasks such as text classification and entity extraction.
Model Features
Sports domain specific
The model is specifically trained for the sports domain, including a tokenizer with more sports-related vocabulary, suitable for sports-related natural language processing tasks.
Large-scale training data
The training corpus includes sports news articles scraped from the web over the past four years, with approximately 8 million training samples, covering multiple sports.
Based on BERT architecture
The model adopts the BERT base (uncased) architecture, with strong language understanding and generation capabilities.
Model Capabilities
Fill-mask
Text classification
Entity extraction
Use Cases
Sports news analysis
Sports news classification
Classify sports news, such as football, basketball, tennis, etc.
Sports entity recognition
Identify entities in sports news, such as athletes, teams, matches, etc.
Sports content generation
Sports news summarization
Generate summaries or headlines for sports news.
Featured Recommended AI Models