Bert Base Uncased Pretrained Mlm Coqa Stories
B
Bert Base Uncased Pretrained Mlm Coqa Stories
Developed by alistvt
Pre-trained language model based on BERT architecture, supporting masked language modeling tasks
Downloads 23
Release Time : 3/2/2022
Model Overview
This model is a pre-trained model based on the BERT-base architecture, trained through masked language modeling (MLM) tasks, potentially suitable for story comprehension and question answering scenarios
Model Features
Based on BERT Architecture
Utilizes the proven BERT-base architecture with strong language understanding capabilities
Pre-trained for Masked Language Modeling
Specifically optimized for MLM tasks, suitable for language understanding and generation tasks
Potential Support for Story Comprehension
The model name includes 'stories', suggesting it may have been specially trained for story texts
Model Capabilities
Text Understanding
Masked Prediction
Language Modeling
Use Cases
Natural Language Processing
Story Comprehension
Understand story text content and answer related questions
Question Answering System
Build text-based question answering systems
Featured Recommended AI Models