Chinese Mobile Bert
C
Chinese Mobile Bert
Developed by Ayou
This model was pre-trained on a 250-million-word Chinese corpus using the MobileBERT architecture, with a training period of 15 days, completing 1 million iterations on a single A100 GPU.
Downloads 25
Release Time : 3/2/2022
Model Overview
This is an efficient Chinese pre-trained language model suitable for various natural language processing tasks.
Model Features
Efficient Training
Completed 1 million iterations on a single A100 GPU with a training period of only 15 days.
Large-scale Chinese Corpus
Pre-trained on a 250-million-word Chinese corpus, demonstrating strong Chinese comprehension capabilities.
Lightweight Architecture
Utilizes the MobileBERT architecture to reduce model complexity while maintaining performance.
Model Capabilities
Text Understanding
Text Generation
Semantic Analysis
Use Cases
Natural Language Processing
Text Classification
Can be used for tasks such as news classification and sentiment analysis.
Question Answering System
Serves as a foundational model for building Chinese question answering systems.
Featured Recommended AI Models