đ BERT base Japanese (character tokenization)
This is a BERT model pretrained on Japanese texts. It processes input texts with word - level tokenization based on the IPA dictionary and then character - level tokenization.
đ Quick Start
This is a BERT model pretrained on texts in the Japanese language. This version of the model processes input texts with word - level tokenization based on the IPA dictionary, followed by character - level tokenization. The codes for the pretraining are available at [cl - tohoku/bert - japanese](https://github.com/cl - tohoku/bert - japanese/tree/v1.0).
⨠Features
- The model architecture is the same as the original BERT base model; 12 layers, 768 dimensions of hidden states, and 12 attention heads.
- The texts are first tokenized by MeCab morphological parser with the IPA dictionary and then split into characters. The vocabulary size is 4000.
- The model is trained with the same configuration as the original BERT; 512 tokens per instance, 256 instances per batch, and 1M training steps.
đĻ Installation
No installation steps are provided in the original README, so this section is skipped.
đģ Usage Examples
No code examples are provided in the original README, so this section is skipped.
đ Documentation
Model architecture
The model architecture is the same as the original BERT base model; 12 layers, 768 dimensions of hidden states, and 12 attention heads.
Training Data
The model is trained on Japanese Wikipedia as of September 1, 2019. To generate the training corpus, WikiExtractor is used to extract plain texts from a dump file of Wikipedia articles. The text files used for the training are 2.6GB in size, consisting of approximately 17M sentences.
Tokenization
The texts are first tokenized by MeCab morphological parser with the IPA dictionary and then split into characters. The vocabulary size is 4000.
Training
The model is trained with the same configuration as the original BERT; 512 tokens per instance, 256 instances per batch, and 1M training steps.
đ§ Technical Details
The model combines word - level tokenization using the IPA dictionary and MeCab morphological parser with character - level tokenization. It follows the original BERT architecture and training configuration, which allows it to effectively learn from Japanese texts. The use of Wikipedia data as training corpus provides a large - scale and diverse set of Japanese language data for the model to learn from.
đ License
The pretrained models are distributed under the terms of the [Creative Commons Attribution - ShareAlike 3.0](https://creativecommons.org/licenses/by - sa/3.0/).
Acknowledgments
For training models, we used Cloud TPUs provided by TensorFlow Research Cloud program.
Information Table
Property |
Details |
Model Type |
BERT base Japanese (character tokenization) |
Training Data |
Japanese Wikipedia as of September 1, 2019, extracted by WikiExtractor, 2.6GB in size, approximately 17M sentences |
License |
[Creative Commons Attribution - ShareAlike 3.0](https://creativecommons.org/licenses/by - sa/3.0/) |