# Unidic Tokenization
Bert Base Japanese Char V2
BERT model pre-trained on Japanese text using character-level tokenization and whole word masking, trained on the Japanese Wikipedia version as of August 31, 2020
Large Language Model Japanese
B
tohoku-nlp
134.28k
6
Bert Large Japanese
BERT large model pretrained on Japanese Wikipedia, utilizing Unidic dictionary tokenization and whole word masking strategy
Large Language Model Japanese
B
tohoku-nlp
1,272
9
Featured Recommended AI Models