Sarashina2 13b
A large language model trained by SB Intuitions, supporting Japanese and English, based on the Llama2 architecture
Downloads 1,167
Release Time : 6/7/2024
Model Overview
Sarashina2-13B is a large language model based on the Llama2 architecture, supporting Japanese and English text generation tasks. The model has been trained on 2.1 trillion tokens, exhibiting strong language understanding and generation capabilities.
Model Features
Multilingual Support
Supports both Japanese and English processing, capable of handling Japanese text without pre-tokenization
Large-scale Training
Trained on 2.1 trillion tokens, demonstrating robust language understanding and generation capabilities
Efficient Tokenization
Utilizes a sentencepiece tokenizer based on unigram language models, supporting byte fallback mechanism
Model Capabilities
Japanese Text Generation
English Text Generation
Multi-turn Dialogue
Text Continuation
Use Cases
Content Creation
Article Continuation
Automatically generates subsequent content based on the opening paragraph
Can produce coherent text content
Dialogue System
Building multi-turn dialogue robots
Capable of basic conversational interactions
Education
Language Learning Assistance
Helps Japanese or English learners practice writing
Provides language demonstrations and feedback
Featured Recommended AI Models