S

Sarashina2 13b

Developed by sbintuitions
A large language model trained by SB Intuitions, supporting Japanese and English, based on the Llama2 architecture
Downloads 1,167
Release Time : 6/7/2024

Model Overview

Sarashina2-13B is a large language model based on the Llama2 architecture, supporting Japanese and English text generation tasks. The model has been trained on 2.1 trillion tokens, exhibiting strong language understanding and generation capabilities.

Model Features

Multilingual Support
Supports both Japanese and English processing, capable of handling Japanese text without pre-tokenization
Large-scale Training
Trained on 2.1 trillion tokens, demonstrating robust language understanding and generation capabilities
Efficient Tokenization
Utilizes a sentencepiece tokenizer based on unigram language models, supporting byte fallback mechanism

Model Capabilities

Japanese Text Generation
English Text Generation
Multi-turn Dialogue
Text Continuation

Use Cases

Content Creation
Article Continuation
Automatically generates subsequent content based on the opening paragraph
Can produce coherent text content
Dialogue System
Building multi-turn dialogue robots
Capable of basic conversational interactions
Education
Language Learning Assistance
Helps Japanese or English learners practice writing
Provides language demonstrations and feedback
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase