Stockmark 13b
Stockmark-13b is a 13-billion-parameter large language model pre-trained from scratch on a corpus of approximately 220 billion Japanese tokens, developed by Stockmark Inc.
Downloads 604
Release Time : 10/21/2023
Model Overview
This is a large language model specialized in Japanese language processing, suitable for natural language processing tasks such as text generation.
Model Features
Large-scale Japanese Pre-training
Trained on a corpus of 220 billion Japanese tokens, focusing on Japanese language processing capabilities
AWS Trainium Support
Supported by AWS's Large Language Model Development Support Program, trained using Trainium accelerators
Quantization Support
Supports 8-bit quantization, can run on GPUs like T4 or V100
Model Capabilities
Japanese text generation
Natural language understanding
Contextual learning
Use Cases
Natural Language Processing
Japanese Text Generation
Generate coherent Japanese text
Can generate coherent text with 128 new tokens
Technical Document Processing
Process technical documents such as patent literature
Featured Recommended AI Models