S

Stockmark 13b

Developed by stockmark
Stockmark-13b is a 13-billion-parameter large language model pre-trained from scratch on a corpus of approximately 220 billion Japanese tokens, developed by Stockmark Inc.
Downloads 604
Release Time : 10/21/2023

Model Overview

This is a large language model specialized in Japanese language processing, suitable for natural language processing tasks such as text generation.

Model Features

Large-scale Japanese Pre-training
Trained on a corpus of 220 billion Japanese tokens, focusing on Japanese language processing capabilities
AWS Trainium Support
Supported by AWS's Large Language Model Development Support Program, trained using Trainium accelerators
Quantization Support
Supports 8-bit quantization, can run on GPUs like T4 or V100

Model Capabilities

Japanese text generation
Natural language understanding
Contextual learning

Use Cases

Natural Language Processing
Japanese Text Generation
Generate coherent Japanese text
Can generate coherent text with 128 new tokens
Technical Document Processing
Process technical documents such as patent literature
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase