N

Nekomata 14b

Developed by rinna
A large language model continuously pre-trained on a mixed Japanese and English dataset based on Qwen-14B, significantly improving performance on Japanese tasks
Downloads 705
Release Time : 12/19/2023

Model Overview

This model is obtained by continuously pre-training Qwen-14B on a mixed Japanese and English corpus of 66 billion tokens, specifically optimized for Japanese processing while retaining the excellent features of the original Qwen model.

Model Features

Efficient Japanese Processing
Uses Qwen vocabulary (vocabulary size >150k), with higher efficiency in Japanese text processing compared to the youri series
Long Context Support
Supports a maximum sequence length of 8192
Multilingual Capability
Trained on mixed Japanese and English corpus, capable of bilingual processing
High-Performance Tokenization
Uses an extended cl100k tiktoken tokenizer, significantly outperforming the llama-2 tokenizer in Japanese tokenization efficiency

Model Capabilities

Japanese Text Generation
English Text Generation
Long Text Processing

Use Cases

Content Creation
Japanese Article Continuation
Automatically generates coherent Japanese articles based on the opening text
Example demonstrates the continuation effect starting with 'Nishida Kitaro wa,'
Language Learning
Bilingual Text Generation
Generates bilingual content in Japanese and English
Featured Recommended AI Models
ยฉ 2025AIbase