Z

Ziya LLaMA 13B Pretrain V1

Developed by IDEA-CCNL
A large-scale pre-trained model with 13 billion parameters based on the LLaMa architecture, optimized for Chinese tokenization, completing 110 billion tokens of incremental pre-training in Chinese and English, significantly improving Chinese generation and comprehension capabilities
Downloads 113
Release Time : 6/1/2023

Model Overview

Jiang Ziya-LLaMA-13B-Pretrain-v1 is a large-scale pre-trained model with 13 billion parameters based on the LLaMa architecture, optimized for Chinese tokenization, and completing 110 billion tokens of incremental pre-training in Chinese and English, significantly improving Chinese generation and comprehension capabilities.

Model Features

Chinese Optimization
Added 7000+ high-frequency Chinese characters to the original LLaMa vocabulary, significantly improving Chinese processing efficiency
Large-scale Incremental Pre-training
Completed 110 billion tokens of incremental pre-training in Chinese and English, the largest publicly available incremental training for LLaMA-13B models
High-performance Training
Used 160 A100 GPUs with 40GB memory, achieving 118 TFLOP/s per GPU, completing training in just 8 days

Model Capabilities

Text Generation
Text Comprehension
Translation
Programming
Text Classification
Information Extraction
Summary Generation
Copywriting
Common-sense QA
Mathematical Calculation

Use Cases

Content Creation
Travel Plan Generation
Generate detailed travel plans based on user requirements
Produces well-structured and content-rich travel plans
Education
Subject QA
Answer knowledge questions across various academic fields
Outperforms the original LLaMA model in Chinese multi-subject evaluations
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase