Z

Ziya LLaMA 13B V1

Developed by IDEA-CCNL
A 13-billion-parameter pre-trained model based on the LLaMa architecture, capable of translation, programming, text classification, information extraction, summarization, copywriting, commonsense Q&A, and mathematical calculations
Downloads 219
Release Time : 5/16/2023

Model Overview

Ziya General Large Model V1 is a 13-billion-parameter pre-trained model based on the LLaMa architecture, having completed a three-stage training process: large-scale pre-training → multi-task supervised fine-tuning → human feedback reinforcement learning.

Model Features

Multi-stage Training
Undergone a three-stage training process: large-scale pre-training, multi-task supervised fine-tuning, and human feedback reinforcement learning
Vocabulary Optimization
Added 7000+ high-frequency Chinese characters to the original LLaMA vocabulary, constructing a mixed vocabulary with 39,410 dimensions
Curriculum Learning Framework
Adopts a curriculum learning framework, dividing data difficulty levels through model self-assessment for progressive training from easy to difficult
Human Feedback Reinforcement Learning
Utilizes PPO reinforcement learning + RM reward model, combined with hindsight fine-tuning/AI feedback/rule-based reward systems

Model Capabilities

Text generation
Translation
Programming
Text classification
Information extraction
Summarization
Copywriting
Commonsense Q&A
Mathematical calculations

Use Cases

Travel Planning
Travel Guide Generation
Generate city travel guides
Can produce detailed itineraries and attraction recommendations
Programming Assistance
Code Generation
Generate code snippets based on requirements
Can solve programming problems from platforms like LeetCode
Content Creation
Copywriting
Generate marketing copy or articles
Can produce various types of text content as required
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase