L

Llama 3.3 Swallow 70B V0.4

Developed by tokyotech-llm
Llama 3.3 Swallow is a large language model with 70 billion parameters, built on Meta Llama 3.3, which enhances Japanese capabilities while retaining English capabilities.
Downloads 1,950
Release Time : 2/17/2025

Model Overview

Llama 3.3 Swallow is built by continuous pre - training on the Meta Llama 3.3 model, using large - scale Japanese and English corpora to enhance bilingual processing capabilities.

Model Features

Enhanced bilingual capabilities
On the basis of retaining English capabilities, the Japanese processing capabilities are significantly improved.
Continuous pre - training
Use large - scale and diverse corpora for continuous pre - training to improve model performance.
Multiple model variants
Provide models in different versions and specifications to meet different needs.

Model Capabilities

Japanese text generation
English text generation
Machine translation
Question - answering system
Code generation
Mathematical reasoning

Use Cases

Education
Japanese learning assistance
Help students understand and generate Japanese texts.
Performs excellently in Japanese evaluation benchmarks.
Translation
Japanese - English and English - Japanese translation
Used for Japanese - English and English - Japanese machine translation tasks.
Performs well in the WMT20 translation task.
Programming
Code generation
Generate code that complies with the Google Python Style Guide.
Performs excellently in the JHumanEval evaluation.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase