๐ ktdsbaseLM-v0.14-onbased-llama3.1
This model is based on LlaMA3.1 and is developed to be applicable to Korean and various Korean cultures. It uses self - produced Korean data from 53 domains to understand Korean social values and cultures.
โจ Features
- Multifunctional: Supports various tasks such as text generation, dialogue inference, document summarization, question - answering, sentiment analysis, and other natural language processing - related tasks. It can be applied in diverse fields like law, finance, science, education, business, and cultural research.
- Korean - specific: Designed to understand the Korean language and various Korean cultural contexts. It reflects the values and cultures of Korean society by leveraging self - produced Korean data from 53 domains.
- Efficient Architecture: Based on the LlaMA3.1 8B model with 8 billion parameters. Its lightweight architecture ensures fast inference speed and memory efficiency, and is optimized for various natural language processing tasks.
๐ฆ Installation
No specific installation steps are provided in the original document.
๐ Documentation
Model Information
Property |
Details |
Base Model |
meta - llama/Llama - 3.1 - 8B - Instruct |
Datasets |
AIDX - ktds/ko_leaderboard |
Language |
ko |
License |
apache - 2.0 |
Metrics |
accuracy |
Pipeline Tag |
text - generation |
Tags |
ko, leaderboard, ktds, llama3.1 |
Model Description
This model is fine - tuned using the SFT method based on the LlaMA3.1 model. It is designed to understand the Korean language and various Korean cultural contexts, and reflects the values and cultures of Korean society by using self - produced Korean data from 53 domains.
Training Data
๐ป Usage Examples
Basic Usage
The model can be used in various application scenarios:
- Education: Generate question - answers and explanations for various learning materials in history, mathematics, science, etc.
- Business: Provide answers to legal, financial, and tax - related questions and summarize documents.
- Research and Culture: Perform natural language processing tasks, sentiment analysis, document generation, and translation in line with Korean society and culture.
- Customer Service: Generate conversations with users and provide customized responses.
Code Example
from transformers import AutoModel, AutoTokenizer
import torch
tokenizer = AutoTokenizer.from_pretrained("AIDX-ktds/ktdsbaseLM-v0.14-onbased-llama3.1")
model = AutoModel.from_pretrained("AIDX-ktds/ktdsbaseLM-v0.14-onbased-llama3.1")
input_text = """ ใ๊ตญ๋ฏผ๊ฑด๊ฐ๋ณดํ๋ฒใ์ 44์กฐ, ใ๊ตญ๋ฏผ๊ฑด๊ฐ๋ณดํ๋ฒ ์ํ๋ นใ์ 19์กฐ,ใ์ฝ๊ด์ ๊ท์ ์ ๊ดํ ๋ฒ๋ฅ ใ์ 5์กฐ, ใ์๋ฒใ์ 54์กฐ ์ฐธ์กฐ ํ๋จ ํด์ค""" + " ๋ต๋ณ:"
inputs = tokenizer(input_text, return_tensors="pt")
with torch.no_grad():
outputs = model.generate(**inputs, max_length=1024, temperature=0.5, do_sample=True, repetition_penalty=1.15)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)
๐ง Technical Details
The model is based on the LlaMA3.1 8B model, with 8 billion parameters. It uses the SFT method for fine - tuning, which is optimized for Korean language and culture - specific tasks. The lightweight architecture of LlaMA3.1 8B ensures fast inference speed and memory efficiency, making it suitable for a variety of natural language processing tasks.
โ ๏ธ Limitations
- Since the model is specialized in the Korean language and Korean culture, the accuracy of responses in other languages or cultures may be affected due to the lack of data in specific areas (e.g., the latest international materials, professional fields).
- It may show limited inference ability for problems that require complex logical thinking, and there is a possibility of generating biased responses if the training data contains biases.
๐ License
The model is released under the apache - 2.0 license.