๐ imatrix quants of lmg-anon/vntl-llama3-8b-v2-hf
This project uses a multilingual fork of Bartowski's imatrix dataset to perform imatrix quants on lmg-anon/vntl-llama3-8b-v2-hf, aiming to improve the performance of LLMs in translating Japanese visual novels to English.
๐ Quick Start
This is a LLaMA 3 Youko qlora fine - tune, created using a new version of the VNTL dataset. The goal is to enhance the performance of large language models (LLMs) in translating Japanese visual novels to English.
โจ Features
- Improved Performance: Rebuilt and expanded the VNTL dataset from scratch. It shows better accuracy and stability compared to the previous version, making far fewer mistakes even at high temperatures.
- Prompt Format Update: Switched to the default LLaMA3 prompt format, addressing the issues users had with the custom one.
- Multi - line Translation Support: Added proper support for multi - line translations, while the old version only handled single lines.
- Higher Translation Accuracy: Overall, it provides better translation accuracy, though the translations are more literal compared to the previous version.
๐ง Technical Details
Notes
For this new version of VNTL 8B, the VNTL's dataset has been rebuilt and expanded. It performs well, outperforming the previous version in accuracy and stability. It makes fewer mistakes even at high temperatures, but it's still recommended to use temperature 0 for the best accuracy.
Sampling Recommendations
For optimal results, it's highly recommended to use neutral sampling parameters (temperature 0 with no repetition penalty) when using this model.
Training Details
This fine - tune was done using similar hyperparameters as the previous version. The only difference is the dataset, which is brand - new.
Property |
Details |
Rank |
128 |
Alpha |
32 |
Effective Batch Size |
45 |
Warmup Ratio |
0.02 |
Learning Rate |
6e - 5 |
Embedding Learning Rate |
1e - 5 |
Optimizer |
grokadamw |
LR Schedule |
cosine |
Weight Decay |
0.01 |
Train Loss |
0.42 |
Translation Prompt
This fine - tune uses the LLaMA 3 prompt format. Here is an example prompt for translation:
<|begin_of_text|><|start_header_id|>Metadata<|end_header_id|>
[character] Name: Uryuu Shingo (็็ ๆฐๅพ) | Gender: Male | Aliases: Onii-chan (ใๅ
ใกใใ)
[character] Name: Uryuu Sakuno (็็ ๆกไน) | Gender: Female<|eot_id|><|start_header_id|>Japanese<|end_header_id|>
[ๆกไน]: ใโฆโฆใใใใ<|eot_id|><|start_header_id|>English<|end_header_id|>
[Sakuno]: ใ... Sorry.ใ<|eot_id|><|start_header_id|>Japanese<|end_header_id|>
[ๆฐๅพ]: ใใใใใใใ่จใฃใกใใชใใ ใใฉใ่ฟทๅญใงใใใฃใใใๆกไนใฏๅฏๆใใใใใใใใๅฟ้
ใใกใใฃใฆใใใ ใไฟบใ<|eot_id|><|start_header_id|>English<|end_header_id|>
[Shingo]: "Nah, I know itโs weird to say this, but Iโm glad you got lost. Youโre so cute, Sakuno, so I was really worried about you."<|eot_id|>
The generated translation for that prompt, with temperature 0, is:
[Shingo]: "Nah, I know itโs weird to say this, but Iโm glad you got lost. Youโre so cute, Sakuno, so I was really worried about you."
Trivia
The Metadata section isn't limited to character information. You can also add trivia and teach the model the correct way to pronounce words it struggles with.
Here's an example:
<|begin_of_text|><|start_header_id|>Metadata<|end_header_id|>
[character] Name: Uryuu Shingo (็็ ๆฐๅพ) | Gender: Male | Aliases: Onii-chan (ใๅ
ใกใใ)
[character] Name: Uryuu Sakuno (็็ ๆกไน) | Gender: Female
[element] Name: Murasamemaru (ๅข้จไธธ) | Type: Quality<|eot_id|><|start_header_id|>Japanese<|end_header_id|>
[ๆกไน]: ใโฆโฆใใใใ<|eot_id|><|start_header_id|>English<|end_header_id|>
[Sakuno]: ใ... Sorry.ใ<|eot_id|><|start_header_id|>Japanese<|end_header_id|>
[ๆฐๅพ]: ใใใใใใใ่จใฃใกใใชใใ ใใฉใ่ฟทๅญใงใใใฃใใใๆกไนใฏๅข้จไธธใใใใใใใใๅฟ้
ใใกใใฃใฆใใใ ใไฟบใ<|eot_id|><|start_header_id|>English<|end_header_id|>
The generated translation for that prompt, with temperature 0, is:
[Shingo]: "Nah, I know itโs not the best thing to say, but Iโm glad you got lost. Sakunoโs Murasamemaru, so I was really worried about you, you know?"