đ Kyara: Knowledge Yielding Adaptive Retrieval Augmentation for LLM Fine - tuning
Kyara is an experimental project that uses knowledge retrieval to enhance language models, especially for under - represented languages like Traditional Chinese.
đ Quick Start
Kyara (Knowledge Yielding Adaptive Retrieval Augmentation) is an experimental project aimed at improving language models through knowledge retrieval processes. The project seeks to enhance the model's ability to adapt knowledge and improve language comprehension, particularly in underrepresented languages like Traditional Chinese. Given the relatively scarce availability of Traditional Chinese data compared to the vast corpus of English data used for model training, Kyara addresses this gap by expanding the limited corpus for this language.
This is a preview model, with the stable version set to be released soon.
⨠Features
- Knowledge Retrieval: Improves language models by integrating knowledge retrieval processes.
- Language Adaptation: Enhances the model's ability to adapt knowledge and improve language comprehension, especially for under - represented languages such as Traditional Chinese.
- Corpus Expansion: Addresses the data scarcity issue in Traditional Chinese by expanding its limited corpus.
đ Documentation
Benchmark
All evaluations are conducted in a zero - shot setting.
Property |
Details |
Model Type |
Kyara (Knowledge Yielding Adaptive Retrieval Augmentation) |
Training Data |
Aims to expand the limited corpus of Traditional Chinese data |
Metric |
Kyara - 1b - it |
Llama3.2 - 1b - it |
TMMLUPlus |
31.92 |
30.48 |
â- STEM |
32.56 |
29.74 |
â- Humanities |
30.60 |
29.89 |
â- Other |
31.08 |
30.32 |
â- Social - Science |
33.42 |
31.98 |
MMLU - Redux |
41.40 |
19.62â |
GSM8K |
31.31 |
31.61 |
MATH - L5 |
5.55 |
2.91 |
CRUX |
14 |
11 |
[AlpacaEval](https://github.com/tatsu - lab/alpaca_eval) |
10.79 |
7.39 |
â : Llama3.2 - 1b - it appears to have failed to follow the output schema of ZeroEval on MMLU, with 45.28% of examples lacking answers, which has resulted in a lower MMLU score.
Links

- Hugging Face: [https://huggingface.co/zake7749/Llama - 3.2 - 1B - it - chinese - kyara/](https://huggingface.co/zake7749/Llama - 3.2 - 1B - it - chinese - kyara/)
- Github: https://github.com/zake7749/kyara
- Paper: [https://hf - mirror.com/zake7749/Llama - 3.2 - 1B - it - chinese - kyara/resolve/main/README.md?download=true](https://hf - mirror.com/zake7749/Llama - 3.2 - 1B - it - chinese - kyara/resolve/main/README.md?download=true)
- English: https://github.com/zake7749/kyara/blob/main/document/README_EN.md
- Chinese: https://github.com/zake7749/kyara
- Kaggle Notebook: [https://www.kaggle.com/code/zake7749/kyara - a - compact - yet - powerful - chinese - llm](https://www.kaggle.com/code/zake7749/kyara - a - compact - yet - powerful - chinese - llm)
Image
đ License
The license for this project is llama3.2.