You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Kyara: Knowledge Yielding Adaptive Retrieval Augmentation for LLM Fine-tuning

DOI

πŸ€— Hugging Face  | πŸš€Github  |  πŸ“‘ Paper  |  πŸ“– English  |  πŸ“– Chinese  |  πŸ’» Kaggle Notebook

kyara

Kyara (Knowledge Yielding Adaptive Retrieval Augmentation) is an experimental project aimed at improving language models through knowledge retrieval processes. The project seeks to enhance the model’s ability to adapt knowledge and improve language comprehension, particularly in underrepresented languages like Traditional Chinese. Given the relatively scarce availability of Traditional Chinese data compared to the vast corpus of English data used for model training, Kyara addresses this gap by expanding the limited corpus for this language.

This is a preview model, with the stable version set to be released soon.

Benchmark

All evaluations are conducted in a zero-shot setting.

Metric Kyara-9b-it Gemma-2-9b-it
TMMLUPlus 59.87 54.77
 - STEM 66.98 58.12
 - Humanities 54.42 48.71
 - Other 55.26 51.43
 - Social-Science 62.81 60.84
MMLU-Redux 72.57 72.82
GSM8K 90.60 87.41
MATH-L5 28.71 19.42
CRUX 48.62 46.00
MT-Bench 8.81 8.53
MT-Bench-TW 8.36 7.80
Chatbot-Arena-Hard 41.1 33.6
Downloads last month
11
Safetensors
Model size
9.24B params
Tensor type
BF16
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for zake7749/gemma-2-9b-it-chinese-kyara-preview

Base model

google/gemma-2-9b
Finetuned
(120)
this model

Collection including zake7749/gemma-2-9b-it-chinese-kyara-preview