This is an exl2 quant of lmg-anon/vntl-llama3-8b-qlora at 8.0 bpw. The measurements.json is provided. This was quanted against the default exl2 calibration dataset at 2k tokens.
Summary
This is the merge of the VNTL LLaMA 3 8B qlora created using a slightly modified version of the VNTL-v3.1-1k dataset, concatenated with the VNTL-Chat dataset.
This also includes a new "chat mode", which works mostly for Japanese grammar questions.
Translation Prompt
This is an prompt example for translation:
<<METADATA>>
[character] Name: Uryuu Shingo (ηη ζ°εΎ) | Gender: Male | Aliases: Onii-chan (γε
γ‘γγ)
[character] Name: Uryuu Sakuno (ηη ζ‘δΉ) | Gender: Female
<<TRANSLATE>>
<<JAPANESE>>
[ζ‘δΉ]: γβ¦β¦γγγγ
<<ENGLISH>>
[Sakuno]: γ... Sorry.γ<|end_of_text|>
<<JAPANESE>>
[ζ°εΎ]: γγγγγγγθ¨γ£γ‘γγͺγγ γγ©γθΏ·εγ§γγγ£γγγζ‘δΉγ―ε―ζγγγγγγγγεΏι
γγ‘γγ£γ¦γγγ γδΏΊγ
<<ENGLISH>>
The generated translation for that prompt, with temperature 0, is:
[Shingo]: γNo, I know this isn't the best way to say it, but I'm glad you got lost. You're so cute that I was worried about you.γ
Chat Prompt
This is an prompt example for chat:
<<CHAT>>
<<HUMAN>>
Hi! Can you break this down to me? Like, what is the meaning of "ζ‘δΉγ―ε―ζγγγ"?
"γγγγγγθ¨γ£γ‘γγͺγγ γγ©γθΏ·εγ§γγγ£γγγζ‘δΉγ―ε―ζγγγγγγγγεΏι
γγ‘γγ£γ¦γγγ γδΏΊγ"
<<LLM>>
The generated response for that prompt, with temperature 0, is:
Certainly! The phrase "ζ‘δΉγ―ε―ζγγγ" literally translates to "Because Sakura-no is cute." In Japanese culture, there is a strong emphasis on physical attractiveness and cuteness, especially when it comes to young girls or women. This phrase suggests that the speaker finds Sakura-no (presumably a girl) to be very attractive and endearing in appearance.
- Downloads last month
- 0
Inference API (serverless) does not yet support peft models for this pipeline type.