This is an experimental llama2 7B lora created using the VNTL-v2-1k dataset. There have been some minor changes in the dataset since version 0.2, and I have made the following adjustments to the training arguments:

  • Model loaded in 8 bits.
  • Sequence length limited to 1024 tokens to speed up experiments.
  • Effective batch size changed to 30 (6 batch size + 5 grad acc).
  • 2 epochs.

Eval Loss: 0.78

This lora was trained alongside a 4-bit lora (qlora), the goal being to see if training a lora would be better than training a qlora. In the end, I don't think that there was much of a difference. At most I saw a consistent 0.01 drop in loss, but the loss graph looked the same, which meant both fine-tunes converged the same way.

This is an prompt example:

<<START>>
Name: Uryuu Shingo (η“œη”Ÿ 新吾) | Gender: Male | Aliases: Onii-chan (γŠε…„γ‘γ‚ƒγ‚“)
Name: Uryuu Sakuno (η“œη”Ÿ ζ‘œδΉƒ) | Gender: Female
<<JAPANESE>>
[ζ‘œδΉƒ]: γ€Žβ€¦β€¦γ”γ‚γ‚“γ€
<<ENGLISH>> (fidelity = absolute)
[Sakuno]: γ€Ž... Sorry.』
<<JAPANESE>>
[新吾]: γ€Œγ†γ†γ‚“γ€γ“γ†θ¨€γ£γ‘γ‚ƒγͺγ‚“γ γ‘γ©γ€θΏ·ε­γ§γ‚ˆγ‹γ£γŸγ‚ˆγ€‚ζ‘œδΉƒγ―ε―ζ„›γ„γ‹γ‚‰γ€γ„γ‚γ„γ‚εΏƒι…γ—γ‘γ‚ƒγ£γ¦γŸγ‚“γ γžδΏΊγ€
<<ENGLISH>> (fidelity = high)

The generated translation for that prompt, with temperature 0, is:

[Shingo]: γ€ŒNo, don't apologize. I'm just glad you're safe. You're so cute, Sakuno, I was worried sick.」
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train lmg-anon/vntl-7b-v0.3-lora