See our Github repo for more details: https://github.com/hao-ai-lab/Consistency_LLM | |
Metadata: | |
AR loss to consistency loss ratio: 10: 1 | |
shareGPT dataset size: 48k | |
n-token sequence length: 32 | |
Jacobi trajectory data cleaning: True | |
Target model: LLaMA2-7B fine-tuned on ShareGPT48k | |
release date: 02/26/2024 |