--- base_model: meta-llama/Meta-Llama-3.1-70B-Instruct language: - en - de - fr - it - pt - hi - es - th library_name: transformers license: llama3.1 pipeline_tag: text-generation tags: - facebook - meta - pytorch - llama - llama-3 --- |Weight Quantization| PPL | |-------------------|--------------------| | FP16 | 4.1892 +/- 0.01430 | | IQ_1S | 8.5005 +/- 0.03298 | Dataset used for re-calibration: Mix of [standard_cal_data](https://github.com/turboderp/exllamav2/tree/master/exllamav2/conversion/standard_cal_data) The generated `imatrix` can be downloaded from [imatrix.dat]()