Llama-3-Taiwan-70B-Instruct-fp8
- Model creator: Yen-Ting Lin
- Original model: Llama-3-Taiwan-70B-Instruct
Description
This repo contains fp8 model files for Llama-3-Taiwan-70B-Instruct.
Quantization parameter
- activation_scheme : static
- quant_method : fp8
- ignored_layers : lm_head
It tooks about 8.5 hrs to quantize on H100.
- Downloads last month
- 4
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for minyichen/Llama-3-Taiwan-70B-Instruct-fp8
Base model
meta-llama/Meta-Llama-3-70B
Finetuned
yentinglin/Llama-3-Taiwan-70B-Instruct