|
--- |
|
license: apache-2.0 |
|
base_model: fal/AuraFlow-v0.3 |
|
base_model_relation: quantized |
|
--- |
|
|
|
|
|
FP8 quantized version of [AuraFlow v0.3](fal/AuraFlow-v0.3) |
|
|
|
## Quantization |
|
|
|
```py |
|
import torch |
|
from huggingface_hub import cached_download |
|
from safetensors.torch import load_file, save_file |
|
|
|
ckpt_path = cached_download( |
|
"https://huggingface.co/fal/AuraFlow-v0.3/resolve/main/aura_flow_0.3.safetensors", |
|
) |
|
|
|
state_dict = load_file(ckpt_path) |
|
|
|
for key, value in state_dict.items(): |
|
state_dict[key] = value.to(torch.float8_e4m3fn) |
|
|
|
save_file(state_dict, "./aura_flow_0.3.float8_e4m3fn.safetensors") |
|
``` |
|
|