Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,8 @@ library_name: transformers
|
|
8 |
---
|
9 |
# Aria-sequential_mlp-FP8-dynamic
|
10 |
FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
|
11 |
-
|
|
|
12 |
```
|
13 |
pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
|
14 |
pip install flash-attn --no-build-isolation
|
|
|
8 |
---
|
9 |
# Aria-sequential_mlp-FP8-dynamic
|
10 |
FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
|
11 |
+
|
12 |
+
### Installation
|
13 |
```
|
14 |
pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
|
15 |
pip install flash-attn --no-build-isolation
|