thwin27 commited on
Commit
e4dcfa0
1 Parent(s): 50f77df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -8,7 +8,8 @@ library_name: transformers
8
  ---
9
  # Aria-sequential_mlp-FP8-dynamic
10
  FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
11
- ### Installationcompressed-tensors
 
12
  ```
13
  pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
14
  pip install flash-attn --no-build-isolation
 
8
  ---
9
  # Aria-sequential_mlp-FP8-dynamic
10
  FP8-Dynamic quantization from [Aria-sequential_mlp](https://huggingface.co/rhymes-ai/Aria-sequential_mlp) made with [llm-compressor](https://github.com/vllm-project/llm-compressor), requires about 24.8 GB of VRAM.
11
+
12
+ ### Installation
13
  ```
14
  pip install transformers==4.45.0 accelerate==0.34.1 sentencepiece==0.2.0 torchvision requests torch Pillow compressed-tensors
15
  pip install flash-attn --no-build-isolation