Update README.md
Browse files
README.md
CHANGED
@@ -27,6 +27,8 @@ This model is a <b>causal</b> language model for the <b>Italian</b> language, ba
|
|
27 |
|
28 |
The model has ~6.6B parameters and a vocabulary of 50.335 tokens. It is a foundation model, pre-trained for causal language modeling, so it is mainly suitable for basic natural language generation. It also has some zero-shot and few-shots inference capabilities, but you will have to fine-tune it in order to use it on more specific downstream tasks.
|
29 |
|
|
|
|
|
30 |
The released checkpoint is quantized in 8-bit, so that it can easily be loaded and used for training and inference on ordinary hardware, and it requires the installation of the <b>transformers</b> library version >= 4.30.1 and the <b>bitsandbytes</b> library, version >= 0.37.2
|
31 |
|
32 |
On Windows operating systems, the <b>bitsandbytes-windows</b> module also needs to be installed on top. However, it appears that the module is not yet updated with some recent features, like the possibility to save the 8-bit quantized models.
|
|
|
27 |
|
28 |
The model has ~6.6B parameters and a vocabulary of 50.335 tokens. It is a foundation model, pre-trained for causal language modeling, so it is mainly suitable for basic natural language generation. It also has some zero-shot and few-shots inference capabilities, but you will have to fine-tune it in order to use it on more specific downstream tasks.
|
29 |
|
30 |
+
<h3>Quantization</h3>
|
31 |
+
|
32 |
The released checkpoint is quantized in 8-bit, so that it can easily be loaded and used for training and inference on ordinary hardware, and it requires the installation of the <b>transformers</b> library version >= 4.30.1 and the <b>bitsandbytes</b> library, version >= 0.37.2
|
33 |
|
34 |
On Windows operating systems, the <b>bitsandbytes-windows</b> module also needs to be installed on top. However, it appears that the module is not yet updated with some recent features, like the possibility to save the 8-bit quantized models.
|