osiria commited on
Commit
3c42e74
1 Parent(s): 22b2c91

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -27,6 +27,8 @@ This model is a <b>causal</b> language model for the <b>Italian</b> language, ba
27
 
28
  The model has ~6.6B parameters and a vocabulary of 50.335 tokens. It is a foundation model, pre-trained for causal language modeling, so it is mainly suitable for basic natural language generation. It also has some zero-shot and few-shots inference capabilities, but you will have to fine-tune it in order to use it on more specific downstream tasks.
29
 
 
 
30
  The released checkpoint is quantized in 8-bit, so that it can easily be loaded and used for training and inference on ordinary hardware, and it requires the installation of the <b>transformers</b> library version >= 4.30.1 and the <b>bitsandbytes</b> library, version >= 0.37.2
31
 
32
  On Windows operating systems, the <b>bitsandbytes-windows</b> module also needs to be installed on top. However, it appears that the module is not yet updated with some recent features, like the possibility to save the 8-bit quantized models.
 
27
 
28
  The model has ~6.6B parameters and a vocabulary of 50.335 tokens. It is a foundation model, pre-trained for causal language modeling, so it is mainly suitable for basic natural language generation. It also has some zero-shot and few-shots inference capabilities, but you will have to fine-tune it in order to use it on more specific downstream tasks.
29
 
30
+ <h3>Quantization</h3>
31
+
32
  The released checkpoint is quantized in 8-bit, so that it can easily be loaded and used for training and inference on ordinary hardware, and it requires the installation of the <b>transformers</b> library version >= 4.30.1 and the <b>bitsandbytes</b> library, version >= 0.37.2
33
 
34
  On Windows operating systems, the <b>bitsandbytes-windows</b> module also needs to be installed on top. However, it appears that the module is not yet updated with some recent features, like the possibility to save the 8-bit quantized models.