stefanoscotta commited on
Commit
ebd6241
1 Parent(s): 366cefc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -24,7 +24,7 @@ An open-source LLaMa2 language model of 7b parameters fine-tuned (using as base
24
  This model is a LLM of 7b parameters based on [NousResearch/Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b), a version of [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b) fine-tuned to follow instructions.
25
  The model was further fine-tuned in order to follow instructions in italian, using [LoRA](https://arxiv.org/abs/2106.09685) approach and a dataset of 120k random pairs of instruction/answer from [raicrits/Orca_ITA_200k](https://huggingface.co/datasets/raicrits/Orca_ITA_200k).
26
 
27
- This repository contains the model weights merged with the LoRA adapters obtained in the fine-tuning procedure (in "float32").
28
 
29
 
30
  - **Developed by:** Stefano Scotta (stefano.scotta@rai.it)
@@ -77,7 +77,7 @@ model_name = "raicrits/Hermes7b_ITA_v1"
77
  model = LlamaForCausalLM.from_pretrained(
78
  model_name,
79
  device_map="auto",
80
- # torch_dtype=torch.bfloat16 #if you want to load quantized model to save GPU memory (it gets only a bit slower)
81
  )
82
 
83
  tokenizer = AutoTokenizer.from_pretrained("Hermes_ITA_Lora_merged_V2", add_eos_token=False)
@@ -113,6 +113,8 @@ The fine-tuning procedure was done using [LoRA](https://arxiv.org/abs/2106.09685
113
 
114
  - learning_rate=2e-4,
115
 
 
 
116
 
117
  **LoRA configuration:**
118
  - r= 8
 
24
  This model is a LLM of 7b parameters based on [NousResearch/Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b), a version of [meta-llama/Llama-2-7b](https://huggingface.co/meta-llama/Llama-2-7b) fine-tuned to follow instructions.
25
  The model was further fine-tuned in order to follow instructions in italian, using [LoRA](https://arxiv.org/abs/2106.09685) approach and a dataset of 120k random pairs of instruction/answer from [raicrits/Orca_ITA_200k](https://huggingface.co/datasets/raicrits/Orca_ITA_200k).
26
 
27
+ This repository contains the model weights merged with the LoRA adapters obtained in the fine-tuning procedure.
28
 
29
 
30
  - **Developed by:** Stefano Scotta (stefano.scotta@rai.it)
 
77
  model = LlamaForCausalLM.from_pretrained(
78
  model_name,
79
  device_map="auto",
80
+ torch_dtype=torch.bfloat16
81
  )
82
 
83
  tokenizer = AutoTokenizer.from_pretrained("Hermes_ITA_Lora_merged_V2", add_eos_token=False)
 
113
 
114
  - learning_rate=2e-4,
115
 
116
+ - mixed precision training: float16
117
+
118
 
119
  **LoRA configuration:**
120
  - r= 8