dawoz commited on
Commit
9c6e528
1 Parent(s): d29fa83

Training in progress, step 120

Browse files
README.md CHANGED
@@ -1,19 +1,20 @@
1
  ---
2
- base_model: HuggingFaceM4/idefics2-8b
3
- license: apache-2.0
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: idefics2-frozenlake
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # idefics2-frozenlake
15
 
16
- This model is a fine-tuned version of [HuggingFaceM4/idefics2-8b](https://huggingface.co/HuggingFaceM4/idefics2-8b) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -50,7 +51,8 @@ The following hyperparameters were used during training:
50
 
51
  ### Framework versions
52
 
53
- - Transformers 4.41.2
54
- - Pytorch 2.1.1+cu121
55
- - Datasets 2.20.0
56
- - Tokenizers 0.19.1
 
 
1
  ---
2
+ base_model: HuggingFaceM4/idefics-9b-instruct
3
+ library_name: peft
4
+ license: other
5
  tags:
6
  - generated_from_trainer
7
  model-index:
8
+ - name: idefics-frozenlake
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # idefics-frozenlake
16
 
17
+ This model is a fine-tuned version of [HuggingFaceM4/idefics-9b-instruct](https://huggingface.co/HuggingFaceM4/idefics-9b-instruct) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
51
 
52
  ### Framework versions
53
 
54
+ - PEFT 0.12.0
55
+ - Transformers 4.44.2
56
+ - Pytorch 2.4.0+cu121
57
+ - Datasets 2.16.1
58
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,10 +1,13 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
4
  "base_model_name_or_path": "HuggingFaceM4/idefics2-8b",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
- "inference_mode": false,
8
  "init_lora_weights": "gaussian",
9
  "layer_replication": null,
10
  "layers_pattern": null,
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "Idefics2ForConditionalGeneration",
5
+ "parent_library": "transformers.models.idefics2.modeling_idefics2"
6
+ },
7
  "base_model_name_or_path": "HuggingFaceM4/idefics2-8b",
8
  "bias": "none",
9
  "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
  "init_lora_weights": "gaussian",
12
  "layer_replication": null,
13
  "layers_pattern": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87e79329754ab9a2d72ff37115cd2032d32033eb572e10a71f63cbf5838520a6
3
  size 93378688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9839be56cb740b79528ed01d5549e2d47fc920db3998b041e46a73673beca51
3
  size 93378688
generation_config.json CHANGED
@@ -1,18 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
- "bad_words_ids": [
4
- [
5
- 32000
6
- ],
7
- [
8
- 32001
9
- ]
10
- ],
11
  "bos_token_id": 1,
12
- "eos_token_id": [
13
- 2,
14
- 32002
15
- ],
16
  "pad_token_id": 0,
17
- "transformers_version": "4.41.2"
18
  }
 
1
  {
2
  "_from_model_config": true,
 
 
 
 
 
 
 
 
3
  "bos_token_id": 1,
4
+ "eos_token_id": 2,
 
 
 
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.44.2"
7
  }
runs/Jul31_15-43-03_ananke/events.out.tfevents.1722433386.ananke.31988.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ce75119eb99c0c325d3ed007177c942413a2d4746f44ca84de99113746fcbb0
3
+ size 5947
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dff79ecc78535c10e2728a33beecbd0699a59c812d86d07b89fe50048982f643
3
- size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b43490a3fe649c6b004f00482289139911ff811809a2541d2f69c7a92cd4b84
3
+ size 5176