Morgan McGuire Morgan McGuire winglian commited on
Commit
7019509
1 Parent(s): 96bd6ae

Add wandb_entity to wandb options, update example configs, update README (#361)

Browse files

* Update wandb_entity and add wandb descriptions

* add wandb to config section

* remove trailing whitespace for pre-commit hook

* remove trailing whitespace for pre-commit hook

---------

Co-authored-by: Morgan McGuire <morganmcguire@Morgans-MacBook-Pro.local>
Co-authored-by: Wing Lian <wing.lian@gmail.com>

README.md CHANGED
@@ -401,11 +401,12 @@ lora_out_dir:
401
  lora_fan_in_fan_out: false
402
 
403
  # wandb configuration if you're using it
404
- wandb_mode:
405
- wandb_project:
 
406
  wandb_watch:
407
- wandb_run_id:
408
- wandb_log_model: # 'checkpoint'
409
 
410
  # where to save the finished model to
411
  output_dir: ./completed-model
@@ -553,6 +554,18 @@ fsdp_config:
553
 
554
  - llama Deepspeed: append `ACCELERATE_USE_DEEPSPEED=true` in front of finetune command
555
 
 
 
 
 
 
 
 
 
 
 
 
 
556
  ### Inference
557
 
558
  Pass the appropriate flag to the train command:
 
401
  lora_fan_in_fan_out: false
402
 
403
  # wandb configuration if you're using it
404
+ wandb_mode: # "offline" to save run metadata locally and not sync to the server, "disabled" to turn off wandb
405
+ wandb_project: # your wandb project name
406
+ wandb_entity: # a wandb Team name if using a Team
407
  wandb_watch:
408
+ wandb_run_id: # set the name of your wandb run
409
+ wandb_log_model: # "checkpoint" to log model to wandb Artifacts every `save_steps` or "end" to log only at the end of training
410
 
411
  # where to save the finished model to
412
  output_dir: ./completed-model
 
554
 
555
  - llama Deepspeed: append `ACCELERATE_USE_DEEPSPEED=true` in front of finetune command
556
 
557
+ ##### Weights & Biases Logging
558
+
559
+ - wandb options
560
+ ```yaml
561
+ wandb_mode:
562
+ wandb_project:
563
+ wandb_entity:
564
+ wandb_watch:
565
+ wandb_run_id:
566
+ wandb_log_model:
567
+ ```
568
+
569
  ### Inference
570
 
571
  Pass the appropriate flag to the train command:
examples/cerebras/qlora.yml CHANGED
@@ -23,6 +23,7 @@ lora_target_modules:
23
  lora_target_linear:
24
  lora_fan_in_fan_out:
25
  wandb_project:
 
26
  wandb_watch:
27
  wandb_run_id:
28
  wandb_log_model:
 
23
  lora_target_linear:
24
  lora_fan_in_fan_out:
25
  wandb_project:
26
+ wandb_entity:
27
  wandb_watch:
28
  wandb_run_id:
29
  wandb_log_model:
examples/falcon/config-7b-lora.yml CHANGED
@@ -24,6 +24,7 @@ lora_target_modules:
24
  lora_target_linear: true
25
  lora_fan_in_fan_out:
26
  wandb_project:
 
27
  wandb_watch:
28
  wandb_run_id:
29
  wandb_log_model:
 
24
  lora_target_linear: true
25
  lora_fan_in_fan_out:
26
  wandb_project:
27
+ wandb_entity:
28
  wandb_watch:
29
  wandb_run_id:
30
  wandb_log_model:
examples/falcon/config-7b-qlora.yml CHANGED
@@ -38,6 +38,7 @@ lora_target_linear: true
38
  lora_fan_in_fan_out:
39
 
40
  wandb_project:
 
41
  wandb_watch:
42
  wandb_run_id:
43
  wandb_log_model:
 
38
  lora_fan_in_fan_out:
39
 
40
  wandb_project:
41
+ wandb_entity:
42
  wandb_watch:
43
  wandb_run_id:
44
  wandb_log_model:
examples/falcon/config-7b.yml CHANGED
@@ -24,6 +24,7 @@ lora_target_modules:
24
  lora_target_linear: true
25
  lora_fan_in_fan_out:
26
  wandb_project:
 
27
  wandb_watch:
28
  wandb_run_id:
29
  wandb_log_model:
 
24
  lora_target_linear: true
25
  lora_fan_in_fan_out:
26
  wandb_project:
27
+ wandb_entity:
28
  wandb_watch:
29
  wandb_run_id:
30
  wandb_log_model:
examples/gptj/qlora.yml CHANGED
@@ -20,6 +20,7 @@ lora_target_modules:
20
  lora_target_linear: true
21
  lora_fan_in_fan_out:
22
  wandb_project:
 
23
  wandb_watch:
24
  wandb_run_id:
25
  wandb_log_model:
 
20
  lora_target_linear: true
21
  lora_fan_in_fan_out:
22
  wandb_project:
23
+ wandb_entity:
24
  wandb_watch:
25
  wandb_run_id:
26
  wandb_log_model:
examples/gptq-lora-7b/config.yml CHANGED
@@ -22,6 +22,7 @@ lora_target_modules:
22
  - v_proj
23
  lora_fan_in_fan_out: false
24
  wandb_project: llama-7b-lora-int4
 
25
  wandb_watch:
26
  wandb_run_id:
27
  wandb_log_model:
 
22
  - v_proj
23
  lora_fan_in_fan_out: false
24
  wandb_project: llama-7b-lora-int4
25
+ wandb_entity:
26
  wandb_watch:
27
  wandb_run_id:
28
  wandb_log_model:
examples/jeopardy-bot/config.yml CHANGED
@@ -18,6 +18,7 @@ lora_dropout:
18
  lora_target_modules:
19
  lora_fan_in_fan_out: false
20
  wandb_project:
 
21
  wandb_watch:
22
  wandb_run_id:
23
  wandb_log_model:
 
18
  lora_target_modules:
19
  lora_fan_in_fan_out: false
20
  wandb_project:
21
+ wandb_entity:
22
  wandb_watch:
23
  wandb_run_id:
24
  wandb_log_model:
examples/llama-2/lora.yml CHANGED
@@ -26,6 +26,7 @@ lora_target_linear: true
26
  lora_fan_in_fan_out:
27
 
28
  wandb_project:
 
29
  wandb_watch:
30
  wandb_run_id:
31
  wandb_log_model:
 
26
  lora_fan_in_fan_out:
27
 
28
  wandb_project:
29
+ wandb_entity:
30
  wandb_watch:
31
  wandb_run_id:
32
  wandb_log_model:
examples/llama-2/qlora.yml CHANGED
@@ -27,6 +27,7 @@ lora_target_linear: true
27
  lora_fan_in_fan_out:
28
 
29
  wandb_project:
 
30
  wandb_watch:
31
  wandb_run_id:
32
  wandb_log_model:
 
27
  lora_fan_in_fan_out:
28
 
29
  wandb_project:
30
+ wandb_entity:
31
  wandb_watch:
32
  wandb_run_id:
33
  wandb_log_model:
examples/mpt-7b/config.yml CHANGED
@@ -20,6 +20,7 @@ lora_target_modules:
20
  - v_proj
21
  lora_fan_in_fan_out: false
22
  wandb_project: mpt-alpaca-7b
 
23
  wandb_watch:
24
  wandb_run_id:
25
  wandb_log_model:
 
20
  - v_proj
21
  lora_fan_in_fan_out: false
22
  wandb_project: mpt-alpaca-7b
23
+ wandb_entity:
24
  wandb_watch:
25
  wandb_run_id:
26
  wandb_log_model:
examples/openllama-3b/config.yml CHANGED
@@ -22,6 +22,7 @@ lora_target_modules:
22
  lora_target_linear:
23
  lora_fan_in_fan_out:
24
  wandb_project:
 
25
  wandb_watch:
26
  wandb_run_id:
27
  wandb_log_model:
 
22
  lora_target_linear:
23
  lora_fan_in_fan_out:
24
  wandb_project:
25
+ wandb_entity:
26
  wandb_watch:
27
  wandb_run_id:
28
  wandb_log_model:
examples/openllama-3b/lora.yml CHANGED
@@ -28,6 +28,7 @@ lora_target_modules:
28
  - o_proj
29
  lora_fan_in_fan_out:
30
  wandb_project:
 
31
  wandb_watch:
32
  wandb_run_id:
33
  wandb_log_model:
 
28
  - o_proj
29
  lora_fan_in_fan_out:
30
  wandb_project:
31
+ wandb_entity:
32
  wandb_watch:
33
  wandb_run_id:
34
  wandb_log_model:
examples/openllama-3b/qlora.yml CHANGED
@@ -22,6 +22,7 @@ lora_target_modules:
22
  lora_target_linear: true
23
  lora_fan_in_fan_out:
24
  wandb_project:
 
25
  wandb_watch:
26
  wandb_run_id:
27
  wandb_log_model:
 
22
  lora_target_linear: true
23
  lora_fan_in_fan_out:
24
  wandb_project:
25
+ wandb_entity:
26
  wandb_watch:
27
  wandb_run_id:
28
  wandb_log_model:
examples/pythia-12b/config.yml CHANGED
@@ -23,6 +23,7 @@ lora_target_modules:
23
  lora_target_linear: true
24
  lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
25
  wandb_project:
 
26
  wandb_watch:
27
  wandb_run_id:
28
  wandb_log_model:
 
23
  lora_target_linear: true
24
  lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
25
  wandb_project:
26
+ wandb_entity:
27
  wandb_watch:
28
  wandb_run_id:
29
  wandb_log_model:
examples/pythia/lora.yml CHANGED
@@ -17,6 +17,7 @@ lora_target_modules:
17
  lora_target_linear:
18
  lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
19
  wandb_project:
 
20
  wandb_watch:
21
  wandb_run_id:
22
  wandb_log_model:
 
17
  lora_target_linear:
18
  lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
19
  wandb_project:
20
+ wandb_entity:
21
  wandb_watch:
22
  wandb_run_id:
23
  wandb_log_model:
examples/redpajama/config-3b.yml CHANGED
@@ -21,6 +21,7 @@ lora_target_modules:
21
  - v_proj
22
  lora_fan_in_fan_out: false
23
  wandb_project: redpajama-alpaca-3b
 
24
  wandb_watch:
25
  wandb_run_id:
26
  wandb_log_model:
 
21
  - v_proj
22
  lora_fan_in_fan_out: false
23
  wandb_project: redpajama-alpaca-3b
24
+ wandb_entity:
25
  wandb_watch:
26
  wandb_run_id:
27
  wandb_log_model:
examples/replit-3b/config-lora.yml CHANGED
@@ -20,6 +20,7 @@ lora_target_modules:
20
  - mlp_down
21
  lora_fan_in_fan_out:
22
  wandb_project: lora-replit
 
23
  wandb_watch:
24
  wandb_run_id:
25
  wandb_log_model:
 
20
  - mlp_down
21
  lora_fan_in_fan_out:
22
  wandb_project: lora-replit
23
+ wandb_entity:
24
  wandb_watch:
25
  wandb_run_id:
26
  wandb_log_model:
examples/xgen-7b/xgen-7b-8k-qlora.yml CHANGED
@@ -37,6 +37,7 @@ lora_target_linear: true
37
  lora_fan_in_fan_out:
38
 
39
  wandb_project:
 
40
  wandb_watch:
41
  wandb_run_id:
42
  wandb_log_model:
 
37
  lora_fan_in_fan_out:
38
 
39
  wandb_project:
40
+ wandb_entity:
41
  wandb_watch:
42
  wandb_run_id:
43
  wandb_log_model:
src/axolotl/utils/wandb.py CHANGED
@@ -9,6 +9,8 @@ def setup_wandb_env_vars(cfg):
9
  elif cfg.wandb_project and len(cfg.wandb_project) > 0:
10
  os.environ["WANDB_PROJECT"] = cfg.wandb_project
11
  cfg.use_wandb = True
 
 
12
  if cfg.wandb_watch and len(cfg.wandb_watch) > 0:
13
  os.environ["WANDB_WATCH"] = cfg.wandb_watch
14
  if cfg.wandb_log_model and len(cfg.wandb_log_model) > 0:
 
9
  elif cfg.wandb_project and len(cfg.wandb_project) > 0:
10
  os.environ["WANDB_PROJECT"] = cfg.wandb_project
11
  cfg.use_wandb = True
12
+ if cfg.wandb_entity and len(cfg.wandb_entity) > 0:
13
+ os.environ["WANDB_ENTITY"] = cfg.wandb_entity
14
  if cfg.wandb_watch and len(cfg.wandb_watch) > 0:
15
  os.environ["WANDB_WATCH"] = cfg.wandb_watch
16
  if cfg.wandb_log_model and len(cfg.wandb_log_model) > 0: