RhinoWithAcape commited on
Commit
9112cb5
·
verified ·
1 Parent(s): 46ece13

Add Physics, Chemistry, Biology specialist LoRA weights and fix model info

Browse files
.gitattributes CHANGED
@@ -1,36 +1,3 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
  *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  grogu-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  *.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ *.bin filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
3
  grogu-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -114,9 +114,9 @@ The system achieves state-of-the-art results on graduate-level science benchmark
114
  │ │ GROGU │ │ PHYSICS │ │CHEMISTRY │ │ BIOLOGY │ │
115
  │ │ (General)│ │ Expert │ │ Expert │ │ Expert │ │
116
  │ │ │ │ │ │ │ │ │ │
117
- │ │ Nemotron │ │ QwQ-32B │ │ QwQ-32B │ │ QwQ-32B │ │
118
- │ │ Qwen 1.5B│ │ Physics │ │ Chemistry│ │ Biology │ │
119
- │ │ + LoRA │ │ LoRA │ │ LoRA │ │ LoRA │ │
120
  │ └────┬─────┘ └────┬─────┘ └────┬─────┘ └────┬─────┘ │
121
  │ │ │ │ │ │
122
  │ └─────────────┴──────┬──────┴─────────────┘ │
@@ -140,10 +140,12 @@ The system achieves state-of-the-art results on graduate-level science benchmark
140
 
141
  | Component | Base Model | Training Method | Parameters |
142
  |-----------|------------|-----------------|------------|
143
- | **Grogu (Generalist)** | Nemotron-Qwen-1.5B | LoRA (r=16, α=32) | ~1.5B + 2.4M LoRA |
144
- | **Physics Expert** | QwQ-32B | LoRA fine-tuning | ~32B + adapters |
145
- | **Chemistry Expert** | QwQ-32B | LoRA fine-tuning | ~32B + adapters |
146
- | **Biology Expert** | QwQ-32B | LoRA fine-tuning | ~32B + adapters |
 
 
147
 
148
  ---
149
 
 
114
  │ │ GROGU │ │ PHYSICS │ │CHEMISTRY │ │ BIOLOGY │ │
115
  │ │ (General)│ │ Expert │ │ Expert │ │ Expert │ │
116
  │ │ │ │ │ │ │ │ │ │
117
+ │ │ Nemotron │ │Qwen2.5 │ │Qwen2.5 │ │ Nemotron │ │
118
+ │ │ Qwen 1.5B│ │ 1.5B │ │ 1.5B │ │ Qwen 1.5B│ │
119
+ │ │ + LoRA │ │ + LoRA │ │ + LoRA │ │ + LoRA │ │
120
  │ └────┬─────┘ └────┬─────┘ └────┬─────┘ └────┬─────┘ │
121
  │ │ │ │ │ │
122
  │ └─────────────┴──────┬──────┴─────────────┘ │
 
140
 
141
  | Component | Base Model | Training Method | Parameters |
142
  |-----------|------------|-----------------|------------|
143
+ | **Grogu (Generalist)** | nvidia/nemotron-qwen-1.5b | LoRA (r=16, α=32) | ~1.5B + 2.4M LoRA |
144
+ | **Physics Expert** | Qwen/Qwen2.5-1.5B-Instruct | LoRA fine-tuning | ~1.5B + adapters |
145
+ | **Chemistry Expert** | Qwen/Qwen2.5-1.5B-Instruct | LoRA fine-tuning | ~1.5B + adapters |
146
+ | **Biology Expert** | nvidia/nemotron-qwen-1.5b | LoRA fine-tuning (ProRL) | ~1.5B + adapters |
147
+
148
+ > **Note**: All 4 agents use efficient 1.5B parameter models, allowing the entire system to run on a single 12GB GPU.
149
 
150
  ---
151
 
biology-lora/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "nvidia/nemotron-qwen-1.5b",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
biology-lora/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:997da1c27c7778fc89ccfc7a268637666d054d713502231909e8d4ba72dfce38
3
+ size 17462432
chemistry-lora/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
chemistry-lora/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f63566a15f8b707a90e953896cf3691bde5d85d9b8ed800d43249815ff3d7b4e
3
+ size 17462432
configs/debate_config.yaml CHANGED
@@ -8,31 +8,31 @@ system:
8
  version: "1.0.0"
9
  organization: "Zenith Global AI Ltd"
10
 
11
- # Model Paths (update with your HuggingFace paths after upload)
12
  models:
13
  grogu:
14
  base_model: "nvidia/nemotron-qwen-1.5b"
15
- lora_path: "zenith-global/grogu-science-moe/grogu-lora"
16
  weight: 1.0
17
- description: "General reasoning agent with curriculum-trained LoRA"
18
 
19
  physics:
20
- base_model: "Qwen/QwQ-32B-Preview"
21
- lora_path: null # Add path if releasing physics specialist
22
  weight: 1.2
23
- description: "Physics domain specialist"
24
 
25
  chemistry:
26
- base_model: "Qwen/QwQ-32B-Preview"
27
- lora_path: null # Add path if releasing chemistry specialist
28
  weight: 1.2
29
- description: "Chemistry domain specialist"
30
 
31
  biology:
32
- base_model: "Qwen/QwQ-32B-Preview"
33
- lora_path: null # Add path if releasing biology specialist
34
  weight: 1.2
35
- description: "Biology domain specialist"
36
 
37
  # Debate Configuration
38
  debate:
@@ -140,8 +140,8 @@ hardware:
140
  description: "Grogu alone (1.5B parameters)"
141
 
142
  full_system:
143
- gpu_memory: "80GB+"
144
- description: "All 4 agents (32B specialists require significant VRAM)"
145
 
146
  recommended:
147
  device: "cuda"
 
8
  version: "1.0.0"
9
  organization: "Zenith Global AI Ltd"
10
 
11
+ # Model Paths
12
  models:
13
  grogu:
14
  base_model: "nvidia/nemotron-qwen-1.5b"
15
+ lora_path: "RhinoWithAcape/grogu-science-moe/grogu-lora"
16
  weight: 1.0
17
+ description: "General reasoning agent with curriculum-trained LoRA (1.5B)"
18
 
19
  physics:
20
+ base_model: "Qwen/Qwen2.5-1.5B-Instruct"
21
+ lora_path: "RhinoWithAcape/grogu-science-moe/physics-lora"
22
  weight: 1.2
23
+ description: "Physics domain specialist (1.5B)"
24
 
25
  chemistry:
26
+ base_model: "Qwen/Qwen2.5-1.5B-Instruct"
27
+ lora_path: "RhinoWithAcape/grogu-science-moe/chemistry-lora"
28
  weight: 1.2
29
+ description: "Chemistry domain specialist (1.5B)"
30
 
31
  biology:
32
+ base_model: "nvidia/nemotron-qwen-1.5b"
33
+ lora_path: "RhinoWithAcape/grogu-science-moe/biology-lora"
34
  weight: 1.2
35
+ description: "Biology domain specialist - Nemotron ProRL (1.5B)"
36
 
37
  # Debate Configuration
38
  debate:
 
140
  description: "Grogu alone (1.5B parameters)"
141
 
142
  full_system:
143
+ gpu_memory: "12GB"
144
+ description: "All 4 agents (4x 1.5B = ~6B total, fits on single GPU)"
145
 
146
  recommended:
147
  device: "cuda"
physics-lora/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
physics-lora/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa38957390b839969f90818c4883a3436ad4a2772547c8b9fb013f6ceb44f638
3
+ size 17462432