Add Physics, Chemistry, Biology specialist LoRA weights and fix model info
Browse files- .gitattributes +1 -34
- README.md +9 -7
- biology-lora/adapter_config.json +38 -0
- biology-lora/adapter_model.safetensors +3 -0
- chemistry-lora/adapter_config.json +38 -0
- chemistry-lora/adapter_model.safetensors +3 -0
- configs/debate_config.yaml +14 -14
- physics-lora/adapter_config.json +38 -0
- physics-lora/adapter_model.safetensors +3 -0
.gitattributes
CHANGED
|
@@ -1,36 +1,3 @@
|
|
| 1 |
-
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
-
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
-
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
-
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
-
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
-
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
-
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
-
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
-
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
-
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
-
|
| 27 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
-
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
-
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
grogu-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
grogu-lora/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
|
@@ -114,9 +114,9 @@ The system achieves state-of-the-art results on graduate-level science benchmark
|
|
| 114 |
│ │ GROGU │ │ PHYSICS │ │CHEMISTRY │ │ BIOLOGY │ │
|
| 115 |
│ │ (General)│ │ Expert │ │ Expert │ │ Expert │ │
|
| 116 |
│ │ │ │ │ │ │ │ │ │
|
| 117 |
-
│ │ Nemotron │ │
|
| 118 |
-
│ │ Qwen 1.5B│ │
|
| 119 |
-
│ │ + LoRA │ │ LoRA
|
| 120 |
│ └────┬─────┘ └────┬─────┘ └────┬─────┘ └────┬─────┘ │
|
| 121 |
│ │ │ │ │ │
|
| 122 |
│ └─────────────┴──────┬──────┴─────────────┘ │
|
|
@@ -140,10 +140,12 @@ The system achieves state-of-the-art results on graduate-level science benchmark
|
|
| 140 |
|
| 141 |
| Component | Base Model | Training Method | Parameters |
|
| 142 |
|-----------|------------|-----------------|------------|
|
| 143 |
-
| **Grogu (Generalist)** |
|
| 144 |
-
| **Physics Expert** |
|
| 145 |
-
| **Chemistry Expert** |
|
| 146 |
-
| **Biology Expert** |
|
|
|
|
|
|
|
| 147 |
|
| 148 |
---
|
| 149 |
|
|
|
|
| 114 |
│ │ GROGU │ │ PHYSICS │ │CHEMISTRY │ │ BIOLOGY │ │
|
| 115 |
│ │ (General)│ │ Expert │ │ Expert │ │ Expert │ │
|
| 116 |
│ │ │ │ │ │ │ │ │ │
|
| 117 |
+
│ │ Nemotron │ │Qwen2.5 │ │Qwen2.5 │ │ Nemotron │ │
|
| 118 |
+
│ │ Qwen 1.5B│ │ 1.5B │ │ 1.5B │ │ Qwen 1.5B│ │
|
| 119 |
+
│ │ + LoRA │ │ + LoRA │ │ + LoRA │ │ + LoRA │ │
|
| 120 |
│ └────┬─────┘ └────┬─────┘ └────┬─────┘ └────┬─────┘ │
|
| 121 |
│ │ │ │ │ │
|
| 122 |
│ └─────────────┴──────┬──────┴─────────────┘ │
|
|
|
|
| 140 |
|
| 141 |
| Component | Base Model | Training Method | Parameters |
|
| 142 |
|-----------|------------|-----------------|------------|
|
| 143 |
+
| **Grogu (Generalist)** | nvidia/nemotron-qwen-1.5b | LoRA (r=16, α=32) | ~1.5B + 2.4M LoRA |
|
| 144 |
+
| **Physics Expert** | Qwen/Qwen2.5-1.5B-Instruct | LoRA fine-tuning | ~1.5B + adapters |
|
| 145 |
+
| **Chemistry Expert** | Qwen/Qwen2.5-1.5B-Instruct | LoRA fine-tuning | ~1.5B + adapters |
|
| 146 |
+
| **Biology Expert** | nvidia/nemotron-qwen-1.5b | LoRA fine-tuning (ProRL) | ~1.5B + adapters |
|
| 147 |
+
|
| 148 |
+
> **Note**: All 4 agents use efficient 1.5B parameter models, allowing the entire system to run on a single 12GB GPU.
|
| 149 |
|
| 150 |
---
|
| 151 |
|
biology-lora/adapter_config.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "nvidia/nemotron-qwen-1.5b",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"corda_config": null,
|
| 7 |
+
"eva_config": null,
|
| 8 |
+
"exclude_modules": null,
|
| 9 |
+
"fan_in_fan_out": false,
|
| 10 |
+
"inference_mode": true,
|
| 11 |
+
"init_lora_weights": true,
|
| 12 |
+
"layer_replication": null,
|
| 13 |
+
"layers_pattern": null,
|
| 14 |
+
"layers_to_transform": null,
|
| 15 |
+
"loftq_config": {},
|
| 16 |
+
"lora_alpha": 32,
|
| 17 |
+
"lora_bias": false,
|
| 18 |
+
"lora_dropout": 0.05,
|
| 19 |
+
"megatron_config": null,
|
| 20 |
+
"megatron_core": "megatron.core",
|
| 21 |
+
"modules_to_save": null,
|
| 22 |
+
"peft_type": "LORA",
|
| 23 |
+
"qalora_group_size": 16,
|
| 24 |
+
"r": 16,
|
| 25 |
+
"rank_pattern": {},
|
| 26 |
+
"revision": null,
|
| 27 |
+
"target_modules": [
|
| 28 |
+
"v_proj",
|
| 29 |
+
"o_proj",
|
| 30 |
+
"q_proj",
|
| 31 |
+
"k_proj"
|
| 32 |
+
],
|
| 33 |
+
"task_type": "CAUSAL_LM",
|
| 34 |
+
"trainable_token_indices": null,
|
| 35 |
+
"use_dora": false,
|
| 36 |
+
"use_qalora": false,
|
| 37 |
+
"use_rslora": false
|
| 38 |
+
}
|
biology-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:997da1c27c7778fc89ccfc7a268637666d054d713502231909e8d4ba72dfce38
|
| 3 |
+
size 17462432
|
chemistry-lora/adapter_config.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"corda_config": null,
|
| 7 |
+
"eva_config": null,
|
| 8 |
+
"exclude_modules": null,
|
| 9 |
+
"fan_in_fan_out": false,
|
| 10 |
+
"inference_mode": true,
|
| 11 |
+
"init_lora_weights": true,
|
| 12 |
+
"layer_replication": null,
|
| 13 |
+
"layers_pattern": null,
|
| 14 |
+
"layers_to_transform": null,
|
| 15 |
+
"loftq_config": {},
|
| 16 |
+
"lora_alpha": 32,
|
| 17 |
+
"lora_bias": false,
|
| 18 |
+
"lora_dropout": 0.05,
|
| 19 |
+
"megatron_config": null,
|
| 20 |
+
"megatron_core": "megatron.core",
|
| 21 |
+
"modules_to_save": null,
|
| 22 |
+
"peft_type": "LORA",
|
| 23 |
+
"qalora_group_size": 16,
|
| 24 |
+
"r": 16,
|
| 25 |
+
"rank_pattern": {},
|
| 26 |
+
"revision": null,
|
| 27 |
+
"target_modules": [
|
| 28 |
+
"v_proj",
|
| 29 |
+
"o_proj",
|
| 30 |
+
"q_proj",
|
| 31 |
+
"k_proj"
|
| 32 |
+
],
|
| 33 |
+
"task_type": "CAUSAL_LM",
|
| 34 |
+
"trainable_token_indices": null,
|
| 35 |
+
"use_dora": false,
|
| 36 |
+
"use_qalora": false,
|
| 37 |
+
"use_rslora": false
|
| 38 |
+
}
|
chemistry-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f63566a15f8b707a90e953896cf3691bde5d85d9b8ed800d43249815ff3d7b4e
|
| 3 |
+
size 17462432
|
configs/debate_config.yaml
CHANGED
|
@@ -8,31 +8,31 @@ system:
|
|
| 8 |
version: "1.0.0"
|
| 9 |
organization: "Zenith Global AI Ltd"
|
| 10 |
|
| 11 |
-
# Model Paths
|
| 12 |
models:
|
| 13 |
grogu:
|
| 14 |
base_model: "nvidia/nemotron-qwen-1.5b"
|
| 15 |
-
lora_path: "
|
| 16 |
weight: 1.0
|
| 17 |
-
description: "General reasoning agent with curriculum-trained LoRA"
|
| 18 |
|
| 19 |
physics:
|
| 20 |
-
base_model: "Qwen/
|
| 21 |
-
lora_path:
|
| 22 |
weight: 1.2
|
| 23 |
-
description: "Physics domain specialist"
|
| 24 |
|
| 25 |
chemistry:
|
| 26 |
-
base_model: "Qwen/
|
| 27 |
-
lora_path:
|
| 28 |
weight: 1.2
|
| 29 |
-
description: "Chemistry domain specialist"
|
| 30 |
|
| 31 |
biology:
|
| 32 |
-
base_model: "
|
| 33 |
-
lora_path:
|
| 34 |
weight: 1.2
|
| 35 |
-
description: "Biology domain specialist"
|
| 36 |
|
| 37 |
# Debate Configuration
|
| 38 |
debate:
|
|
@@ -140,8 +140,8 @@ hardware:
|
|
| 140 |
description: "Grogu alone (1.5B parameters)"
|
| 141 |
|
| 142 |
full_system:
|
| 143 |
-
gpu_memory: "
|
| 144 |
-
description: "All 4 agents (
|
| 145 |
|
| 146 |
recommended:
|
| 147 |
device: "cuda"
|
|
|
|
| 8 |
version: "1.0.0"
|
| 9 |
organization: "Zenith Global AI Ltd"
|
| 10 |
|
| 11 |
+
# Model Paths
|
| 12 |
models:
|
| 13 |
grogu:
|
| 14 |
base_model: "nvidia/nemotron-qwen-1.5b"
|
| 15 |
+
lora_path: "RhinoWithAcape/grogu-science-moe/grogu-lora"
|
| 16 |
weight: 1.0
|
| 17 |
+
description: "General reasoning agent with curriculum-trained LoRA (1.5B)"
|
| 18 |
|
| 19 |
physics:
|
| 20 |
+
base_model: "Qwen/Qwen2.5-1.5B-Instruct"
|
| 21 |
+
lora_path: "RhinoWithAcape/grogu-science-moe/physics-lora"
|
| 22 |
weight: 1.2
|
| 23 |
+
description: "Physics domain specialist (1.5B)"
|
| 24 |
|
| 25 |
chemistry:
|
| 26 |
+
base_model: "Qwen/Qwen2.5-1.5B-Instruct"
|
| 27 |
+
lora_path: "RhinoWithAcape/grogu-science-moe/chemistry-lora"
|
| 28 |
weight: 1.2
|
| 29 |
+
description: "Chemistry domain specialist (1.5B)"
|
| 30 |
|
| 31 |
biology:
|
| 32 |
+
base_model: "nvidia/nemotron-qwen-1.5b"
|
| 33 |
+
lora_path: "RhinoWithAcape/grogu-science-moe/biology-lora"
|
| 34 |
weight: 1.2
|
| 35 |
+
description: "Biology domain specialist - Nemotron ProRL (1.5B)"
|
| 36 |
|
| 37 |
# Debate Configuration
|
| 38 |
debate:
|
|
|
|
| 140 |
description: "Grogu alone (1.5B parameters)"
|
| 141 |
|
| 142 |
full_system:
|
| 143 |
+
gpu_memory: "12GB"
|
| 144 |
+
description: "All 4 agents (4x 1.5B = ~6B total, fits on single GPU)"
|
| 145 |
|
| 146 |
recommended:
|
| 147 |
device: "cuda"
|
physics-lora/adapter_config.json
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"corda_config": null,
|
| 7 |
+
"eva_config": null,
|
| 8 |
+
"exclude_modules": null,
|
| 9 |
+
"fan_in_fan_out": false,
|
| 10 |
+
"inference_mode": true,
|
| 11 |
+
"init_lora_weights": true,
|
| 12 |
+
"layer_replication": null,
|
| 13 |
+
"layers_pattern": null,
|
| 14 |
+
"layers_to_transform": null,
|
| 15 |
+
"loftq_config": {},
|
| 16 |
+
"lora_alpha": 32,
|
| 17 |
+
"lora_bias": false,
|
| 18 |
+
"lora_dropout": 0.05,
|
| 19 |
+
"megatron_config": null,
|
| 20 |
+
"megatron_core": "megatron.core",
|
| 21 |
+
"modules_to_save": null,
|
| 22 |
+
"peft_type": "LORA",
|
| 23 |
+
"qalora_group_size": 16,
|
| 24 |
+
"r": 16,
|
| 25 |
+
"rank_pattern": {},
|
| 26 |
+
"revision": null,
|
| 27 |
+
"target_modules": [
|
| 28 |
+
"v_proj",
|
| 29 |
+
"o_proj",
|
| 30 |
+
"q_proj",
|
| 31 |
+
"k_proj"
|
| 32 |
+
],
|
| 33 |
+
"task_type": "CAUSAL_LM",
|
| 34 |
+
"trainable_token_indices": null,
|
| 35 |
+
"use_dora": false,
|
| 36 |
+
"use_qalora": false,
|
| 37 |
+
"use_rslora": false
|
| 38 |
+
}
|
physics-lora/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa38957390b839969f90818c4883a3436ad4a2772547c8b9fb013f6ceb44f638
|
| 3 |
+
size 17462432
|