Steelskull commited on
Commit
8de079d
1 Parent(s): 93eb9da

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -4
README.md CHANGED
@@ -13,8 +13,11 @@ tags:
13
 
14
  # Lumosia-MoE-4x10.7
15
 
 
 
16
  This is a very experimantal model. its a MoE of all good performing Solar models (based off of personal experiance not open leaderboard),
17
- why? dunno whated to see what would happen
 
18
 
19
  context is maybe 32k? waiting for GGUF to upload.
20
 
@@ -42,7 +45,8 @@ Lumosia-MoE-4x10.7 is a Mixure of Experts (MoE) made with the following models:
42
 
43
  ## 🧩 Configuration
44
 
45
- ```yamlbase_model: DopeorNope/SOLARC-M-10.7B
 
46
  gate_mode: hidden
47
  dtype: bfloat16
48
  experts:
@@ -53,11 +57,13 @@ experts:
53
  - source_model: kyujinpy/Sakura-SOLAR-Instruct
54
  positive_prompts: [""]
55
  - source_model: jeonsworld/CarbonVillain-en-10.7B-v1
56
- positive_prompts: [""]```
 
57
 
58
  ## 💻 Usage
59
 
60
- ```python
 
61
  !pip install -qU transformers bitsandbytes accelerate
62
 
63
  from transformers import AutoTokenizer
 
13
 
14
  # Lumosia-MoE-4x10.7
15
 
16
+ The name Lumosia was selected as its a MoE of Multiple SOLAR Merges so it really "Lights the way".... its 3am.
17
+
18
  This is a very experimantal model. its a MoE of all good performing Solar models (based off of personal experiance not open leaderboard),
19
+
20
+ Why? Dunno whated to see what would happen
21
 
22
  context is maybe 32k? waiting for GGUF to upload.
23
 
 
45
 
46
  ## 🧩 Configuration
47
 
48
+ ```
49
+ yamlbase_model: DopeorNope/SOLARC-M-10.7B
50
  gate_mode: hidden
51
  dtype: bfloat16
52
  experts:
 
57
  - source_model: kyujinpy/Sakura-SOLAR-Instruct
58
  positive_prompts: [""]
59
  - source_model: jeonsworld/CarbonVillain-en-10.7B-v1
60
+ positive_prompts: [""]
61
+ ```
62
 
63
  ## 💻 Usage
64
 
65
+ ```
66
+ python
67
  !pip install -qU transformers bitsandbytes accelerate
68
 
69
  from transformers import AutoTokenizer