JingzeShi commited on
Commit
97375cb
·
verified ·
1 Parent(s): 767c289

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -7
README.md CHANGED
@@ -11,12 +11,11 @@ pipeline_tag: text-generation
11
 
12
  # **Doge 60M**
13
 
14
-
15
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/673ab3647afcea17eb4378fd/VRDlr57ADKLA-PQxFMb3V.png)
16
 
17
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
18
 
19
- In addition, Doge uses Dynamic Mask Attention as sequence transformation and can use Multi-Layer Perceptron or Cross Domain Mixture of Experts as state transformation. Dynamic Mask Attention allows the Transformer to use self-attention during training and state space during inference, and Cross Domain Mixture of Experts can directly inherit the weights of Multi-Layer Perceptron for further training. This model is trained by Jingze Shi, it only allows text input and text generation, for detailed algorithm and model architecture, please refer to [Wonderful Matrices](https://arxiv.org/abs/2412.11834), the ongoing research repository is [Wonderful Matrices](https://github.com/LoserCheems/WonderfulMatrices).
20
 
21
 
22
  ## Uses
@@ -35,11 +34,15 @@ In addition, Doge uses Dynamic Mask Attention as sequence transformation and can
35
 
36
  ## Model Details
37
 
38
- > NOTE: This model has not been fine-tuned for instruction
 
 
 
 
39
 
40
  > TODO: The larger model is under training and will be uploaded soon.
41
 
42
- **Training**:
43
 
44
  | Model | Training Data | Steps | Content Length | Tokens | LR | Batch Size | Precision |
45
  |---|---|---|---|---|---|---|---|
@@ -50,8 +53,8 @@ In addition, Doge uses Dynamic Mask Attention as sequence transformation and can
50
 
51
  | Model | MMLU | TriviaQA | ARC-E | ARC-C | PIQA | HellaSwag | OBQA | Winogrande | tokens / s on CPU |
52
  |---|---|---|---|---|---|---|---|---|---|
53
- | [Doge-20M](https://huggingface.co/JingzeShi/Doge-20M) | 25.43 | 0 | 36.83 | 22.53 | 58.38 | 27.25 | 25.60 | 50.20 | 142 |
54
- | [Doge-60M](https://huggingface.co/JingzeShi/Doge-60M) | 26.41 | 0 | 50.00 | 25.34 | 61.43 | 31.45 | 28.00 | 49.64 | 62 |
55
 
56
  > All evaluations are done using five-shot settings, without additional training on the benchmarks.
57
 
 
11
 
12
  # **Doge 60M**
13
 
14
+ ![architecture](Doge.png)
 
15
 
16
  Doge is an ongoing research project where we aim to train a series of small language models to further explore whether the Transformer framework allows for more complex feedforward network structures, enabling the model to have fewer cache states and larger knowledge capacity.
17
 
18
+ In addition, Doge uses Dynamic Mask Attention as sequence transformation and can use Multi-Layer Perceptron or Cross Domain Mixture of Experts as state transformation. Dynamic Mask Attention allows the Transformer to use self-attention during training and state space during inference, and Cross Domain Mixture of Experts can directly inherit the weights of Multi-Layer Perceptron for further training. This model is trained by Jingze Shi, it only allows text input and text generation, for detailed algorithm and model architecture, please refer to [Wonderful Matrices](https://arxiv.org/abs/2412.11834), all training details are in [here](https://github.com/LoserCheems/WonderfulMatrices).
19
 
20
 
21
  ## Uses
 
34
 
35
  ## Model Details
36
 
37
+ We build the Doge by doing Per-Training on [Smollm-Corpus](https://huggingface.co/datasets/HuggingFaceTB/smollm-corpus).
38
+
39
+ > NOTE: If you want to continue pre-training this model, you can find the unconverged checkpoint [here](https://huggingface.co/JingzeShi/Doge-60M-checkpoint).
40
+
41
+ > NOTE: These models has not been fine-tuned for instruction, the instruction model is [here](https://huggingface.co/JingzeShi/Doge-60M-Instruct).
42
 
43
  > TODO: The larger model is under training and will be uploaded soon.
44
 
45
+ **Pre-Training**:
46
 
47
  | Model | Training Data | Steps | Content Length | Tokens | LR | Batch Size | Precision |
48
  |---|---|---|---|---|---|---|---|
 
53
 
54
  | Model | MMLU | TriviaQA | ARC-E | ARC-C | PIQA | HellaSwag | OBQA | Winogrande | tokens / s on CPU |
55
  |---|---|---|---|---|---|---|---|---|---|
56
+ | [Doge-20M](https://huggingface.co/JingzeShi/Doge-20M) | 25.43 | 0.03 | 36.83 | 22.78 | 58.38 | 27.25 | 25.60 | 50.20 | 142 |
57
+ | [Doge-60M](https://huggingface.co/JingzeShi/Doge-60M) | 26.41 | 0.18 | 50.46 | 25.34 | 61.43 | 31.45 | 28.00 | 50.75 | 62 |
58
 
59
  > All evaluations are done using five-shot settings, without additional training on the benchmarks.
60