fix readme
Browse files- .ipynb_checkpoints/README-checkpoint.md +28 -4
- README.md +28 -4
.ipynb_checkpoints/README-checkpoint.md
CHANGED
@@ -1,12 +1,21 @@
|
|
1 |
---
|
|
|
2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |
# [Gated State Space](https://arxiv.org/abs/2206.13947)
|
6 |
|
7 |
-
This repo contains pretrain model for the gated state space paper. The model has been trained on [C4 dataset](https://huggingface.co/datasets/c4). I have used [Lucidrains' implementation](https://github.com/lucidrains/gated-state-spaces-pytorch) ([commit](https://github.com/lucidrains/gated-state-spaces-pytorch/tree/32cd036e775112cc469e94fa1165fe111393708b)) for the model. I think the main benefit of this model is the ability to scale beyond the training context length. As authors noted in the paper, they trained the model on 4k sequence length but it generalized beyond that length. I have written a blog post on how I started the training [here](https://naxalpha.substack.com/p/devlog-experiment-a2a468-gated-state)
|
8 |
|
9 |
-
[
|
10 |
|
11 |
## How to use this.
|
12 |
|
@@ -38,6 +47,21 @@ Since it is not based on [transformers](https://github.com/huggingface/transform
|
|
38 |
```
|
39 |
|
40 |
|
41 |
-
## Training
|
42 |
|
43 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
library_name: lucidrains/gated-state-spaces-pytorch
|
3 |
license: mit
|
4 |
+
datasets:
|
5 |
+
- c4
|
6 |
+
pipeline_tag: text-generation
|
7 |
+
tags:
|
8 |
+
- text generation
|
9 |
+
- pytorch
|
10 |
+
- causal-lm
|
11 |
+
- gated-state-space
|
12 |
---
|
13 |
|
14 |
# [Gated State Space](https://arxiv.org/abs/2206.13947)
|
15 |
|
16 |
+
This repo contains pretrain model for the gated state space paper. The model has been trained on [C4 dataset](https://huggingface.co/datasets/c4). I have used [Lucidrains' implementation](https://github.com/lucidrains/gated-state-spaces-pytorch) ([commit](https://github.com/lucidrains/gated-state-spaces-pytorch/tree/32cd036e775112cc469e94fa1165fe111393708b)) for the model. I think the main benefit of this model is the ability to scale beyond the training context length. As authors noted in the paper, they trained the model on 4k sequence length but it generalized beyond that length. I have written a **blog post on how I started the training [here](https://naxalpha.substack.com/p/devlog-experiment-a2a468-gated-state)**.
|
17 |
|
18 |
+
**[Wandb Report is available at this link](https://wandb.ai/naxalpha/gated-state-space/reports/Gated-State-Space-Training-v1--VmlldzozMTYzMzY3?accessToken=zy10rrpofi9k7l52aqwiej8bk0ub302rdswfkxmf8y94dt2j6z4kxbca6ar3sc52)**
|
19 |
|
20 |
## How to use this.
|
21 |
|
|
|
47 |
```
|
48 |
|
49 |
|
50 |
+
## Training Information
|
51 |
|
52 |
+
Here are the details of the training:
|
53 |
+
|
54 |
+
- Objective: `Alternate between simple cross entropy and GPT-2 XL distillation`
|
55 |
+
- Gradient Accumulation: `4`
|
56 |
+
- Batch Size: `8`
|
57 |
+
- Sequence Length `128`
|
58 |
+
- Learning Rate: `2e-5`
|
59 |
+
- Optimizer: `AdamW`
|
60 |
+
- Gradient Norm Clipping: `1.0`
|
61 |
+
- Hardware: `RTX 3090` on [vast.ai](vast.ai)
|
62 |
+
- Training Cost: `~20$`
|
63 |
+
- Training Time: `~2 days`
|
64 |
+
- Number of steps: `434,000`
|
65 |
+
- Tokens seen: `444 million`
|
66 |
+
|
67 |
+
Training code is available in this repo. [Link to the training script](https://huggingface.co/naxalpha/gated-state-space/blob/main/app.py).
|
README.md
CHANGED
@@ -1,12 +1,21 @@
|
|
1 |
---
|
|
|
2 |
license: mit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |
# [Gated State Space](https://arxiv.org/abs/2206.13947)
|
6 |
|
7 |
-
This repo contains pretrain model for the gated state space paper. The model has been trained on [C4 dataset](https://huggingface.co/datasets/c4). I have used [Lucidrains' implementation](https://github.com/lucidrains/gated-state-spaces-pytorch) ([commit](https://github.com/lucidrains/gated-state-spaces-pytorch/tree/32cd036e775112cc469e94fa1165fe111393708b)) for the model. I think the main benefit of this model is the ability to scale beyond the training context length. As authors noted in the paper, they trained the model on 4k sequence length but it generalized beyond that length. I have written a blog post on how I started the training [here](https://naxalpha.substack.com/p/devlog-experiment-a2a468-gated-state)
|
8 |
|
9 |
-
[
|
10 |
|
11 |
## How to use this.
|
12 |
|
@@ -38,6 +47,21 @@ Since it is not based on [transformers](https://github.com/huggingface/transform
|
|
38 |
```
|
39 |
|
40 |
|
41 |
-
## Training
|
42 |
|
43 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
library_name: lucidrains/gated-state-spaces-pytorch
|
3 |
license: mit
|
4 |
+
datasets:
|
5 |
+
- c4
|
6 |
+
pipeline_tag: text-generation
|
7 |
+
tags:
|
8 |
+
- text generation
|
9 |
+
- pytorch
|
10 |
+
- causal-lm
|
11 |
+
- gated-state-space
|
12 |
---
|
13 |
|
14 |
# [Gated State Space](https://arxiv.org/abs/2206.13947)
|
15 |
|
16 |
+
This repo contains pretrain model for the gated state space paper. The model has been trained on [C4 dataset](https://huggingface.co/datasets/c4). I have used [Lucidrains' implementation](https://github.com/lucidrains/gated-state-spaces-pytorch) ([commit](https://github.com/lucidrains/gated-state-spaces-pytorch/tree/32cd036e775112cc469e94fa1165fe111393708b)) for the model. I think the main benefit of this model is the ability to scale beyond the training context length. As authors noted in the paper, they trained the model on 4k sequence length but it generalized beyond that length. I have written a **blog post on how I started the training [here](https://naxalpha.substack.com/p/devlog-experiment-a2a468-gated-state)**.
|
17 |
|
18 |
+
**[Wandb Report is available at this link](https://wandb.ai/naxalpha/gated-state-space/reports/Gated-State-Space-Training-v1--VmlldzozMTYzMzY3?accessToken=zy10rrpofi9k7l52aqwiej8bk0ub302rdswfkxmf8y94dt2j6z4kxbca6ar3sc52)**
|
19 |
|
20 |
## How to use this.
|
21 |
|
|
|
47 |
```
|
48 |
|
49 |
|
50 |
+
## Training Information
|
51 |
|
52 |
+
Here are the details of the training:
|
53 |
+
|
54 |
+
- Objective: `Alternate between simple cross entropy and GPT-2 XL distillation`
|
55 |
+
- Gradient Accumulation: `4`
|
56 |
+
- Batch Size: `8`
|
57 |
+
- Sequence Length `128`
|
58 |
+
- Learning Rate: `2e-5`
|
59 |
+
- Optimizer: `AdamW`
|
60 |
+
- Gradient Norm Clipping: `1.0`
|
61 |
+
- Hardware: `RTX 3090` on [vast.ai](vast.ai)
|
62 |
+
- Training Cost: `~20$`
|
63 |
+
- Training Time: `~2 days`
|
64 |
+
- Number of steps: `434,000`
|
65 |
+
- Tokens seen: `444 million`
|
66 |
+
|
67 |
+
Training code is available in this repo. [Link to the training script](https://huggingface.co/naxalpha/gated-state-space/blob/main/app.py).
|