Fredrik Carlsson commited on
Commit
764085d
1 Parent(s): 2d7742f

GPT-2 Large

Browse files
Files changed (3) hide show
  1. .gitattributes +2 -0
  2. config.json +3 -38
  3. tf_model.h5 +3 -0
.gitattributes CHANGED
@@ -25,3 +25,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ tf_model.h5 filter=lfs diff=lfs merge=lfs -text
29
+ config.json filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,38 +1,3 @@
1
- {
2
- "_name_or_path": "gpt2-large",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "gradient_checkpointing": false,
12
- "initializer_range": 0.02,
13
- "layer_norm_epsilon": 1e-05,
14
- "model_type": "gpt2",
15
- "n_ctx": 1024,
16
- "n_embd": 1280,
17
- "n_head": 20,
18
- "n_inner": null,
19
- "n_layer": 36,
20
- "n_positions": 1024,
21
- "output_attentions": true,
22
- "resid_pdrop": 0.1,
23
- "scale_attn_weights": true,
24
- "summary_activation": null,
25
- "summary_first_dropout": 0.1,
26
- "summary_proj_to_labels": true,
27
- "summary_type": "cls_index",
28
- "summary_use_proj": true,
29
- "task_specific_params": {
30
- "text-generation": {
31
- "do_sample": true,
32
- "max_length": 50
33
- }
34
- },
35
- "transformers_version": "4.8.1",
36
- "use_cache": true,
37
- "vocab_size": 50257
38
- }
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:364655f02df5d1801be741b988e8917db45ded8da2230c849857a5a3c2f51da1
3
+ size 869
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb894e748ca33b2f5bee99d2cd7af98895fdf0e433fb9c1ec690e3d5ecd68a72
3
+ size 3096618024