Tie word embeddings

#3
by jbochi - opened
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -22,7 +22,7 @@
22
  "relative_attention_max_distance": 128,
23
  "relative_attention_num_buckets": 32,
24
  "task_specific_params": {},
25
- "tie_word_embeddings": false,
26
  "transformers_version": "4.23.1",
27
  "use_cache": true,
28
  "vocab_size": 256000
 
22
  "relative_attention_max_distance": 128,
23
  "relative_attention_num_buckets": 32,
24
  "task_specific_params": {},
25
+ "tie_word_embeddings": true,
26
  "transformers_version": "4.23.1",
27
  "use_cache": true,
28
  "vocab_size": 256000