{ "emb_size": 512, "feedforward_size": 1024, "hidden_size": 512, "hidden_act": "gelu_fast", "attention_head_size": 64, "heads_num": 6, "layers_num": 8, "decoder_layers_num": 8, "dropout": 0.0, "data_processor": "t5", "embedding": ["word"], "relative_position_embedding": true, "remove_embedding_layernorm": true, "tgt_embedding": ["word"], "share_embedding": true, "encoder": "transformer", "mask": "fully_visible", "layernorm_positioning": "pre", "feed_forward": "gated", "remove_attention_scale": true, "layernorm": "t5", "remove_transformer_bias": true, "decoder": "transformer", "target": ["lm"] }