{ "config": { "activation_function": "gelu", "bias": true, "embedding_size": 768, "head_type": "masked_lm", "label2id": null, "layer_norm": true, "layers": 2, "shift_labels": false, "vocab_size": 50265 }, "hidden_size": 768, "model_class": "RobertaAdapterModel", "model_name": "roberta-base", "model_type": "roberta", "name": "micro_par_bn_v_4_pretrain", "version": "0.1.2" }