{ | |
"attn_implementation": "flash_attention_2", | |
"connection_type": "ff", | |
"dropout": 0.1, | |
"encoder_backbone_type": "mistral", | |
"encoder_lora_name": "encoder_lora", | |
"encoder_lora_target_modules": "all", | |
"encoder_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3", | |
"is_freeze_encoder": true, | |
"is_freeze_universal_learner": false, | |
"loar_r": 16, | |
"lora_alpha": 32, | |
"num_added_tokens": 0, | |
"pooling_method": "mean", | |
"universal_learner_backbone_type": "xlm-r", | |
"universal_learner_lora_name": "universal_learner_lora", | |
"universal_learner_lora_target_modules": "all", | |
"universal_learner_name_or_path": "Hieuman/xlm-aligned-for-mistral-v0.3" | |
} |