{ "attn_implementation": "flash_attention_2", "connection_type": "ff", "dropout": 0.1, "encoder_backbone_type": "mistral", "encoder_lora_name": "encoder_lora", "encoder_lora_target_modules": "all", "encoder_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3", "is_freeze_encoder": true, "is_freeze_universal_learner": false, "loar_r": 16, "lora_alpha": 32, "num_added_tokens": 0, "pooling_method": "mean", "universal_learner_backbone_type": "xlm-r", "universal_learner_lora_name": "universal_learner_lora", "universal_learner_lora_target_modules": "all", "universal_learner_name_or_path": "Hieuman/xlm-aligned-for-mistral-v0.3" }