{ "model_type": "mistral", "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3", "task": "CAUSAL_LM", "peft_type": "LORA", "adapter_config": "adapter_config.json", "tokenizer_class": "MistralTokenizer", "hidden_size": 4096, "num_attention_heads": 32, "num_hidden_layers": 28, "vocab_size": 50257 }