Enable flash_attention_2 support since the underlying Mistral model supports it (#3) 5ab7ef3 verified lievan winglian commited on Apr 21