from llava.train.train_LoHa import train if __name__ == "__main__": train(attn_implementation="flash_attention_2")