AdrienB134 commited on
Commit
597a940
1 Parent(s): a4c6545
Files changed (1) hide show
  1. app.py +5 -1
app.py CHANGED
@@ -45,7 +45,11 @@ def model_inference(
45
  # )
46
 
47
  #We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
48
- model = Qwen2VLForConditionalGeneration.from_pretrained("Qwen/Qwen2-VL-2B-Instruct",attn_implementation="flash_attention_2", trust_remote_code=True, torch_dtype="auto").cuda().eval()
 
 
 
 
49
 
50
  # default processer
51
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2-VL-2B-Instruct")
 
45
  # )
46
 
47
  #We recommend enabling flash_attention_2 for better acceleration and memory saving, especially in multi-image and video scenarios.
48
+ model = Qwen2VLForConditionalGeneration.from_pretrained(
49
+ "Qwen/Qwen2-VL-2B-Instruct",
50
+ #attn_implementation="flash_attention_2",
51
+ trust_remote_code=True,
52
+ torch_dtype="auto").cuda().eval()
53
 
54
  # default processer
55
  processor = AutoProcessor.from_pretrained("Qwen/Qwen2-VL-2B-Instruct")