from transformers import AutoModel, CLIPImageProcessor, CLIPTokenizer | |
import torch | |
import spaces | |
model_name_or_path = "BAAI/EVA-CLIP-8B" | |
image_size = 224 | |
def load_model(): | |
processor = CLIPImageProcessor.from_pretrained("openai/clip-vit-large-patch14") | |
model = AutoModel.from_pretrained( | |
model_name_or_path, | |
torch_dtype=torch.bfloat16, | |
trust_remote_code=True).to('cuda').eval() | |
tokenizer = CLIPTokenizer.from_pretrained(model_name_or_path) | |
return model, tokenizer, processor | |
load_model() |