Can't load model with SentenceTransformers 3.0.1 AttributeError: 'LatentAttentionConfig' object has no attribute '_attn_implementation_internal'
Hello, I am trying to load the model using the following code
from sentence_transformers import SentenceTransformer
nv_model = SentenceTransformer('nvidia/NV-Embed-v1', trust_remote_code=True)
When I do it downloads the shards and then errors out with the error "AttributeError: 'LatentAttentionConfig' object has no attribute '_attn_implementation_internal'". Is there a specific version of sentence_transformers I should be using?
Hi, @jswarner85 . Can you try to install the package as below?
pip uninstall -y transformer-engine
pip install torch==2.2.0
pip install transformers==4.42.4
pip install flash-attn==2.2.0
pip install sentence-transformers==2.7.0
I am experiencing the same issue
I think the problem is with latest transformers==4.44.1. I downgraded to transformers==4.42.4 and it works now. (I did not try versions between 4.42.4 and 4.44.1)
I tried setting the transformers version to 4.42.4 and 4.44.1 and both times I am getting the same, but different, error: ImportError: libcudart.so.11.0: cannot open shared object file: No such file or directory.
Is anyone who has gotten the model working able to share their python,cuda, transformers version worked for you? I'm on sagemaker using image "pytorch 2.2.0 python 3.10 gpu optimized".
Hi, @konkalita . Thanks for validating the transformers package version. Readme trouble shooting is updated accordingly.
@jswarner85
Sure, although from my experience libcudart.so.11.0
-like errors are usually related to some mismatch between cuda and torch libs not related to this model.
accelerate==0.33.0
einops==0.8.0
nvidia-cublas-cu12==12.1.3.1
nvidia-cuda-cupti-cu12==12.1.105
nvidia-cuda-nvrtc-cu12==12.1.105
nvidia-cuda-runtime-cu12==12.1.105
nvidia-cudnn-cu12==9.1.0.70
nvidia-cufft-cu12==11.0.2.54
nvidia-curand-cu12==10.3.2.106
nvidia-cusolver-cu12==11.4.5.107
nvidia-cusparse-cu12==12.1.0.106
nvidia-nccl-cu12==2.20.5
nvidia-nvjitlink-cu12==12.6.20
nvidia-nvtx-cu12==12.1.105
peft==0.12.0
safetensors==0.4.4
sentence-transformers==3.0.1
tokenizers==0.19.1
torch==2.4.0
transformers==4.42.4
triton==3.0.0
NVIDIA-SMI 550.90.07 Driver Version: 550.90.07 CUDA Version: 12.4
I think the problem is with latest transformers==4.44.1. I downgraded to transformers==4.42.4 and it works now. (I did not try versions between 4.42.4 and 4.44.1)
transformers==4.43.4 works for me
I confirm version 4.43.4 works for me
Fixing it in #53