max_seq_length seems not to be properly reported in sentence_bert_config.json

#35
by yjoonjang - opened

Hi authors, I was looking at the max_seq_length of this model and found out that it's not properly written in sentence_bert_config.json(https://huggingface.co/nvidia/NV-Embed-v2/blob/main/sentence_bert_config.json)

In the "Usage" section (https://huggingface.co/nvidia/NV-Embed-v2#usage-huggingface-transformers), the max_seq_length is 32768. Maybe the max_seq_length you wrote (4096) in sentence_bert_config.json might represent the hidden state (or the size of the output vector)

Can you please check?
It would be awesome if @tomaarsen could check too.

Thank you.

Hello!
You're right, it should be 32768 in sentence_bert_config.json. The reason it didn't affect the README snippet is because the 4096 immediately gets overridden due to model.max_seq_length = 32768. But we should probably update the config.

  • Tom Aarsen

Sign up or log in to comment