max_seq_length seems not to be properly reported in sentence_bert_config.json
Hi authors, I was looking at the max_seq_length of this model and found out that it's not properly written in sentence_bert_config.json(https://huggingface.co/nvidia/NV-Embed-v2/blob/main/sentence_bert_config.json)
In the "Usage" section (https://huggingface.co/nvidia/NV-Embed-v2#usage-huggingface-transformers), the max_seq_length is 32768. Maybe the max_seq_length you wrote (4096) in sentence_bert_config.json might represent the hidden state (or the size of the output vector)
Can you please check?
It would be awesome if
@tomaarsen
could check too.
Thank you.
Hello!
You're right, it should be 32768 in sentence_bert_config.json
. The reason it didn't affect the README snippet is because the 4096 immediately gets overridden due to model.max_seq_length = 32768
. But we should probably update the config.
- Tom Aarsen