manzilzaheer's picture
Create README.md
d6813d2 verified
|
raw
history blame
998 Bytes
metadata
language:
  - en
base_model:
  - google/gemma-2-9b-it

Gofer Embeddings v0.8

Gofer is a dense-vector embedding model, trained especially for retrieval. As of December 2, 2024, Gofer achieves the #1 position overall on the MTEB Retrieval leaderboard, with a score of 63.01.

Important Notes

  • This is not an official Google product.
  • This is a research project.

Results summary

Results compared to BGE-EN-ICL on several large datasets

Model DBPedia FEVER HotPotQA MSMARCO NQ
BGE-EN-ICL 51.63 92.83 85.14 46.79 73.88
Gofer Embeddings v0.8 52.58 93.225 86.921 47.537 73.75

Model & Data

Our base encoder model is Gemma2 9B.

We use the BGE-EN-ICL training data.

Research Team

  • Nicholas Monath
  • Michael Boratko
  • Seungyeon Kim
  • Andrew McCallum
  • Rob Fergus
  • Manzil Zaheer