# Dense Encoder - Distilbert - Frozen Token Embeddings This model is a distilbert-base-uncased model trained for 30 epochs (235k steps), 64 batch size with MarginMSE Loss on MS MARCO dataset. The token embeddings were frozen. | Dataset | Model with updated token embeddings | Model with frozen embeddings | | --- | :---: | :---: | | TREC-DL 19 | 70.68 | 68.60 | | TREC-DL 20 | 67.69 | 70.21 | | FiQA | 28.89 | 28.60 | | Robust04 | 39.56 | 39.08 | | TREC-COVID v2 | 69.80 | 69.84 | | TREC-NEWS | 37.97 | 38.27 | | Avg. 4 BEIR tasks | 44.06 | 43.95 |