Dense Encoder - Distilbert - Frozen Token Embeddings

This model is a distilbert-base-uncased model trained for 30 epochs (235k steps), 64 batch size with MarginMSE Loss on MS MARCO dataset.

The token embeddings were frozen.

Dataset Model with updated token embeddings Model with frozen embeddings
TREC-DL 19 70.68 68.60
TREC-DL 20 67.69 70.21
FiQA 28.89 28.60
Robust04 39.56 39.08
TREC-COVID v2 69.80 69.84
TREC-NEWS 37.97 38.27
Avg. 4 BEIR tasks 44.06 43.95