Changes since the previous TMLR submission:
In this version, we explicitly state in the paper that this work is a reproducibility study under limited computational resources.
As requested by the reviewers, we made several corrections to the paper for the cases where the language was strong.
In response to the reviewers' request, we have included (i) a new experiment in which the model is trained with a batch size of 512, and (ii) an ablation study on the effect of batch size on the retrieval performance of the model. We conclude that the impressive results reported in the original paper are most likely due to training with a large batch size for many training steps.
We report higher scores than in our previous submission regarding MRL's performance when trained with the original training setup. To this extent, in the new version of the paper, we do not follow the CLDRD training setup to train a more effective MRL model. However, we still report the retrieval performance of MRL following the CLDRD training setup - due to its significance in providing a better understanding of the difference in the performance of MRL and its main competitor, CLDRD - as part of our ablation study in Section 5.3.
Changes since the last version (first revision)
- The addition of recent sota dense retrievers is in Table 1.
- A more in-depth explanation of logvar in section 5.5.
- Correction of two typographical errors in section 5.1