Interesting, but this aspect makes me double-take: "We demonstrate that Mistral-7B, when fine-tuned solely on synthetic data, attains competitive
performance on the BEIR [ 40 ] and MTEB [27] benchmarks".<p>E5/BGE large are an order of magnitude smaller than Mistral-7B. So is this just "bigger model wins" in disguise?<p>I need to read the whole paper carefully, but this jumped out at me.