TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Improving Text Embeddings with Large Language Models

48 点作者 cmcollier超过 1 年前

3 条评论

binarymax超过 1 年前
Interesting, but this aspect makes me double-take: &quot;We demonstrate that Mistral-7B, when fine-tuned solely on synthetic data, attains competitive performance on the BEIR [ 40 ] and MTEB [27] benchmarks&quot;.<p>E5&#x2F;BGE large are an order of magnitude smaller than Mistral-7B. So is this just &quot;bigger model wins&quot; in disguise?<p>I need to read the whole paper carefully, but this jumped out at me.
评论 #38848144 未加载
nalzok超过 1 年前
&gt; Subjects: Computation and Language (cs.CL); Information Retrieval (cs.IR)<p>I&#x27;m surprised they didn&#x27;t put `Machine Learning (cs.LG)` and `Machine Learning (stat.ML)`.
3abiton超过 1 年前
I am confused, aren&#x27;t LLMs already embeddings of text?
评论 #38852279 未加载