TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

DPO fine-tuned Mistral 7B beats Llama 70B on MT Bench

3 点作者 clmnt超过 1 年前

2 条评论

amilios超过 1 年前
Can anyone corroborate this anecdotally? I.e. has anyone actually looked at the output of the two models side-by-side for common tasks? There's lots of talks these days about academic benchmarks being pretty "broken" for modern LMs, and not really properly showcasing the differences between models. I wonder if that's the case here or if the model is genuinely better.
评论 #37849556 未加载
brucethemoose2超过 1 年前
&gt; &lt;|system|&gt;, &lt;|user|&gt; and &lt;|model|&gt;.<p>Oh hey, thats almost Metharme&#x27;s format.<p>It must originate from an older model, as most new models dont use that syntax.