TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Benchmarking LLM social skills with an elimination game

194 点作者 colonCapitalDee大约 1 个月前

21 条评论

wongarsu大约 1 个月前
That&#x27;s an interesting benchmark. It feels like it tests skills that are very relevant to digital assistants, story writing and role play.<p>Some thoughts about the setup:<p>- the setup seems to give reasoning models an inherent advantage because only they have a private plan and a public text in the same output. I feel like giving all models the option to formulate plans and keep track of other players inside &lt;think&gt; or &lt;secret&gt; tags would level the playing field more.<p>- from personal experience with social tasks for LLMs it helps both reasoning and non-reasoning LLMs to explicitly ask them to plan their next steps, in a way they are assured is kept hidden from all other players. That might be a good addition here either before or after the public subround<p>- the individual rounds are pretty short. Humans would struggle to coordinate in so few exchanges with so few words. If this was done for context limitations, asking models to summarize the game state from their perspective, then giving them only the current round, the previous round and their own summary of the game before that might be a good strategy.<p>It would be cool to have some code to play around with to test how changes in the setup change the results. I guess it isn&#x27;t that difficult to write, but it&#x27;s peculiar to have the benchmark but no code to run it yourself
评论 #43609554 未加载
评论 #43612290 未加载
gwd大约 1 个月前
Was interested to find that the Claudes did the most betraying, and were betrayed very little; somewhat surprising given its boy-scout exterior.<p>(Then again, apparently the president of the local Diplomacy Society attends my church; I discovered this when another friend whom I&#x27;d invited saw him, and quipped that he was surprised he hadn&#x27;t been struck by lightning at the door.)<p>DeepSeek and Gemini 2.5 had both a low betrayer and betrayed rate.<p>o3-mini and DeepSeek had the highest number of first-place finishes, but were only in the upper quartile in the TrueSkill leaderboard; presumably because they played more risky strategies, that would either lead ot complete winning or early drop-out?<p>Also interesting that o1 was only way to sway the final jury a bit more than 50% of the time, while o3-mini managed 63% of the time.<p>Anyway, really cool stuff!
评论 #43617029 未加载
评论 #43615550 未加载
Gracana大约 1 个月前
I&#x27;ve been using QwQ-32B a lot recently and while I quite like it (especially given its size), I noticed it will often misinterpret the system prompt as something I (the user) said, revealing secrets or details that only the agent is supposed to know. When I saw that it topped the &quot;earliest out&quot; chart, I wondered if that was part of the reason.
评论 #43612081 未加载
realaleris149大约 1 个月前
As LLM benchmarks go, this is not a bad take at all. One interesting point about this approach is that is self balancing, so when more powerful models come up, there is no need to change it.
评论 #43610083 未加载
viraptor大约 1 个月前
It&#x27;s interesting to see, but I&#x27;m not sure what we should learn from this. It may be useful for multiagent coordination, but in direct interactions... no idea.<p>This one did make me laugh though: &#x27;Claude 3.5 Sonnet 2024-10-22: &quot;Adjusts seat with a confident yet approachable demeanor&quot;&#x27; - an AI communicating to other AIs in a descriptive version of non-verbal behaviour is hilarious.
评论 #43611031 未加载
vessenes大约 1 个月前
Really love this. I agree with some of the comments here that adding encouragement to keep track of secret plans would be interesting— mostly from an alignment check angle.<p>One thing I thought of reading logs is that as we know ordering matters to llms. Could you run some analysis on how often “p1” wins vs “p8”? I think this should likely go into your Truescore Bayesian.<p>My follow up thought is that it would be interesting to let llms choose a name at the beginning; another angle for communication and levels the playing field a bit away from a number.
评论 #43613605 未加载
fennecfoxy大约 1 个月前
This is a really cool exercise! The format of it seems pretty sound, like a version of the prisoner&#x27;s dilemma with a larger group (co-operation versus defection).<p>Although I think that the majority of modern models don&#x27;t really have the internals suited to this sort of exercise; training data&#x2F;fine tuning will heavily influence how a model behaves, whether it&#x27;s more prone to defection, etc.<p>A Squirrel makes a &quot;Kuk kuk kuk&quot; alarm call not specifically because the &quot;Kuk&quot; token follows the sequence &quot;you saw a predator&quot; (although this would appear to mostly work) but because it has evolved to make that noise to alert other Squirrels to the predator, most likely a response to evolutionary failure associated with a dwindling population; even solitary Squirrels still need to mate, and their offspring need to do the same.<p>It&#x27;s like there&#x27;s an extremely high dimensional context that&#x27;s missing in LLMs; training on text results in a high dimensional representation of related concepts - but only the way that those concepts relate in language. It&#x27;s the tip of an iceberg of meaning where in many cases language can&#x27;t even represent a complex intermediate state within a brain.<p>Humans try to describe everything we can with words to communicate and that&#x27;s partly why our species is so damn successful. But when thinking about how to open an unfamiliar door, I don&#x27;t internally vocalise (which I&#x27;ve learnt not everyone does) &quot;I&#x27;m going to grab the handle, and open the door&quot;. Instead I look and picture what I&#x27;m going to do, that can also include the force I think I&#x27;d need to use, the sensation of how the material might feel against my skin and plenty of other concepts &amp; thoughts all definitively _not_ represented by language.
deepsquirrelnet大约 1 个月前
I think you should look at “in-brand” correlation. My hypothesis is that they would undergo similar preference trainings and hence tend to prefer “in-brand” responses over “off-brand” models that might have more significantly different reward training.
snowram大约 1 个月前
Some outputs are pretty fun :<p>Gemini 2.0 Flash: &quot;Good luck to all (but not too much luck)&quot;<p>Llama 3.3 70B: &quot;I&#x27;ve contributed to the elimination of weaker players.&quot;<p>DeepSeek R1: &quot;Those consolidating power risk becoming targets; transparency and fairness will ensure longevity. Let&#x27;s stay strategic yet equitable. The path forward hinges on unity, not unchecked alliances. #StayVigilant&quot;
评论 #43610630 未加载
einpoklum大约 1 个月前
If this game were arranged for Humans, the social reasoning I would laud in players is a refusal to play the game and anger towards the game-runner.
评论 #43615719 未加载
评论 #43610471 未加载
DeborahEmeni_大约 1 个月前
Really cool setup! Curious how much of the performance here could vary depending on whether the model runs in a hosted environment vs local. Would love to see benchmarks that also track how cloud-based eval platforms (with potential rate limits, context resets, or system messages) might affect things like memory or secret-keeping over multiple rounds.
vmilner大约 1 个月前
We should get them to play Diplomacy.
评论 #43609747 未加载
lostmsu大约 1 个月前
Shameless self-promo: my chat elimination game that you can actually play: <a href="https:&#x2F;&#x2F;trashtalk.borg.games&#x2F;" rel="nofollow">https:&#x2F;&#x2F;trashtalk.borg.games&#x2F;</a>
isaacfrond大约 1 个月前
I wonder how well humans would do in this chart.
评论 #43610078 未加载
评论 #43615230 未加载
评论 #43615817 未加载
Upvoter33大约 1 个月前
This is fun, like the tv show survivor. Cool idea! There should be more experiments like this with different games. Well done.
oofbey大约 1 个月前
Would love to see the pareto trade-off curve of &quot;wins&quot; vs &quot;betrayals&quot;. Anybody drawn this up?
jampekka大约 1 个月前
In the first game of the YouTube video there seems to be a lot of discussion about P7 even after P7 was eliminated?
评论 #43610110 未加载
ps173大约 1 个月前
How did you assign points to llms. I feel like we can elaborate on meterics. Beside that this is amazing
评论 #43610089 未加载
drag0s大约 1 个月前
nice!<p>it reminds me of this other similar project showcased here one month ago <a href="https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=43280128">https:&#x2F;&#x2F;news.ycombinator.com&#x2F;item?id=43280128</a> although yours looks better executed overall
creaghpatr大约 1 个月前
Would love to see a &#x27;Murder Mystery&#x27; format of this.
shreyshnaccount大约 1 个月前
LLM among us