首页

68 条评论

dudeinhawaii大约 1 个月前
I see a lot of threads pitting models against each other (or whole swarms of them) in the hope that &quot;wisdom of crowds&quot; will magically appear. After a stack of experiments of my own—and after watching the recent ASU&#x2F;Microsoft-Research work [1].. I&#x27;ve landed on a simpler takeaway:<p><i>An LLM is a terrible verifier of another LLM.</i> Subbarao Kambhampati&#x27;s &quot;(How) Do LLMs Reason&#x2F;Plan?&quot; talk shows GPT-4 confidently producing provably wrong graph-coloring proofs until a symbolic SAT solver is introduced as the referee [1]. Stechly et al. quantify the problem: letting GPT-4 critique its own answers *reduces* accuracy, whereas adding an external, sound verifier boosts it by ~30 pp across planning and puzzle tasks [2]. In other words, verification is *harder* than generation for today&#x27;s autoregressive models, so you need a checker that actually reasons about the world (compiler, linter, SAT solver, ground-truth dataset, etc.).<p>Because of that asymmetry, stacking multiple LLMs rarely helps. The &quot;LLM-Modulo&quot; position paper argues that auto-regressive models simply can&#x27;t do self-verification or long-horizon planning on their own and should instead be treated as high-recall idea generators wrapped by a single, sound verifier [3]. In my tests, replacing a five-model &quot;debate&quot; with one strong model + verifier gives equal or better answers with far less latency and orchestration overhead.<p>[1] <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=0u2hdSpNS2o" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=0u2hdSpNS2o</a> - (How) Do LLMs Reason&#x2F;Plan? (talk at Microsoft Research, 11 Apr 2025)<p>[2] <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2402.08115" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2402.08115</a><p>[3] <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2402.01817" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2402.01817</a> (related to the talk in #1)
评论 #43842083 未加载
评论 #43841207 未加载
评论 #43840833 未加载
评论 #43842076 未加载
评论 #43853446 未加载
评论 #43848090 未加载
评论 #43840850 未加载
评论 #43848830 未加载
评论 #43846984 未加载
odo1242大约 1 个月前
Something I do sometimes is:<p>- Have an AI chat model come up with an answer to a problem.<p>- Have it write a report discussing the details of the problem and why it&#x27;s answer is correct, directed at a person or AI model who has no knowledge of the initial problem or technical field.<p>- Have a second AI model with no knowledge of the problem grade the report, and write it&#x27;s own report either (a) asking for clarification &#x2F; more information about the problem that the original model didn&#x27;t provide or (b) pointing out an inconsistency in the argument posed by the original model. Give this report back to the original model and ask it to write it&#x27;s own report back with either the necessary information or changes.<p>- Repeat until either the second AI model is convinced by the first AI model&#x27;s explanation or the first AI model has implemented all the changes requested by the second AI model.<p>It&#x27;s super clunky but has given pretty good results in the cases where I tried it lol
评论 #43839530 未加载
评论 #43838163 未加载
评论 #43838897 未加载
评论 #43838522 未加载
评论 #43837692 未加载
评论 #43838677 未加载
评论 #43837680 未加载
评论 #43838493 未加载
评论 #43839444 未加载
评论 #43841323 未加载
评论 #43840124 未加载
Lerc大约 1 个月前
I kind of want to try something like this at a larger scale in an always-on mode where I have a &#x27;senate&#x27; of debate. Rather than responding to prompts on a case by case basis, provide a list of tasks (potentially with deadlines) and let the senate work on them, break off into groups to manage subtasks, challenge results , make suggestions. Even potentially a tree of analysts where suggestions only gets passed up the tree when the parent node thinks a lower analysis is particularly insightful.<p>I definitely think that directing models to approach a problem from a specific perspective can generate better or worse results. Creating a diverse set of perspectives along with critical analysis of their results should be able to produce some impressive results.<p>Things like this would generate a massive number of tokens, but the cost per token is definitely heading in the right direction to allow for this. There is also the possibility of setting up an AI only IRC server where anybody can connect their own models for a shared debating chamber.
评论 #43837003 未加载
评论 #43837562 未加载
评论 #43836867 未加载
评论 #43837385 未加载
评论 #43837623 未加载
评论 #43843025 未加载
评论 #43839732 未加载
评论 #43838964 未加载
评论 #43837046 未加载
cube2222大约 1 个月前
This is really cool!<p>One strategy I often use (which is much simpler and more limited than this), is to finish my message with: “Please do a round of thinking in &lt;thinking&gt;&lt;&#x2F;thinking&gt; tags, then a round of self-critique in &lt;critique&gt;&lt;&#x2F;critique&gt; tags, and then a final round of &lt;thinking&gt;, before responding.”<p>It works very well. Similarly just asking it to “find the 5 biggest issues with its proposal” works pretty good (the 5 forcing it to find <i>something</i>, even if it’s mostly irrelevant).
评论 #43838977 未加载
评论 #43836316 未加载
评论 #43838152 未加载
electroly大约 1 个月前
This seems to be different than I expected from the title. I thought it would be explicitly adversarial.<p>1. You are the assistant. Please answer the question directly.<p>2. You are the cross-examiner. The assistant is wrong. Explain why.<p>3. You are the assistant. The cross-examiner is wrong. Defend your claim.<p>4. You are a judge. Did either party make their case, or is another round of argumentation required?<p>I haven&#x27;t tried this. No idea if it works. But I find it&#x27;s helpful to ask ChatGPT, in separate prompts, &quot;XYZ is true, explain why&quot; and &quot;XYZ is false, explain why&quot; and see which one seems more convincing.
评论 #43837482 未加载
评论 #43837815 未加载
评论 #43838329 未加载
评论 #43836834 未加载
hnuser123456大约 1 个月前
I&#x27;m having a lot of fun experimenting with stuff like this. I&#x27;m trying to put together an unrealengine blueprints style graph editor to allow people to design workflows like this where you start with the user prompt input, which goes to one agent, which makes an initial attempt, and then that conversation history gets passed to another &quot;agent&quot; with a different system prompt telling it to be a harsh critic, but to also give a pass&#x2F;fail signal, and loop back until the critic judges pass, then send that back to the user as output. Ideally as a little website that can call your own LLM endpoints and save&#x2F;load&#x2F;share workflow graphs.<p>Mistral small 3.1 and gemma 3 feel like the first semi-competent models that can be run locally, but that competence is just a seed, and they still need to be guided with a framework that keeps them on track.<p>Try giving it python execution in a loop and tell it to explore the world. It&#x27;ll start trying to download and read news and stuff.
评论 #43835984 未加载
评论 #43836749 未加载
评论 #43835865 未加载
评论 #43836444 未加载
评论 #43836520 未加载
jedberg大约 1 个月前
We&#x27;re really going to need to figure out how to power all these GPUs with green power real quick, or we&#x27;re going to melt the planet having AIs debate with themselves on the optimal solution to tik-tac-toe...
评论 #43836887 未加载
评论 #43838700 未加载
评论 #43853498 未加载
Xcelerate大约 1 个月前
I think this is how we get ML models to come up with novel ideas. Diagonalize against all the ideas they’ve already tried and dismissed via self-argument but keep certain consistency constraints. (Obviously much easier said than done.)
评论 #43835965 未加载
评论 #43835912 未加载
albertgoeswoof大约 1 个月前
How far is this going to go? Are we going to have a team of AI agents that runs a scrum team and meets for stand ups every couple of hours?<p>Are we going to replicate government bureaucracy with agents all debating topics all day long to find the best opinion?
评论 #43841666 未加载
评论 #43838630 未加载
评论 #43842615 未加载
faramarz大约 1 个月前
That&#x27;s cool! thanks for making it easy to fork and play with this!<p>I&#x27;ve just begun my own iteration of adding Nash Equilibrium (NECoRT?) and reframing the &quot;prompt engineering&quot; to be a multi-agent negotiation. Curious what others think? <a href="https:&#x2F;&#x2F;github.com&#x2F;faramarz&#x2F;NECoRT&#x2F;">https:&#x2F;&#x2F;github.com&#x2F;faramarz&#x2F;NECoRT&#x2F;</a><p>my reasoning is that enterprise LLMs wont have any issue with the extra compute costs and would rather reconcile complex financials with various modeling optimizations.<p>I&#x27;m very new to public repo and contributions, and hope someone can point out if I&#x27;m doing it wrong.<p>my intention was to fork the ops codebase so I can test out my theory, and push as PR eventually
alexmolas大约 1 个月前
There are two examples in the repo, one with CoRT and another one without. And the one without it it&#x27;s much better than the one that uses it. Weird choice of examples...
评论 #43836996 未加载
joshstrange大约 1 个月前
I&#x27;ve thought about trying this cross-model as well. Have Claude generate something, have OpenAI check it, have Gemini check that check. Firing multiple of these in parallel.<p>There was a post here a week or so ago doing the &quot;model checking model&quot;-type thing with GH PRs IIRC that was interesting. I haven&#x27;t had a chance to play with this idea yet.
K0balt大约 1 个月前
I’ll second this. I often use a “research assistant “ and skeptical“department head” personas working together&#x2F;against each other as a research team. It works well and is occasionally hilarious, replete with the occasional HR complaint when things go off the rails. ( I typically use local uncensored models)
k2xl大约 1 个月前
I&#x27;ve done something similar for learning about a controversial topic. I ask it to act as if it is called Bob is a well informed supporter of one side (like Ukraine) and then act as if it is something named Alice who is a well informed supporter of another side (Russia) and they have to debate each other over a few prompts with a moderator named &#x27;Sue&#x27;<p>Then after a few rounds of the debate where Sue asks a bunch of questions, I ask it to go to the judges - Mark, Phil, Sarah (and I add a few personalities to each of them... Sometimes I pretend they are famous moral philosophers) and then I have them each come up with a rubric and decide who is the winner.<p>Really fun, and helps me understand different sides of issues.
评论 #43838130 未加载
caseyy大约 1 个月前
I tried something similar when Llama2 came out, pitting two assistants, who each believed the other is the user, against each other. Ultimately, it was the same model talking with itself. The system prompts for both had various instructions to disagree and criticise the opinion of the user. I provided the first message to get things started. Usually, it’s be along the lines of “nuclear proliferation is harmful to humanity”.<p>After 15 or so iterations, both assistants would keep repeating the same things and find agreement anyway. Sometimes, the chat became unhinged and useless, but 95&#x2F;100 times, it was agreement.<p>Happy someone else made it work.
评论 #43839718 未加载
评论 #43837229 未加载
评论 #43838986 未加载
bilekas大约 1 个月前
This is an interesting approach, it reminds me of YT creator actually. I&#x27;ll find the YT creator, but basically he would make some script that would play the game like a race-course, with the goal being the finish line and iterate it N number of times, the script would keep iterating until it found the fastest solution.<p>I believe they called that machine learning.. Or re-enforced training.<p>I&#x27;m being slightly facetious, but my ignorant understanding of AI these days is basically the same no ?<p><a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=SX08NT55YhA" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=SX08NT55YhA</a>
WhitneyLand大约 1 个月前
Why try this idea on base models only?<p>The whole point of reasoning models is to automatically use COT and related techniques to bring out more capabilities.<p>It would be interesting to see if this is doing anything that’s not already being exploited.
ChadMoran大约 1 个月前
Fast Agent has this as a first-class citizen called &quot;Evaluator Optimizer&quot; pattern. Where it in a loop with a defined number of max refinements judge itself and give the output a rating, demanding it improve it&#x27;s output.<p>Highly encourage others to check out Fast Agent. It has been delightful to use. It has interactive chat mode which I love and it&#x27;s really tight and easy to implement.<p><a href="https:&#x2F;&#x2F;github.com&#x2F;evalstate&#x2F;fast-agent">https:&#x2F;&#x2F;github.com&#x2F;evalstate&#x2F;fast-agent</a>
Der_Einzige大约 1 个月前
Debate as a reasoning tactic is massively undervalued. There&#x27;s tons of papers on this at places like NeurIPS, ICML, ICLR, etc.<p>Hell, even a whole quanta article. <a href="https:&#x2F;&#x2F;www.quantamagazine.org&#x2F;debate-may-help-ai-models-converge-on-truth-20241108&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.quantamagazine.org&#x2F;debate-may-help-ai-models-con...</a><p>I got to meet and talk to the authors of this paper at NeurIPS. They&#x27;re class acts!
lepisma大约 1 个月前
Debates have worked good for me while learning something new:<p><a href="https:&#x2F;&#x2F;lepisma.xyz&#x2F;2024&#x2F;10&#x2F;19&#x2F;interventional-debates-for-studying-gray-topics&#x2F;index.html" rel="nofollow">https:&#x2F;&#x2F;lepisma.xyz&#x2F;2024&#x2F;10&#x2F;19&#x2F;interventional-debates-for-st...</a><p>I believe there are researches on this too.
aaroninsf大约 1 个月前
Question: has the the adversarial approach been roled into any coding copilots&#x2F;assistant frameworks?<p>Costs of various kinds aside I&#x27;ve wanted that from assistance&#x27;s inception — with precisely the features many call out and home-roll here, difference by both model&#x2F;provider, and, &quot;role&quot;...<p>It seems like if you have the money&#x2F;compute to burn, and can live with the reasoning wall-clock time,<p>this has got to be the best approach for the foreseeable future, for a lot of specific requirements.<p>(I also have wondered if this would illuminate the edges of what modern production models are capable of, &quot;aggregating and integrating&quot; over a variety of contributions might make more clear what the limits of their abilities are.)
badmonster大约 1 个月前
Have you experimented with weighting the self-evaluations based on specific criteria (e.g., correctness, clarity, creativity), or using external validators to guide the AI’s final choice? Curious how much tuning the evaluation step impacts overall performance.
mortarion大约 1 个月前
I think Gemini 2.5 already does something similar. If you read the &quot;thinking descriptions&quot; that it outputs it often thinks about going back to older thoughts to verify and criticize.
yieldcrv大约 1 个月前
Reminds me of baby agi from 2 years ago<p>but I guess that was before chain of thought models
zekenie大约 1 个月前
I feel like itd be cool to try prompts based on an adversarial justice system… attorney agents arguing both sides, a judge ruling on “the law”—adherence to instructions etc
评论 #43838741 未加载
hu3大约 1 个月前
Here&#x27;s some related challenge I&#x27;m facing. Maybe someone can help me:<p>I also managed to make AI critique itself and that improved code generation a ton.<p>For a TypeScript backend project that runs with Bun, I tell AI to also generate and run unit tests after every code change suggested by AI.<p>How do you solve the risk of AI writting and executing unit tests with something like `rm -rf &#x2F;` and wiping your files?<p>Docker works but I like to keep things simple.<p>Deno supports revoking file access but I&#x27;d like to keep using Bun.
评论 #43844428 未加载
评论 #43838063 未加载
评论 #43838082 未加载
评论 #43838819 未加载
评论 #43839733 未加载
schnitzelstoat大约 1 个月前
I probably don&#x27;t understand the modern, complex models. But doesn&#x27;t it basically predict the next token given the context and the better models use more training data and can consider a larger context, and have more parameters to better retain information from the training data etc.<p>But the fundamental way they operate is the same - predicting the next token given previous tokens. Where&#x2F;how does reasoning happen here?
评论 #43844021 未加载
stormfather大约 1 个月前
I made a trading bot that ingested news. The prompt to assess impact was to simulate a debate between Charlie Munger and Warren Buffet on whether to invest.
评论 #43838953 未加载
thunderbong大约 1 个月前
A lot of the comments here are reminiscent of the early Google days when everyone was finding ways to search better!
j45大约 1 个月前
There appear to be no shortage of token saving attempts that can end up using more tokens, whether it&#x27;s a monthly paid plan or API.<p>Having an approach to recognize what is needed from the AI software, and anticipate how it may default to respond based on it&#x27;s programming is critical.
pkdpic大约 1 个月前
So glad to see a write up on this finally. I&#x27;m no machine learning phd but I always wondered why this wasn&#x27;t more of a thing. Like an extension of a GAN conceptually, sort of, not really at all Im sure.<p>Also I think I kind of assumed OpenAI might be doing this behind the curtain?
mritchie712大约 1 个月前
Did something similar (OverkiLLM) to this waayyyy back in August with open LLMs. I&#x27;m sure it&#x27;d work much better now:<p><a href="https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;overkillm" rel="nofollow">https:&#x2F;&#x2F;www.definite.app&#x2F;blog&#x2F;overkillm</a>
rriley大约 1 个月前
Makes me wonder what would happen if we combine LLMs with recursive genetic algorithms. Similar to <a href="https:&#x2F;&#x2F;github.com&#x2F;DivergentAI&#x2F;dreamGPT">https:&#x2F;&#x2F;github.com&#x2F;DivergentAI&#x2F;dreamGPT</a>
noworriesnate大约 1 个月前
I’ve had success telling the model it really needs to poop and if it gets to the point quickly it’ll be able to leave the meeting and go do that. It actually works amazingly well.<p>It’s also a lot more ethical than verbal abuse, which some people say improves the results as well.<p>Programming isn’t what it used to be.
评论 #43837606 未加载
ausbah大约 1 个月前
at some point this doesn’t make LLMs feel useful. I have to wait 10x as long just so my LLM can have a somewhat higher chance of actually answer my question correctly?
cwillu大约 1 个月前
Any api that lets you constrain output to a formal syntax should let you do away with the “first output a number, and only then explain yourself” boilerplate.
killerstorm大约 1 个月前
This is similar to Tree-of-Thought with self-evaluation.
daxfohl大约 1 个月前
Maybe have a &quot;reconcile&quot; option, for it to see if it can mix and match the best parts of each alternative rather than just choosing one.
grzracz大约 1 个月前
Your readme demo images are wrong: the terminal one is the non-CoRT one and the GUI one is the one with CoRT. Confused me for a while
Svoka大约 1 个月前
Oh. I was just asking &quot;Use dialectic method on your solution&quot; in the end of the prompt... It does make it think harder.
ashoeafoot大约 1 个月前
Give it reward and punishment evaluations, exploring the noise in parallel, extinction for the non rewarding answers ?
keyle大约 1 个月前
When will we get the `4o` vs `o3` background conversation in &quot;thinking&quot; leading to a more correct result?
kevinrineer大约 1 个月前
This sounds like the zeitgeist is approaching genetic algorithms, which are super fun. Adversarial stuff is great.
throwawayForMe2大约 1 个月前
I wonder if the Scholastic method of the Schoolmen would be useful with its argument and counter argument style.
alex1138大约 1 个月前
Every single one of my prompts would be &quot;Are you suuuuuuure you&#x27;re not hallucinating that?&quot;
Garlef大约 1 个月前
Similarly, letting the LLM generate a socratic dialogue can work pretty well to get deeper into a topic.
mangoman大约 1 个月前
a paper with a similar idea on scaling test time reasoning, this is sorta how all the thinking models work under the hood. <a href="https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2501.19393" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;abs&#x2F;2501.19393</a>
gnarlouse大约 1 个月前
This seems like low hanging fruit; are we seriously supposed to believe this is new and novel?
评论 #43843823 未加载
irthomasthomas大约 1 个月前
my favourite pattern rn: llm &quot;write a savage, yet grounded roast of: $content&quot; llm -c &quot;Write an equally savage rebuttal&quot; llm -c &quot;first arbitrate and then synthesize a final review.&quot;
asdfman123大约 1 个月前
And when I do this people say I&#x27;m overanalyzing
评论 #43838827 未加载
animitronix大约 1 个月前
Adversarial networks have been a thing for a while
stevefan1999大约 1 个月前
That is just reinforcement learning in disguise
akomtu大约 1 个月前
The modern Alchemy: the belief that you can extract gold (intelligence) from iron (autocomplete by imitation) by mixing iron with itself.
csours大约 1 个月前
Yes, give the computers anxiety too!
lonetripper大约 1 个月前
all this hard thinking yet humanity fails to come up with just one girlfriend for me
robofanatic大约 1 个月前
soon there will be AI debates. Different models debating with each other on a topic
mparnisari大约 1 个月前
So like rubber ducking for AI?
评论 #43837151 未加载
评论 #43837035 未加载
jbellis大约 1 个月前
does it actually make a difference to do M rounds of N vs one round of M*N?
评论 #43839747 未加载
firgrove大约 1 个月前
this is amazing - I love seeing novel approaches to optimizing
celltalk大约 1 个月前
One of my doctoral propositions is, dialog leads to true artificial intelligence.
getcrunk大约 1 个月前
Hello cnn’s
parrit大约 1 个月前
I want to see &quot;Meh&quot; vs. &quot;Holy crap&quot; as a benchmark in a paper published by Google. Or more likely I suspect, Andrej.
codr7大约 1 个月前
Better yet, let it argue with another AI, preferably using voice; instant entertainment.
antisthenes大约 1 个月前
Cool. Now I can justify talking to myself.
m3kw9大约 1 个月前
Isn’t this best of n?
评论 #43836481 未加载
lenerdenator大约 1 个月前
I, too, like to give Terminator lite anxiety.
hansmayer大约 1 个月前
Right, so... but you do realise its still just producing random output based on how you reconfigured it&#x27;s weights, right? Sometimes it will happen to resonate with what you need. But it still neither thinking nor arguing with itself.
DyslexicAtheist大约 1 个月前
&gt; <i>&quot;I made my AI think&quot;</i> ...<p>utterly moronic.<p>They don&#x27;t “think” ... not even in the most autistic sense of the word.<p>They can generate solutions by combining existing knowledge in unique ways. But they don&#x27;t “think”.
评论 #43842516 未加载