TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Bitter Lesson is about AI agents

139 点作者 ankit219大约 2 个月前

23 条评论

noosphr大约 2 个月前
For a blog post of 1,200 words the bitter lesson has done more damage to AI research and funding than blowing up a nuclear bomb at neurips would.<p>Every time I try to write a reasonable blog post about why it&#x27;s wrong it blows up to tens of thousands of words and no one can be bothered to read it, let alone the supporting citations.<p>In the spirit of low effort anec-data pulled from memory:<p>The raw compute needed to brute force any problem can only be known after the problem is solved. There is no sane upper limit to how much computation, memory and data any given task will take and humans are terrible at estimating how hard tasks actually are. We are after all only 60 years late for the undergraduate summer project that would solve computer vision.<p>Today VLMs are the best brute force approach to solving computer vision we have, and they look like they will take a PB of state to solve and the compute needed to train them will be available some time around 2040.<p>What do we do with the problems that are too hard to solve with the limited compute that we have? Lie down for 80 years and wait for compute to catch up? Or solve a smaller problem using specialized tricks that don&#x27;t require a $10B super computer to build?<p>The bitter lesson is nothing of the sort, there is plenty of space for thinking hard, and there always will be.
评论 #43456845 未加载
评论 #43457142 未加载
评论 #43456961 未加载
评论 #43456829 未加载
评论 #43457393 未加载
评论 #43458260 未加载
评论 #43466857 未加载
评论 #43457220 未加载
评论 #43456200 未加载
lsy大约 2 个月前
Going back to the original &quot;Bitter Lesson&quot; article, I think the analogy to chess computers could be instructive here. A lot of institutional resources were spent trying to achieve &quot;superhuman&quot; chess performance, it was achieved, and today almost the entire TAM for computer chess is covered by good-enough Stockfish, while most of the money tied up in chess is in matching human players with each other across the world, and playing against computers is sort of what you do when you&#x27;re learning, or don&#x27;t have an internet connection, or you&#x27;re embarrassed about your skill and don&#x27;t want to get trash-talked by an Estonian teenager.<p>The &quot;Second Bitter Lesson&quot; of AI might be that &quot;just because massive amounts of compute make something <i>possible</i> doesn&#x27;t mean that there will be a commensurately massive market to justify that compute&quot;.<p>&quot;Bitter Lesson&quot; I think also underplays the amount of energy and structure and design that has to go into compute-intensive systems to make them succeed: Deep Blue and current engines like Stockfish take advantage of tablebases of opening and closing positions that are more like GOFAI than deep tree search. And the current crop of LLMs are not only taking advantage of expanded compute, but of the hard-won ability of companies in the 21st century to not only build and resource massive server farms, but mobilize armies of contractors in low-COL areas to hand-train models into usefulness.
评论 #43455249 未加载
评论 #43455626 未加载
PollardsRho大约 2 个月前
The time span on which these developments take place matter a lot for whether the bitter lesson is relevant to a particular AI deployment. The best AI models of the future will not have 100K lines of hand-coded edge cases, and developing those to make the models of today better won&#x27;t be a long-term way to move towards better AI.<p>On the other hand, most companies don&#x27;t have unlimited time to wait for improvements on the core AI side of things, and even so building competitive advantages like a large existing customer base or really good private data sets to train next-gen AI tools have huge long-term benefits.<p>There&#x27;s been an extraordinary amount of labor hours put into developing games that could run, through whatever tricks were necessary, on whatever hardware actually existed for consumers at the time the developers were working. Many of those tricks are no longer necessary, and clearly the way to high-definition real-time graphics was not in stacking 20 years of tricks onto 2000-era hardware. I don&#x27;t think anyone working on that stuff actually thought that was going to happen, though. Many of the companies dominating the gaming industry now are the ones that built up brands and customers and experience in all of the other aspects of the industry, making sure that when better underlying scaling came there they had the experience, revenue, and know-how to make use of that tooling more effectively.
评论 #43456026 未加载
评论 #43457370 未加载
abstractcontrol大约 2 个月前
&gt; Investment Strategy: Organizations should invest more in computing infrastructure than in complex algorithmic development.<p>&gt; Competitive Advantage: The winners in AI won’t be those with the cleverest algorithms, but those who can effectively harness the most compute power.<p>&gt; Career Focus: As AI engineers, our value lies not in crafting perfect algorithms but in building systems that can effectively leverage massive computational resources. That is a fundamental shift in mental models of how to build software.<p>I think the author has a fundamental misconception what making best use of computational resources requires. It&#x27;s algorithms. His recommendation boils down to not do the one thing that would allow us to make the best use of computational resources.<p>His assumptions would only be correct if all the best algorithms were already known, which is clearly not the case at present.<p>Rich Sutton said something similar, but when he said it, he was thinking of old engineering intensive approaches, so it made sense in the context in which he said it and for the audience he directed it at. It was hardly groundbreaking either, the people whom he wrote the article for all thought the same thing already.<p>People like the author of this article don&#x27;t understand the context and are taking his words as gospel. There is no reason not to think that there won&#x27;t be different machine learning methods to supplant the current ones, and it&#x27;s certain they won&#x27;t be found by people who are convinced that algorithmic development is useless.
评论 #43455997 未加载
评论 #43455722 未加载
serjester大约 2 个月前
This misses that if the agent is occasionally going haywire, the user is leaving and never coming back. AI deployments are about managing expectations - you’re much better off with an agent that’s 80 +&#x2F;- 10% successful than 90 +&#x2F;- 40%. The more you lean into full automation, the more guardrails you give up and the more variance your system has. This is a real problem.
评论 #43454291 未加载
评论 #43455149 未加载
评论 #43454265 未加载
dtagames大约 2 个月前
Good stuff but the original &quot;Bitter Lesson&quot; article has the real meat, which is that by applying more compute power we get better results (just more accurate token predictions, really) than with human guiderails.
typon大约 2 个月前
The counter argument is a bitter lesson that Tesla is learning from Waymo and the lesson might be bitter enough to tank the company. Waymo&#x27;s approach to self driving isn&#x27;t end to end - they have classical control combined with tons of deep learning, creating a final product that actually works in the real world, meanwhile the purely data driving approach from Tesla has failed to deliver a working product.
评论 #43461477 未加载
评论 #43455811 未加载
评论 #43454594 未加载
评论 #43460891 未加载
extr大约 2 个月前
I bring this up often at work. There is more ROI in assuming models will continue to improve, and planning&#x2F;engineering with that future in mind, rather than using a worse model and spending a lot of dev time shoring up it&#x27;s weaknesses, prompt engineering, etc. The best models today will be cheaper tomorrow. The worst models today will literally cease to exist. You want to lean into this - have the AI handle as much as it possibly can.<p>Eg: We were using Flash 1.5 for awhile. Spent a lot of time prompt engineering to get it to do exactly what we wanted and be more reliable. Probably should have just done multi-shot and said &quot;take best of 3&quot;, because as soon as Flash 2.0 came out, all the problems evaporated.
评论 #43455103 未加载
评论 #43457259 未加载
xg15大约 2 个月前
It&#x27;s not wrong, but I find the underlying corrolay pretty creepy that actually trying to understand those problems and fix errors at edge cases is also a fool&#x27;s errand, because why try to understand a specific behavior if you can just (try to) finetune it away?<p>So we&#x27;ll have to get used for good to a future where AI is unpredictable, usually does what you want, but has a 0.1% chance of randomly going haywire and no one will know how to fix it?<p>Also, the focus on hardware seems to imply that it&#x27;s strictly a game of capital - who has access to the most compute resources wins, the others can stop trying. Wouldn&#x27;t this lead to massive centralization?
评论 #43456457 未加载
moojacob大约 2 个月前
&gt; For instance, in customer service, an RL agent might discover that sometimes asking a clarifying question early in the conversation, even when seemingly obvious, leads to much better resolution rates. This isn’t something we would typically program into a wrapper, but the agent found this pattern through extensive trial and error. The key is having enough computational power to run these experiments and learn from them.<p>I am working on a gpt wrapper in customer support. I’ve focused on letting the LMs do what they do best, which is writing responses using context. The human is responsible for managing the context instead. That part is a much harder problem than RL folks expect it to be. How does your AI agent know all the nuance of a business? How does it know you switched your policy on returns? You’d have to have a human sign off on all replies to customer inquiries. But then, why not make an actual UI at that point instead of an “agent” chatbox.<p>Games are simple, we know all the rules. Like chess. Deepmind can train on 50 million games. But we don’t know all the rules in customer support. Are you going to let an AI agent train itself on 50 million customer interactions and be happy with it sucking for the first 20 million?
评论 #43455574 未加载
elicksaur大约 2 个月前
&gt; For instance, in customer service, an RL agent might discover that sometimes asking a clarifying question early in the conversation, even when seemingly obvious, leads to much better resolution rates.<p>Why does this read to me as the bot finding a path of “Annoy the customer until they hang up and mark the case as solved.” ?
patcon大约 2 个月前
YES to the nature analogy.<p>We are not guaranteed a world pliable to our human understanding. The fact that we feel entitled to such things is just a product of our current brief moment in the information-theoretic landscape, where humans have created and have domination over most of the information environment we navigate. This is a rare moment for any actor. Most of our long history has been spent in environments that are unmanaged ecologies that have blossomed around any one actor.<p>imho neither we nor any single AI agent will understand the world as fully as we do. We should retire the idea that we are destined to be privileged to that knowledge.<p><a href="https:&#x2F;&#x2F;nodescription.net&#x2F;notes&#x2F;#2021-05-04" rel="nofollow">https:&#x2F;&#x2F;nodescription.net&#x2F;notes&#x2F;#2021-05-04</a>
advael大约 2 个月前
The bitter lesson is a good demonstration of how people have really short memories and distributed work loses information<p>Every AI &quot;breakthrough&quot; comes at a lag because the people who invent a new architecture or method aren&#x27;t the ones who see its full potential. Because of the financial dynamics at play, the org or team that sees the crazy-looking result often didn&#x27;t invent the pieces they used. Even if they did, it&#x27;s been years and in a fast-moving field that stuff has already started to feel &quot;standard&quot; and &quot;generic&quot;. The real change they saw was something like more compute or more data<p>Basically, even the smartest people in the world are pretty dumb, in the sense of generalizing observations poorly
t_mann大约 2 个月前
Has anyone empirically assessed the claims of the Bitter Lesson? The article may sound convincing, but ultimately it&#x27;s just a few anecdotes. It seems to have a lot of &#x27;cultural&#x27; impact in AI research, so it would be good to have some structured data-based analysis before we dismiss entire research directions.
评论 #43457444 未加载
评论 #43457011 未加载
sgt101大约 2 个月前
I don&#x27;t get how RL can be applied in a domain where there is no simulator.<p>So for customer service, to do RL on real customers... well this sounds like it&#x27;s going to be staggeringly slow and very expensive in terms of peeved customers.
评论 #43456947 未加载
gpapilion大约 2 个月前
More generally beats better. That’s the continual lesson from data intensive workloads. More compute, more data, more bandwidth.<p>The part that I’ve been scratching my head at is whether we see a retreat from aspects of this due to the high costs associated with it. For cpu based workloads this was a workable solution, since the price has been reducing. gpus have generally scaled pricing as a constant of available flops, and the current hardware approach equates to pouring in power to achieve better results.
TylerLives大约 2 个月前
It&#x27;s actually about LLMs. They&#x27;re fundamentally limited by our preconceptions. Can we go back to games and AlphaZero?
ed大约 2 个月前
It’d be nice if this post included a high-level cookbook for training the 3rd approach. The hand-waving around RL sounds great, but how do you accurately simulate a customer for learning at scale?
评论 #43456630 未加载
dangus大约 2 个月前
I think this goes for almost all software. Hardware is still getting impressively faster every year despite moore’s law expiring.
RachelF大约 2 个月前
I think an even more bitter lesson is coming very soon: AI will run out of human-generated content to train on.<p>Already AI companies are probably training AI with AI generated slop.<p>Sure there will be tweaks etc, but can we make it more intelligent than its teachers?
评论 #43456739 未加载
评论 #43458657 未加载
Foreignborn大约 2 个月前
please, please stop letting AI rewrite for you. i’m so tired of reading AI slop.<p>instead, ask it to be a disagreeable editor and have it ask questions about your draft. you’ll go so much further, and the writing won’t be nonsensical.
cainxinth大约 2 个月前
&gt; My plants don’t need detailed instructions to grow. Given the basics (water, sunlight, and nutrients), they figure out the rest on their own.<p>They do need detailed instructions to grow. The instructions are encoded in their DNA, and they didn’t just figure them out in real time.
_wire_大约 2 个月前
If only artificial intelligence was intelligent!<p>Oh, well...