TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

A Revolution in How Robots Learn

82 点作者 jsomers6 个月前

8 条评论

m_ke6 个月前
I did a review of state of the art in robotics recently in prep for some job interviews and the stack is the same as all other ML problems these days, take a large pretrained multi modal model and do supervised fine tuning of it on your domain data.<p>In this case it&#x27;s &quot;VLA&quot; as in Vision Language Action models, where a multimodal decoder predicts action tokens and &quot;behavior cloning&quot; is a fancy made up term for supervised learning, because all of the RL people can&#x27;t get themselves to admit that supervised learning works way better than reinforcement learning in the real world.<p>Proper imitation learning where a robot learns from 3rd person view of humans doing stuff does not work yet, but some people in the field like to pretend that teleoperation and &quot;behavior cloning&quot; is a form of imitation learning.
评论 #42252437 未加载
评论 #42251265 未加载
评论 #42255911 未加载
drcwpl6 个月前
One particularly fascinating aspect of this essay is the comparison between human motor learning and robotic dexterity development, particularly the concept of “motor babbling.” The author highlights how babies use seemingly random movements to calibrate their brains with their bodies, drawing a parallel to how robots are being trained to achieve precise physical tasks. This framing makes the complexity of robotic learning, such as a robot tying shoelaces or threading a needle, more relatable and underscores the immense challenge of replicating human physical intelligence in machines. For me it is also a vivid reminder of how much we take our own physical adaptability for granted.
评论 #42256754 未加载
x11antiek6 个月前
<a href="https:&#x2F;&#x2F;archive.is&#x2F;fsuxe" rel="nofollow">https:&#x2F;&#x2F;archive.is&#x2F;fsuxe</a>
ratedgene6 个月前
Hey, I wonder if we can use LLMs to learn learning patterns, I guess the bottleneck would be the curse of dimensionality when it comes to real world problems, but I think maybe (correct me if I&#x27;m wrong) geographic&#x2F;domain specific attention networks could be used.<p>Maybe it&#x27;s like:<p>1. Intention, context 2. Attention scanning for components 3. Attention network discovery 4. Rescan for missing components 5. If no relevant context exists or found 6. Learned parameters are initially greedy 7. Storage of parameters gets reduced over time by other contributors<p>I guess this relies on there being the tough parts: induction, deduction, abductive reasoning.<p>Can we fake reasoning to test hypothesis that alter the weights of whatever model we use for reasoning?
评论 #42250789 未加载
Animats6 个月前
A research result reported before, but, as usual, the New Yorker has better writers.<p>Is there something which shows what the tokens they use look like?
评论 #42263254 未加载
josefritzishere6 个月前
There&#x27;s a big asterisk on the word &quot;learn&quot; in that headline.
codr76 个月前
Oh my, that has to be one of the worst jobs ever invented.
nobodywillobsrv6 个月前
Anyone find is suspcious that all these paywalled fluff tech legacy media articles keep on ending up on hn? Feels like an op. Who in tech actually reads NYT for example?
评论 #42257657 未加载