TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Teaching Robots to Understand Semantic Concepts

137 点作者 rey12rey将近 8 年前

3 条评论

skywhopper将近 8 年前
The headline seems to misuse the word &quot;semantic&quot; (not to mention &quot;understand&quot;). Does the door-opening robot now understand how to open all hinged doors with a similar opening mechanism? Or was it just trained to imitate a sequence of changes in a 2D image from a fixed angle? Can the same software and robot also be taught to open windows? Boxes? We are talking about &quot;semantics&quot; explicitly here. Does it understand &quot;open&quot; versus &quot;closed&quot; for these different types of closures&#x2F;portals?<p>I don&#x27;t want to discount the value of this research. It&#x27;s absolutely necessary to do this sort of basic proof-of-concept testing of these ideas. But the claim being made implicitly here is way beyond what&#x27;s actually going on. The software understands nothing, and the &quot;semantics&quot; extend to simple image-matching of objects, but there&#x27;s no deeper meaning associated with the labels, so I think calling that &quot;semantics&quot; is a major stretch.<p>This approach is not going to teach a robot how to pick fruit, or serve food, or clean floors anytime soon. In the best case where this is even a workable approach, research like this is just the first of millions more tiny steps along the path. Anyway I think it&#x27;s naive to assume that a good way to approach automation is to write software to let robots learn by watching humans do the desired task. As cool as that sounds, chances are that approach would ultimately be a massively inefficient way to solve the problem. It&#x27;d be like trying to invent the automobile by building a steam-powered horse robot that can tow carriages. The critical purpose is being overlooked in favor of a cool-looking but totally impractical toy demo.
评论 #14823875 未加载
评论 #14824337 未加载
评论 #14824783 未加载
QAPereo将近 8 年前
To reference an earlier article on HN, this reads like a future of, &quot;Robot, observe these field workers picking fruit for a week. Now practice in this field for a day. OK, now the job is yours.<p>What I can&#x27;t tell at all from this article is whether that day is years or decades away.
评论 #14823755 未加载
评论 #14825413 未加载
评论 #14823743 未加载
评论 #14823633 未加载
评论 #14823783 未加载
评论 #14823637 未加载
richard___将近 8 年前
Summary--<p>Semantics part: Seems like the idea is we can &quot;transfer&quot; knowledge from prior labeled samples so that we don&#x27;t need to do as much new work labeling sample images with semantic labels.<p>Grasping part: &quot;Emulating human movements with self-supervision and imitation.&quot; High-level imitation based on visual frame differences avoids needing to manually control actuators. Not sure how this works exactly<p>Two-stream model: ventral network asks What class, dorsal network asks Is this how we should grasp this object. The benefit is that we can make use of all the automatically generated (robot-generated) grasping data without having a human supervise all that automated grasping, e.g. &quot;This process is a successful way to pickup this object, and also this object is an apple.&quot; The ventral network ties back this the grasping data (without object labels) to object labels, which allows for semantic control of the trained robot e.g. &quot;Pickup that apple&quot;.