TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: What do people mean when they say transformers aren't fully understood?

3 点作者 olivierestsage超过 1 年前
I am not a computer scientist (let alone an ML specialist), just one of many interested observers trying to understand how LLMs work. I am familiar with many of the sources that are recommended for coming to terms with the transformer model, such as &quot;The Illustrated Transformer.&quot;<p>However, in online discussions about transformers -- including on this site -- I frequently see references made to some kind of mysterious, unknown element; basically, the idea that &quot;we don&#x27;t REALLY know how transformers produce such good results, at least in detail.&quot;<p>My question is: is that claim truly accurate? I am struggling to understand how a technology that is not fully understood, at least by specialists in the ML field, could be harnessed to such great effect. Thank you!

1 comment

rvz超过 1 年前
&gt; I frequently see references made to some kind of mysterious, unknown element; basically, the idea that &quot;we don&#x27;t REALLY know how transformers produce such good results, at least in detail.&quot; My question is: is that claim truly accurate?<p>It is the same reason why almost no one trusts LLMs to fully pilot a plane with zero humans on board.<p>These systems have a low amount of trust for these use-cases because they fundamentally lack understanding and it is on the basis of being unable to explain themselves transparently. Even when LLMs attempt to, they are able to convince the untrained eye with their wrong answers. Not even the AI researchers can explain why some of these AI systems get confused over bad input or a adversarial input.<p>There is research into trying to &#x27;understand&#x27; these systems for decades but just falls short into demonstrating any consistent evidence of showing transparency in these black-box systems such as LLMs; making them unaccountable.