TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: Anyone working on something better than LLMs?

42 点作者 xucian大约 1 年前
if you think about it, next-token prediction is just stupid. it&#x27;s so resource intensive<p>yet, it&#x27;s mimicking emergent thought quite beautifully. it&#x27;s shockingly unintuitive how a simple process scaled enormously can lead to this much practical intelligence (practical in the sense that&#x27;s useful, but it&#x27;s not the way we think). I&#x27;m aware there are multiple layers, filters, processes etc., I&#x27;m just talking about the foundation, which is next-token prediction.<p>when I first heard that it&#x27;s not predicting words, but parts of words, I immediately saw a red flag. yes, there are compounded words like strawberry (straw + berry) and you can capture meaning at a higher-resolution, but most words are not compounded, and just in general we&#x27;re trying to simulate meaning instead of &#x27;understanding&#x27; it. &#x27;understanding&#x27; simply means knowing a man is to a woman what a king is to a queen, but without the need to learn about words and letters (that should be just an interface).<p>I feel we&#x27;re yet to discover the &quot;machine code&quot; for ASI. it&#x27;s like we have no compiler, but we directly interpret code. imagine the speed-ups if we could just spare the processor from understanding our stupid, inefficient language.<p>I&#x27;d really like to see a completely new approach working in the Meaning Space, which transcends the imperfect Language Space. This will require lots of data pre-processing, but it&#x27;s a fun journey -- basically a parser human-machine and machine-human. I&#x27;m sure I&#x27;m not the first one thinking about it<p>so what we&#x27;ve got so far?

12 条评论

plaidfuji大约 1 年前
As others have noted, Yann LeCun is looking beyond autoregressive (next-token-prediction) models. Here’s one of his slide decks that raises some interesting new concepts:<p><a href="https:&#x2F;&#x2F;drive.google.com&#x2F;file&#x2F;d&#x2F;1BU5bV3X5w65DwSMapKcsr0ZvrMRU_Nbi&#x2F;view?pli=1" rel="nofollow">https:&#x2F;&#x2F;drive.google.com&#x2F;file&#x2F;d&#x2F;1BU5bV3X5w65DwSMapKcsr0ZvrMR...</a>
评论 #40353194 未加载
评论 #40330510 未加载
BugsJustFindMe大约 1 年前
&gt; <i>it&#x27;s shockingly unintuitive how a simple process scaled enormously can lead to this much practical intelligence</i><p>A biological neuron doesn&#x27;t do much. On its own, a simple process. Yet when you put a 100 billion of them together in the right 1000-connected configuration you get a human brain.
评论 #40328684 未加载
评论 #40324070 未加载
stevenAthompson大约 1 年前
&gt; in general we&#x27;re trying to simulate meaning instead of &#x27;understanding&#x27; it. &#x27;understanding&#x27; simply means knowing a man is to a woman what a king is to a queen, but without the need to learn about words and letters (that should be just an interface).<p>I have no idea what I&#x27;m talking about, but what you describe is exactly what LLM&#x27;s do.<p>Words are tokens that represent concepts. We&#x27;ve found a way to express the relationships between many tokens in a giant web. The tokens are defined by their relationships to each other. Changing the tokens we use probably won&#x27;t make much more difference than changing the language the LLM is built from.<p>We could improve the method we use to store and process those relationships, but it will still be fundamentally the same idea: Large webs of inter-related tokens representing concepts.
评论 #40329075 未加载
jtietema大约 1 年前
I think you might find Lex Fridmans interview with Yann LeCun interesting[1]. It discusses exactly this, how LLMs just mimmick intelligent behaviour but have no understanding of the world at all. It also discusses other approaches we should look at instead of current LLMs.<p>[1] <a href="https:&#x2F;&#x2F;youtu.be&#x2F;5t1vTLU7s40" rel="nofollow">https:&#x2F;&#x2F;youtu.be&#x2F;5t1vTLU7s40</a>
评论 #40323803 未加载
评论 #40328974 未加载
exe34大约 1 年前
&gt; understanding&#x27; simply means knowing a man is to a woman what a king is to a queen,<p>Turns out this is beautifully represented by embeddings alone!
评论 #40329878 未加载
waldrews大约 1 年前
Meaning Space transcending the imperfect Language Space? Yes, there&#x27;s been some recent thinking in this direction, e.g. Zhuangzi, &quot;Words exist because of meaning. Once you&#x27;ve gotten the meaning, you can forget the words. Where can I find a man who has forgotten words so I can talk with him?&quot;
评论 #40329176 未加载
jncfhnb大约 1 年前
&gt; understanding&#x27; simply means knowing a man is to a woman what a king is to a queen, but without the need to learn about words and letters (that should be just an interface).<p>Citation needed
评论 #40329312 未加载
评论 #40324025 未加载
quadrature大约 1 年前
&gt;Yes, there are compounded words like strawberry (straw + berry) and you can capture meaning at a higher-resolution, but most words are not compounded<p>What&#x27;s really cool about tokenization is that it breaks down words based on how often parts of the word are used. This helps a lot with understanding different forms of words, like when you add &quot;-ing&quot; to a verb, make words plural, or change tenses. It&#x27;s like seeing language as a bunch of building blocks.
评论 #40329720 未加载
xnx大约 1 年前
Does anyone have a guess what angle John Carmack is working on with Keen Technologies? <a href="https:&#x2F;&#x2F;dallasinnovates.com&#x2F;john-carmacks-keen-technologies-partners-to-accelerate-development-of-artificial-general-intelligence&#x2F;" rel="nofollow">https:&#x2F;&#x2F;dallasinnovates.com&#x2F;john-carmacks-keen-technologies-...</a>
评论 #40329766 未加载
samus大约 1 年前
Look no further than here for decoding more than one following token:<p><a href="https:&#x2F;&#x2F;hao-ai-lab.github.io&#x2F;blogs&#x2F;cllm&#x2F;" rel="nofollow">https:&#x2F;&#x2F;hao-ai-lab.github.io&#x2F;blogs&#x2F;cllm&#x2F;</a>
评论 #40330600 未加载
gaganyaan大约 1 年前
I don&#x27;t think you quite understand how tokenization works. Try typing &quot;strawberry&quot; in here:<p><a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer</a><p>Tokens aren&#x27;t just individual parts of compound words, they&#x27;re sliced up in a way that&#x27;s convenient statistically. The tokenizer has each individual character as a token, so it could be purely character-based if desired, it&#x27;s just easier to compute when some common sequences like &quot;berry&quot; are represented by a single token. Try typing &quot;strawberry&quot; into the tokenizer and see it tokenized as &quot;str&quot;, &quot;aw&quot;, and &quot;berry&quot;.<p>Also, next token prediction is not stupid. A &quot;sufficiently advanced&quot; next token predictor must be at least as intelligent as a human, if it could predict any humans&#x27; next token in any scenario. Obviously, we&#x27;re not there yet, but there&#x27;s no reason to think right now that next token prediction will face any sort of limitation. Especially with new models coming out that are seeing better perfomance purely from training them much longer on the same datasets.
评论 #40329807 未加载
byyoung3大约 1 年前
its not stupid.
评论 #40330000 未加载