TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: What's the practical use of larger context LLMs?

3 点作者 stevemadere将近 2 年前
I see that lots of folks are working on building LLMs that can handle more context without breaking the bank on GPU.<p>Is there a real practical reason for that right now or is it just something that everybody agrees is obvious without economic justification?

2 条评论

ftxbro将近 2 年前
So they have had LLMs with small contexts like one or two words or a dozen letters for a long time, ever since like Laplace or Shannon or Markov. They were called Markov chains. No one really guessed this (although it was known to be theoretically possible in the sense of ai-completeness), but it turns out that longer ones turn out to even in practice unlock so many cognitive capabilities bordering on superhuman. If this is the main difference between the Markov chains that they have been using for autocomplete for decades versus the ones that will beat you at the GREs or the bar exams or every AP test, then it is natural they are curious what happens when they make the context even longer.
评论 #36562708 未加载
seanthemon将近 2 年前
Longer context means more memory, effectively a longer history the LLM remembers. One issue i&#x27;m having is say functions works wonderfully, but context window is tight even with 16k tokens, with a bigger context, sky is the limit.
评论 #36562768 未加载