TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Models with Large Context Window

4 点作者 Roshni1990r大约 1 年前
Recently Gradient AI announced that they released the Llama 3 models with a 1M context window for both 8B and 70B sizes. And now they just dropped a 4M context window for the 8B size.<p>what do you think about this and models with Large Context Window?

2 条评论

PaulHoule大约 1 年前
I use SBERT-type models for ranking, retrieval, classification, etc.<p>These work properly for a document that fits in the context window but not for a document which is larger. You can cut the document into smaller pieces but it just isn’t the same.<p>I would like to see longer context models like that and could accept some inference cost for it.<p>I think the summarization-type use cases will benefit from the context window but the computational complexity of a problem posed can grow explosively as a function of problem size: so x10 context window might not really mean it can do a x10 sized problem consistently.
stainlu大约 1 年前
I am using moonshot ai they cliamed to have a 2m context window