TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Extending the context length to 1M tokens

116 pointsby cmcconomy6 months ago

5 comments

aliljet6 months ago
This is fantastic news. I&#x27;ve been using Qwen2.5-Coder-32B-Instruct with Ollama locally and it&#x27;s honestly such a breathe of fresh air. I wonder if any of you have had a moment to try this newer context length locally?<p>BTW, I fail to effectively run this on my 2080 ti, I&#x27;ve just loaded up the machine with classic RAM. It&#x27;s not going to win any races, but as they say, it&#x27;s not the speed that matter, it&#x27;s the quality of the effort.
评论 #42177831 未加载
评论 #42176314 未加载
评论 #42175226 未加载
lr19706 months ago
&gt; We have extended the model’s context length from 128k to 1M, which is approximately 1 million English words<p>Actually English language tokenizers map on average 3 words into 4 tokens. Hence 1M tokens is about 750K English words not a million as claimed.
评论 #42179262 未加载
评论 #42179102 未加载
lostmsu6 months ago
Is this model downloadable?
评论 #42175492 未加载
swazzy6 months ago
Note unexpected three body problem spoilers in this page
评论 #42175014 未加载
评论 #42175102 未加载
anon2916 months ago
Can we all agree that these models far surpass human intelligence now? I mean they process hours worth of audio in less time than it would take a human to even listen. I think the singularity passed and we didn&#x27;t even notice (which would be expected)
评论 #42174987 未加载
评论 #42176838 未加载
评论 #42175002 未加载
评论 #42175223 未加载
评论 #42175324 未加载
评论 #42175019 未加载
评论 #42175008 未加载
评论 #42174949 未加载
评论 #42175171 未加载
评论 #42175118 未加载
评论 #42175095 未加载