TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Increase GPT4's context size by asking it to compress your prompts

40 pointsby stolenmerchabout 2 years ago

10 comments

danShumwayabout 2 years ago
If you look into this into more detail:<p>A) people have had a hard time reproducing it, and<p>B) more damning, the &quot;compressed&quot; version uses <i>more</i> tokens than the original (<a href="https:&#x2F;&#x2F;gist.github.com&#x2F;VictorTaelin&#x2F;d293328f75291b23e203e9d9db9bd136" rel="nofollow">https:&#x2F;&#x2F;gist.github.com&#x2F;VictorTaelin&#x2F;d293328f75291b23e203e9d...</a>)<p>&gt; In practice, plain, well-designed summaries should be optimal to fit larger documents in the context.<p>&gt; This concept has potential, though; building lookup tables seems to outperform long text summarization.<p>It&#x27;s a clever idea, and I agree that lookup tables and external storage of memory is likely going to be important at some point, but I suspect that&#x27;s going to come out of giving LLM more ability to externally reference &quot;long-term&quot; memory rather than compressing everything into immediate context.
评论 #35453894 未加载
评论 #35463995 未加载
Imnimoabout 2 years ago
A more accurate title would be &quot;decrease GPT4&#x27;s context size by asking it to obfuscate your prompts&quot;. You&#x27;re getting a prompt that consumes more tokens and isn&#x27;t particularly faithful to the original.
sippeangeloabout 2 years ago
One reply points out that even though the text length is 1&#x2F;4 of the original, the “compressed” output is actually LONGER in token count!<p><a href="https:&#x2F;&#x2F;twitter.com&#x2F;dogeofcoin&#x2F;status&#x2F;1642918892602290179" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;dogeofcoin&#x2F;status&#x2F;1642918892602290179</a>
daydreamabout 2 years ago
A promising and potentially useful approach for adding additional context to your prompts.<p>How likely is it that context size will greatly increase in the coming year or two? Are there fundamental limits, or could we reasonably expect greatly increased context size in the future?
评论 #35455160 未加载
Alex_Bellabout 2 years ago
ChatGPT 3.5 in response to a brief analysis of the community gave out: &quot;It seems like the conversation you provided is discussing the idea of compressing prompts or summaries of larger documents in order to fit them into a smaller context. While the idea has potential, it appears that people have had difficulty reproducing it and that lookup tables and external storage of memory may be more effective in the long term. Additionally, there is discussion about the possibility of greatly increasing context size in the future, with the example given that there was an almost 10x jump between GPT-3.5-turbo and GPT-4 in terms of token capacity.&quot;
ftxbroabout 2 years ago
The compressed one is shorter number of characters but does it have fewer tokens? <a href="https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer" rel="nofollow">https:&#x2F;&#x2F;platform.openai.com&#x2F;tokenizer</a>
getpostabout 2 years ago
The context length includes all of GPT4&#x27;s replies, is that correct? If I want to minimize the context length, should I be asking GPT4 to limit its replies to just the fix that is requested?<p>GPT4 seems to prefer long-winded replies, i.e., when I ask for what amounts to a one-line fix, it repeats the entire block of code with the one line correction. In contrast, Replit&#x27;s Ghostwriter often gives a concise reply showing only the one-line fix, and I have to ask for the entire block when it isn&#x27;t clear where the fix is applied.
milesevaabout 2 years ago
This post from @gfodor is also a really interesting read on the topic of compressing prompts, not sure what the main use case would be though!<p><a href="https:&#x2F;&#x2F;twitter.com&#x2F;gfodor&#x2F;status&#x2F;1643444605332099072?s=20" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;gfodor&#x2F;status&#x2F;1643444605332099072?s=20</a>
dr_kiszonkaabout 2 years ago
ZipGPT is a novel, LLM-based non-deterministic compression algorithm. Give our beta version a try. The results may surprise you!
catchnear4321about 2 years ago
This is the most disappointing part about the frenzy.<p>Obvious ideas being held up as something more than that.<p>If you think this is incredible, give yourself some time to go play with LLMs. Try things. Clever things. Stupid things.<p>It helps set a more reasonable scale for these sensational sounding snippets.