TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

JPEG-LM: LLMs as Image Generators with Canonical Codec Representations

5 点作者 hardmaru9 个月前

1 comment

magicalhippo9 个月前
Initially I was assuming they were not including the Huffman encoding step, but no:<p><i>The bytes in the files do not have consistent meanings and would depend on their context and the implicit Huffman tables. [...]</i><p><i>However, we observe that conventional, vanilla language modeling surprisingly conquers these challenges without special designs as training goes (e.g., JPEG-LM generates realistic images barely with any corrupted JPEG patches).</i><p>That surprised me, but then I&#x27;m not in the field.