TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Self-Supervised Learning from Images with JEPA (2023)

40 点作者 Brysonbw大约 2 个月前

5 条评论

byyoung3大约 2 个月前
It’s not new and only superior in a very narrow set of categories.
评论 #43513769 未加载
blixt大约 2 个月前
Needs a (2023) tag. But definitely the release of ARC2 and image outputs from 4o got me thinking about the JEPA family too.<p>I don&#x27;t know if it&#x27;s right (and I&#x27;m sure JEPA has lots of performance issues) but seems good to have a fully latent space representation, ideally across all modalities, so that the concept &quot;an apple a day keeps the doctor away&quot; becoming image&#x2F;audio&#x2F;text is a choice of decoder rather than dedicated token ranges being chosen even before the actual creation process in the model begins.
niemandhier大约 2 个月前
GPTs are in the “exploit” phase of the “explore-exploit” trade-off.<p>JEPA is still in the explore phase, it’s good to read the paper and have an understanding of the architecture to gain an alternative perspective.
laughingcurve大约 2 个月前
Not new, not notable right now, not sure why it&#x27;s getting upvoted (just kidding, it&#x27;s because people see YLC and upvote based on names)
评论 #43513575 未加载
评论 #43515581 未加载
justanotheratom大约 2 个月前
JEPA is presumably superior to Transformers. Can any expert enlighten us on the implications of this paper?
评论 #43516306 未加载