TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Gated Linear Networks

150 点作者 asparagui将近 5 年前

4 条评论

fxtentacle将近 5 年前
That is an amazing paper, a great result and new neutral architectures are long overdue.<p>But I don&#x27;t believe that this has any significance in practice.<p>GPU memory is the limiting factor for most current AI approaches. And that&#x27;s where the typical convolutional architectures shine, because they effectively compress the input data, then work on the compressed representation, then decompress the results. With gated linear networks, I&#x27;m required to always work on the full input data, because it&#x27;s a one step prediction. As the result, I&#x27;ll run out of GPU memory before I reach a learning capacity that is comparable to conv nets.
评论 #23531429 未加载
评论 #23529663 未加载
评论 #23535339 未加载
Immortal333将近 5 年前
&quot;We show that this architecture gives rise to universal learning capabilities in the limit, with effective model capacity increasing as a function of network size in a manner comparable with deep ReLU networks.&quot;<p>What exactly this statement means?
评论 #23530929 未加载
评论 #23529528 未加载
评论 #23529492 未加载
评论 #23529505 未加载
nl将近 5 年前
I didn&#x27;t realise Hutter was on leave from ANU at DeepMind.
caretak3r将近 5 年前
As a relative neophyte in this realm, this is fascniating to read. Comparing this to the the models&#x2F;methods to derive said properties, is good education for me.