TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Swish: A Self-Gated Activation Function

2 点作者 goberoi超过 7 年前

1 comment

goberoi超过 7 年前
Why is this interesting? In short: a great new activation function that may challenge the dominance of ReLU.<p>Longer story:<p>Today, ReLU is the most popular activation function for deep networks (along with its variants like leaky ReLU or parametric ReLU).<p>This paper from the Google Brain team is ~2 weeks old, and shows that SWISH, a new activation function, &quot;improves top-1 classification accuracy on ImageNet by 0.9% for Mobile NASNetA and 0.6% for Inception-ResNet-v2&quot; by simply replacing ReLU with SWISH.<p>SWISH is equal to x * sigmoid(x), so not that much harder to compute either.