TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Exploring APIs of Knet.jl and Flux.jl in Comparison to Tensorflow-Keras

6 点作者 randyzwitch将近 6 年前

1 comment

ChrisRackauckas将近 6 年前
Wow, the training time difference is much larger than I thought it would be. The main reason I use Flux.jl is because of its flexibility to just throw existing Julia libraries into it (DiffEqFlux.jl for neural ODEs, neural SDEs, neural PDEs, neural jump diffusions, etc. took surprisingly little work). However, I assumed that the CPU kernels would all be relatively comparable between the different neural network frameworks. This is a quite a compelling example that, at least for small neural networks and little data, the overhead can be quite large (quite large as in 10x!).<p>As the size of the neural networks grow this will fade away due to spending more time in the BLAS kernels. However, for non-standard applications which don&#x27;t spend their time mostly in a neural network (like many neural differential equations according to our profiling), this difference would make or break an application.