TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Bayesian Neural Networks

259 点作者 reqo6 个月前

7 条评论

dccsillag6 个月前
Bayesian Neural Networks just seem like a failed approach, unfortunately. For one, Bayesian inference and UQ fundamentally depends on the choice of the prior, but this is rarely discussed in the Bayesian NN literature and practice, and is further compounded by how fundamentally hard to interpret and choose these priors are (what is the intuition behind a NN&#x27;s parameters?). Add to that the fact that the Bayesian inference is very much approximate, and you should see the trouble.<p>If you want UQ, &#x27;frequentist nonparametric&#x27; approaches like Conformal Prediction and Calibration&#x2F;Multi-Calibration methods seem to work quite well (especilly when combined with the standard ML machinery of taking a log-likelihood as your loss), and do not suffer from any of the issues above while also giving you formal guarantees of correctness. They are a strict improvement over Bayesian NNs, IMO.
评论 #42210714 未加载
评论 #42210454 未加载
评论 #42211431 未加载
评论 #42213030 未加载
评论 #42210339 未加载
评论 #42211539 未加载
评论 #42211553 未加载
duvenaud6 个月前
Author here! What a surprise. This was an abandoned project from 2019, that we never linked or advertised anywhere as far as I know. Anyways, happy to answer questions.
评论 #42211309 未加载
评论 #42211237 未加载
评论 #42214691 未加载
评论 #42210924 未加载
levocardia6 个月前
What frustrates me about Bayesian NNs is that talking about &quot;priors&quot; doesn&#x27;t make nearly as much sense as it does in a regression context. A prior over parameter weights has no interpretation in the way that a prior over a regression coefficient, or even a spline smoothness, does. What you really want -- and what natural intelligence probably has -- are priors over <i>aspects of the world</i>.<p>Francois Chollet&#x27;s paper on measuring intelligence was really informative for me on this front; the &quot;priors&quot; you should have about the world are not half-cauchys over certain hyperparameters or whatever, but priors about agent-ness, object-ness, goal-oriented-ness, and so on. How to encode that in a network...well, that&#x27;s the real trick, right?
评论 #42217708 未加载
datastoat6 个月前
I like Bayesian inference for few-parameter models where I have solid grounds for choosing my priors. For neural networks, I like to ask people &quot;what&#x27;s your prior for ReLU versus LeakyReLU versus sigmoid?&quot; and I&#x27;ve never gotten a convincing answer.
评论 #42213967 未加载
评论 #42210422 未加载
评论 #42210016 未加载
评论 #42209708 未加载
oli56796 个月前
<a href="https:&#x2F;&#x2F;publications.aston.ac.uk&#x2F;id&#x2F;eprint&#x2F;373&#x2F;1&#x2F;NCRG_94_004.pdf" rel="nofollow">https:&#x2F;&#x2F;publications.aston.ac.uk&#x2F;id&#x2F;eprint&#x2F;373&#x2F;1&#x2F;NCRG_94_004...</a><p>mixture density networks are quite interesting if you want probabilistic estimates of neural. here, your model learns to output and array of gaussian distribution coefficient distributions, and mixture weights.<p>these weights are specific to individual observations, and trained to maximise likelihood.
评论 #42217660 未加载
ok1234566 个月前
BNNs were an attractive choice in scenarios where the data is expensive to collect, like actual physical experiments. But boosting and other tree-based regression methods give you similar performance with a more straightforward framework for limited tabular data.
sideshowb6 个月前
I like Bayes, but I thought the &quot;surprising&quot; result is that double descent is supposed to prevent nns from overfitting?
评论 #42210492 未加载