TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Are machine-learning packages like genomes?

1 点作者 Ecalpal超过 5 年前
Biologist here.<p>I&#x27;m trying to learn a bit about machine learning just to keep my general knowledge fresh.<p>Reading about machine-learning model&#x2F;package development, I&#x27;ve realised that it must be easy to allow such models to degenerate; to patch and glue and create hidden dependencies that can be hard to reverse-engineer.<p>This reminds me very much of a genome, where functionality has been added over billions of years using whatever inputs were available at a time and producing something good enough at the time.<p>I&#x27;m not sure how relevant such analogies are to ML, but it feels like this must be the natural way of things: The code wants to degenerate (path of least resistance), but for the model to be clear--and generalisable--this must be resisted.<p>Do you feel this is fair&#x2F;accurate?<p>Again, I&#x27;m a biologist, not a technical expert. I just found this similarity intriguing and it would be interesting to hear your thoughts on the challenges&#x2F;opportunities of allowing code to be more like code (and less like a genome, since they&#x27;re notoriously hard to understand or reverse-engineer).

1 comment

madhadron超过 5 年前
When applied to the models themselves it&#x27;s not a particularly helpful analogy. Machine learning models are largely curve fitting in high dimensional spaces. Overfitting, overspecialization, and the like are problems, and you could relate them to ecological notions and selection, but it&#x27;s not terribly helpful in practice.<p>Where it <i>does</i> apply is to the cascade of dependencies among data sets, code that generates them, and sources of signal that you see in large data platforms.
评论 #22107268 未加载