TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Machine Learning: Models with Learned Parameters

314 点作者 madisonmay超过 8 年前

6 条评论

antirez超过 8 年前
I strongly advise everybody with one day free (and not much better things to do) to implement a basic fully connected feedforward neural network (the classical stuff, basically), and give it a try against the MNIST handwritten digits database. It&#x27;s a relatively simple project that learns you the basic. On top of that to understand the more complex stuff becomes more approachable. To me this is the parallel task to implement a basic interpreter in order to understand how higher level languages and compilers work. You don&#x27;t need to write compilers normally, as you don&#x27;t need to write your own AI stack, but it&#x27;s the only path to fully understand the basics.<p>You&#x27;ll see it learning to recognize the digits, you can print the digits that it misses and you&#x27;ll see they are actually harder even for humans sometimes, or sometimes you&#x27;ll see why it can&#x27;t understand the digit while it&#x27;s trivial for you (for instance it&#x27;s an 8 but just the lower circle is so small).<p>Also back propagation is an algorithm which is simple to develop an intuition about. Even if you forget the details N years later the idea is one of the stuff you&#x27;ll never forget.
评论 #12558123 未加载
评论 #12558774 未加载
评论 #12558059 未加载
评论 #12558080 未加载
nkozyra超过 8 年前
This is well-written and I applaud any step toward demystifying the sometimes scary sounding concepts that drive much of the ML algorithms.<p>Knowing you can pretty quickly whip up a KNN or ANN in a few hundred lines of code or fewer is one of the more eye-opening parts of the delving in. For the most part, supervised learning follows a pretty reliable path and each algorithm obviously varies in approach, but I know I originally thought &quot;deep learning? ugh, sounds abstract and complicated&quot; before realizing it was all just a deep ANN.<p>Long story short: dig in. It&#x27;s unlikely to be as complex as you think. And if you&#x27;ve ever had an algorithms class (or worked as a professional software dev) none of it should be too daunting. Your only problem will be keeping up the charade if people around you think ML&#x2F;AI is some sort of magic.
评论 #12557919 未加载
评论 #12557896 未加载
djkust超过 8 年前
Hi folks, authors here in case you have questions.<p>This is actually part 3 in a series. For developers who are still getting oriented around machine learning, you might enjoy the first two articles, too. Part 1 shows how the machine learning process is fundamentally the same as the scientific thinking process. Part 2 explains why MNIST is a good benchmark task. Future parts will show how to extend the simple model into the more sophisticated stuff we see in research papers.<p>We intend to continue until as long as there are useful things to show &amp; tell. If there are particular topics you&#x27;d like to see sooner than later, please leave a note!
评论 #12558594 未加载
yodsanklai超过 8 年前
I took Andrew NG&#x27;s ML class on Coursera. It was certainly interesting to see how ML works but I&#x27;m not sure what to do with this. Particularly, I&#x27;m still unsure how to tell beforehand if a problem is too complex to be considered, how much data it&#x27;ll require, what computing power is needed.<p>Are there a lot of problems that fall between the very hard and the very easy ones? and for which enough data can be found?
评论 #12559748 未加载
throwaway13048u超过 8 年前
So this may be a place as good as any -- I&#x27;ve got a decent math background, and am self teaching myself ML while waiting for work to come in.<p>I&#x27;m working on undertstanding CNNs, and I can&#x27;t seem to find the answer (read: don&#x27;t know what terms to look for) that explain how you train the convolutional weights.<p>For instance, a blur might be<p>[[ 0 0.125 0 ] , [ 0.125 0.5 0.125 ] , [0 0.125 0]]<p>But in practice, I assume you would want to have these actual weights themselves trained, no?<p>But, in CNNs, the same convolutional step is executed on the entire input to the convolutional step, you just move around where you take your &quot;inputs&quot;.<p>How do you do the training, then? Do you just do backprop on each variable of the convolution stem from its output, with a really small learning rate, then repeat after shifting over to the next output?<p>Sorry if this seems like a poorly thought out question, I&#x27;m definitely not phrasing this perfectly.
评论 #12559972 未加载
评论 #12558768 未加载
评论 #12558540 未加载
aantix超过 8 年前
There&#x27;s been a couple of times where I needed to classify a large set of web pages and used a Bayes classifier.<p>I would start to get misclassified pages and it was so difficult to diagnose as to why these misclassifications were occurring. Bad examples? Bad counter examples? Wrong algorithm for the job? Ugh.<p>I ended up writing a set of rules. It wasn&#x27;t fancy but at the end of the day, I understood the exact criteria for each classification and they were easily adjustable.
评论 #12558426 未加载