TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: Machine learning cheat sheet

166 pointsby Emoreabout 14 years ago

7 comments

teuobkabout 14 years ago
In case you see the cheat sheet and think, "Wow, I'd love to understand that," there's an excellent (albeit challenging) complete course on machine learning in Stanford's "engineering everywhere" online repository. <a href="http://see.stanford.edu/see/courseinfo.aspx?coll=348ca38a-3a6d-4052-937d-cb017338d7b1" rel="nofollow">http://see.stanford.edu/see/courseinfo.aspx?coll=348ca38a-3a...</a>
评论 #2516075 未加载
评论 #2516089 未加载
iskanderabout 14 years ago
All the algorithms requiring training can be optimized using stochastic gradient descent-- which is very effective for large data sets (see <a href="http://leon.bottou.org/research/stochastic" rel="nofollow">http://leon.bottou.org/research/stochastic</a>)<p>Also, here are some additions for the online learning column:<p>* Online SVM: <a href="http://www.springerlink.com/index/Y8666K76P6R5L467.pdf" rel="nofollow">http://www.springerlink.com/index/Y8666K76P6R5L467.pdf</a><p>* Online gaussian mixture estimation: <a href="http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.87.1698&#38;rep=rep1&#38;type=pdf" rel="nofollow">http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.87....</a><p>One more thing: why no random forests? Or decision tree ensembles of any sort?
评论 #2516772 未加载
imurrayabout 14 years ago
KNN <i>"no learning involved"</i>: one probaby wants to cross-validate K at the least, if not learn the metric.<p>Some methods say online learning isn't applicable. As pointed out elsewhere, objectives for K-means and mixture models could be fitted with stochastic gradient descent. In general there is always an online option. For example, keep a restricted set of items and chuck out ones that seem less useful as others come in.<p>(Aside: I have a <i>very</i> introductory lecture to machine learning on the web: <a href="http://videolectures.net/bootcamp2010_murray_iml/" rel="nofollow">http://videolectures.net/bootcamp2010_murray_iml/</a> — not for anyone that knows the methods on this cheat sheat!)
评论 #2516771 未加载
cloudkjabout 14 years ago
Nice summary; I like the format as well. However, the title of the cheat sheet is misleading since (a) many of the algorithms listed can be used for non-linear classification and (b) some of them can be considered supervised learning, such as naive Bayes and perceptron since they're trained with sample inputs and expected outputs (supervisory signals).<p>Otherwise, this is awesome. Hopefully you will add to it, and make it available in web form.
评论 #2515800 未加载
sesabout 14 years ago
Fantastic work, I have an ML exam coming up and this should really help. If I'm honest its one of the subjects I've struggled with the most. It seems experts in the field while incredibly intelligent, have a hard time breaking the material down into structured and easily digestible pieces of information.
MatthewBabout 14 years ago
No idea what i'm looking at but it definitely looks cool.
axxlabout 14 years ago
I'm taking this class next semester, downloaded it so hopefully I'll understand it later and it will come in use. Thanks!