TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Level-Up Your Machine Learning

336 pointsby cjrdalmost 11 years ago

16 comments

xiaomaalmost 11 years ago
I think this is fantastic advice.<p>As someone who has spent an embarrassing amount of time on various independent education, one of the key things I&#x27;ve taken from it is just how efficient text books are. Not only can you read more quickly than people can speak, but it&#x27;s also active by nature. I&#x27;ve often found my attention wandering during videos, but it&#x27;s just not possible to read without putting in a minimum amount of focus. It&#x27;s also a lot easier to modulate your reading speed based on how easy material is for you than it is to do the same during a lecture video.<p>Some general thoughts on MOOCs:<p>Coursera and edX tend to be great for small, self-contained topics and the automated graders for programming assignments is great as well. The forums are also useful, though not ideal (since there are no practice questions students can get help with that don&#x27;t fall under the honor code).<p>Where modern MOOCs really fall down is prerequisites. It&#x27;s surprisingly difficult to do something like structure an entire CS degree from Coursera classes. Though many classes are taught by famous CS professors, they are from different institutions that break material into courses in different ways. Worse still a lot of the classes are either watered-down or shortened or both.<p>MIT&#x27;s Open Courseware archives are actually a lot better for this. There are no certificates, and no credentials, but nearly all the material is freely available. The one biggest inefficiency though, is all the time spent in the lectures. At least they can be played back at a higher speed, but the lectures really do take a lot more time and cover less than the textbooks. For courses that have good textbooks, I think the best approach is to skip the lectures except in portions where you feel like you need more review.<p>Finally Khan Academy is fantastic for answering specific, mechanical questions (e.g. how to calculate eigen values), but a bit light on material. I&#x27;d use it as a supplement for the other resources.
评论 #8062735 未加载
评论 #8062081 未加载
评论 #8063109 未加载
rfreyalmost 11 years ago
I love textbooks and spend more of my childrens&#x27; inheritance on them than I should.<p>But what MOOCs give me is the <i>exercises</i>. I often think I understand a problem, but it&#x27;s only after getting 2.1&#x2F;10 on a Coursera quiz that I realize I&#x27;ve missed a key step or concept.<p>Many textbooks have exercises but few have solutions. I&#x27;ve been working through Barto and Sutton&#x27;s Reinforcement Learning for example (again), and although I do the exercises and programming questions, I never know if I&#x27;ve gotten it <i>right</i>. My experience with MOOCs shows I probably haven&#x27;t in a large number of cases.<p>The best of both worlds is when I can follow a MOOC with the textbook to gain more depth, for example with the PGM course on Coursera.
评论 #8062228 未加载
评论 #8062467 未加载
评论 #8062325 未加载
grayclhnalmost 11 years ago
Two free books that I haven&#x27;t seen mentioned, that are from more of a stats perspective<p>* James, Witten, Hastie, and Tibshirani&#x27;s <i>An Introduction to Statistical Learning, with Applications in R</i><p><a href="http://www-bcf.usc.edu/~gareth/ISL/" rel="nofollow">http:&#x2F;&#x2F;www-bcf.usc.edu&#x2F;~gareth&#x2F;ISL&#x2F;</a><p>* Hastie, Tibshirani, and Freedman&#x27;s <i>Elements of statistical learning</i> (more advanced)<p><a href="http://statweb.stanford.edu/~tibs/ElemStatLearn/" rel="nofollow">http:&#x2F;&#x2F;statweb.stanford.edu&#x2F;~tibs&#x2F;ElemStatLearn&#x2F;</a>
scottlocklinalmost 11 years ago
PGM is a tough book. I&#x27;m not sure it&#x27;s the right book for &quot;level 3&quot; unless you want to be a level-3 who is good at PGMs.<p>The problem with ML is there are so many different kinds. Bishop&#x27;s book is a decent light weight survey, but it doesn&#x27;t come close to covering all the interesting fields. You could read that and Hastie&#x2F;Tibshirani&#x27;s book and still know almost nothing about online training (hugely important for &quot;big data&quot; and timeseries), reinforcement learning (mentioned, but not in any depth), agent learning, &quot;compression&quot; sequence predicting techniques, time series oriented techniques (recurrent ANNs for starters, but there is a ton to know here, and most interesting data is time ordered), image recognition tools, conformal prediction, speech recognition tools, ML in the presence of lots of noise, and unsupervised learning. I don&#x27;t own PGM, but it probably wouldn&#x27;t help much in these matters either. I know guys who are probably level 4 at machine learning who don&#x27;t know about most of these subjects. On the other hand, Peter Flach&#x27;s book &quot;Machine Learning&quot; at least mentions them and makes pointers to other resources.<p>&quot;Deep learning&quot; is becoming kind of a buzzword for a big basket of tricks. I think it&#x27;s worth knowing about drop-out training, and the tricks used to do semi-supervised learning, but the buzzword is silly. Technically &quot;deep learning&quot; just means &quot;improved gradient descent.&quot; I figure level-4 is anyone making progress coming up with new techniques.<p>That said, reading good books is one way to make progress. Knowing the right people is the other way.
评论 #8071766 未加载
telalmost 11 years ago
PRML is great. I haven&#x27;t read PGM, but I took a relatively intensive course on it which had great lecture notes. Which I&#x27;d like to also suggest—lecture notes are often &quot;skeletal books&quot; which can bring you up to speed on a topic quickly given that you (a) are willing to work a bit more and (b) can fill in the missing fleshy bits with your own experience.<p>I&#x27;d also really like to suggest DGL (<a href="http://books.google.com/books/about/A_Probabilistic_Theory_of_Pattern_Recogn.html?id=5uCTngEACAAJ" rel="nofollow">http:&#x2F;&#x2F;books.google.com&#x2F;books&#x2F;about&#x2F;A_Probabilistic_Theory_o...</a>) and Bickel and Doksum (<a href="http://www.amazon.com/Mathematical-Statistics-Basic-Selected-Topics/dp/0132306379" rel="nofollow">http:&#x2F;&#x2F;www.amazon.com&#x2F;Mathematical-Statistics-Basic-Selected...</a>). These are two of my <i>favorite</i> core ML&#x2F;stats books.
vkhucalmost 11 years ago
There are some (free) good books that haven&#x27;t been mentioned yet:<p>1) &quot;Data Mining and Analysis: Fundamental Concepts and Algorithms&quot; by Zaki and Meira <a href="http://www.cs.rpi.edu/~zaki/PaperDir/DMABOOK.pdf" rel="nofollow">http:&#x2F;&#x2F;www.cs.rpi.edu&#x2F;~zaki&#x2F;PaperDir&#x2F;DMABOOK.pdf</a><p>This book covers many ML topics with concrete examples.<p>2) &quot;Computer Vision: Models, Learning, and Inference&quot; by Simon Prince: <a href="http://web4.cs.ucl.ac.uk/staff/s.prince/book/book.pdf" rel="nofollow">http:&#x2F;&#x2F;web4.cs.ucl.ac.uk&#x2F;staff&#x2F;s.prince&#x2F;book&#x2F;book.pdf</a><p>Despite a CV book, the first half of it is like a statistics book that comes with examples in CV which are very easy to follow.
dimaturaalmost 11 years ago
I would also suggest K. Murphy&#x27;s Machine Learning for the journeyman level. In the intermediate apprentice-journeyman level Alpaydin&#x27;s Introduction to Machine Learning is very friendly.
评论 #8062466 未加载
kashifralmost 11 years ago
Form my own journey I would say that a good place to start for graphical models might be &quot;Bayesian Reasoning and Machine Learning&quot; by Barber. It&#x27;s free (<a href="http://web4.cs.ucl.ac.uk/staff/D.Barber/pmwiki/pmwiki.php?n=Brml.Online" rel="nofollow">http:&#x2F;&#x2F;web4.cs.ucl.ac.uk&#x2F;staff&#x2F;D.Barber&#x2F;pmwiki&#x2F;pmwiki.php?n=...</a>). I haven&#x27;t read through it, but I&#x27;ve heard good things. However, it doesn&#x27;t cover some basic things like SVM, RVM, Neural Networks...<p>For those I&#x27;d suggest &quot;Pattern Recognition and Machine Learning&quot; by Bishop. I&#x27;ve read throughout this and it&#x27;s really well organized and thought out. For more mathematically advanced ML stuff I&#x27;d suggest &quot;Foundations of Machine Learning&quot; by Mohri. For a good reference for anything else I&#x27;d suggest &quot;Machine Learning: A Probabilistic Perspective&quot; by Murphy. For more depth on graphical models look at &quot;Probabilistic Graphical Models: Principles and Techniques&quot; by Koller.<p>On the NLP front there&#x27;s the standard texts &quot;Speech and Language Processing&quot; by Jurafsky and &quot;Foundations of Statistical Natural Language Processing&quot; by Manning.<p>I also like &quot;An Introduction to Statistical Learning&quot; by James, Witten, Hastie and Tibshirani.
评论 #8066812 未加载
cipher0almost 11 years ago
Great recommendations, some people might also find this interesting as a general guideline to &quot;Data science&quot; <a href="http://nirvacana.com/thoughts/becoming-a-data-scientist/" rel="nofollow">http:&#x2F;&#x2F;nirvacana.com&#x2F;thoughts&#x2F;becoming-a-data-scientist&#x2F;</a><p>[edit] scroll down and look at the map.
eli_gottliebalmost 11 years ago
What I&#x27;d really appreciate is ideas on how to learn or review the core math concepts. I haven&#x27;t actually <i>done</i> any multivariable calculus, vector&#x2F;matrix calculus, or linear algebra in <i>years</i>, even though I took them in undergrad.
joaomsaalmost 11 years ago
Wholeheartily agree with the author&#x27;s sentiments on the value of textbooks. Not because of the medium itself, but because they&#x27;re (usually) accompanied by well thoughtout examples and practice problems.<p>When initially starting a dense subject such as PGM, having my hands held through the introductory material with incremental practice problems as the topic elaborated, helped me get a much more intimate grasp. Initially only reading superficially and watching lectures, I kept getting stumped trying to form a cohesive mental map of all the interleaved concepts.
GabrielF00almost 11 years ago
What are the HN community&#x27;s thoughts on Learning from Data by Abu Mostafa, Magdon-Ismail and Lin (<a href="http://amlbook.com/" rel="nofollow">http:&#x2F;&#x2F;amlbook.com&#x2F;</a>)? The lectures from their course are here: <a href="http://work.caltech.edu/lectures.html" rel="nofollow">http:&#x2F;&#x2F;work.caltech.edu&#x2F;lectures.html</a><p>I haven&#x27;t started it yet, but this book was recommended by some folks at my company.
评论 #8061831 未加载
lowglowalmost 11 years ago
I&#x27;m in the middle of PGM right now. It&#x27;s actually really easy to follow if you put some time into it. I&#x27;m reading PRML next. I didn&#x27;t realize there was a &#x27;path&#x27; to learning ML though, thanks for that.<p>We could use some more ML recs on <a href="https://books.techendo.com/" rel="nofollow">https:&#x2F;&#x2F;books.techendo.com&#x2F;</a>
kp25almost 11 years ago
I would like to start my ML Journey in Python, then get to R.<p>How about learning things in python? Good recommendations?
评论 #8062807 未加载
评论 #8062971 未加载
jpetersonalmost 11 years ago
For a really nice introductory book, try &quot;Machine Learning&quot; by Tom Mitchell.
orasisalmost 11 years ago
Textbooks? Really?<p>How about start with a great lecturer like -<p>Nando de Freitas - <a href="https://www.youtube.com/channel/UC0z_jCi0XWqI8awUuQRFnyw" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;channel&#x2F;UC0z_jCi0XWqI8awUuQRFnyw</a><p>David Mackay -<a href="http://videolectures.net/course_information_theory_pattern_recognition/" rel="nofollow">http:&#x2F;&#x2F;videolectures.net&#x2F;course_information_theory_pattern_r...</a><p>or the (sometimes too dense) Andrew Ng - <a href="https://www.coursera.org/course/ml" rel="nofollow">https:&#x2F;&#x2F;www.coursera.org&#x2F;course&#x2F;ml</a>
评论 #8061823 未加载
评论 #8062152 未加载
评论 #8062101 未加载