TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Introduction to Support Vector Machines in Machine Learning

166 pointsby rangerranvirabout 5 years ago

7 comments

rusty-rustabout 5 years ago
Large parts of this blog are straight copy-paste from “An introduction to statistical learning” by Gareth James et. al.
astrophysicianabout 5 years ago
If you&#x27;re new to ML or datascience, I would recommend working to build a strong basis in Bayesian statistics. It will help you understand how all of the &quot;canonical&quot; ML methods relate to one another, and will give you a basis for building off of them.<p>In particular, aspire to learn probabilistic graphical models + the libraries to train them (like pyro, tensorflow probability, Edward, Stan). They have a steep learning curve, especially if you&#x27;re new to the game, but the reward is great.<p>All of these methods have their place. SVM&#x27;s have their place, but also aren&#x27;t great for probability calibration and non-linear SVM&#x27;s like every single kernel method can scale absolutely terribly. Neural networks have their place, sometimes as a component of a larger statistical model, sometimes as a feature selector, sometimes in and of themselves. They&#x27;re also very often the wrong choice for a problem.<p>Don&#x27;t fall into the beginner trap: sometimes people tend to mistake &#x27;what is the hottest research topic&#x27; for &#x27;what is the right solution to my problem given my constraints, (data limitations, time limitations, skill limitations, etc.)&#x27;. Be realistic, don&#x27;t use magical thinking, and have a strong basis in statistics to weed out the beautiful non-bullshit from the bullshit that is frustratingly prevalent (everyone and their mother is an ML expert today).<p>EDIT: I want to also clarify: I don&#x27;t mean to suggest the author is new to ML, I just mean this as general advice for anyone coming here who is new to DS&#x2F;ML. The article looks great!
评论 #23037687 未加载
评论 #23037138 未加载
评论 #23037356 未加载
评论 #23036686 未加载
评论 #23036613 未加载
评论 #23037009 未加载
评论 #23037329 未加载
评论 #23040099 未加载
评论 #23036592 未加载
评论 #23039690 未加载
评论 #23037630 未加载
smbrianabout 5 years ago
Stay away, in my opinion. I spent a year supporting a SVM in a production machine learning application, and it made me wish the ML research community hadn&#x27;t been so in love with them for so long.<p>They&#x27;re the perfect blend of theoretically elegant and practically impractical. Training scales as O(n^3), serialized models are heavyweight, prediction is slow. They&#x27;re like Gaussian Processes, except warped and without any principled way of choosing the kernel function. Applying them to structured data (mix of categorical &amp; continuous features, missing values) is difficult. The hyperparameters are non-intuitive and tuning them is a black art.<p>GBMs&#x2F;Random Forests are a better default choice, and far more performant. Even simpler than that, linear models &amp; generalized linear models are my go-to most of the time. And if you genuinely need the extra predictiveness, deep learning seems like better bang for your buck right now. Fast.ai is a good resource if that&#x27;s interesting to you.
评论 #23036846 未加载
评论 #23036066 未加载
评论 #23037965 未加载
评论 #23037010 未加载
bitforgerabout 5 years ago
ITT: Whether SVMs are still relevant in the deep learning era. Some junior researchers will say neural networks are all you need. Industry folks will talk about how they still use decision trees.<p>Personally, I&#x27;m quite bullish on the resurgence of SVMs as SOTA. What did it for me was Mikhail Belkin&#x27;s talk at IAS.[1]<p>[1] <a href="https:&#x2F;&#x2F;m.youtube.com&#x2F;watch?index=15&amp;list=PLdDZb3TwJPZ5dqqg_S-rgJqSFeH4DQqFQ&amp;v=5-Kqb80h9rk" rel="nofollow">https:&#x2F;&#x2F;m.youtube.com&#x2F;watch?index=15&amp;list=PLdDZb3TwJPZ5dqqg_...</a>
评论 #23036298 未加载
评论 #23038514 未加载
评论 #23036121 未加载
starchild_3001about 5 years ago
I&#x27;ve been an ML practioner since 2009. I&#x27;ve used every method imaginable or popular, I think. With the exception of non-linear SVMs. Linear SVM =&gt; All good, just the hingle loss optimization. Non-linear SVM, a bit of overkill with basis expansion. Just too slow, or too complex a model?<p>My impression: SVMs are more of theoretical interest than practical interest. Yeah, learn your statistics. Loss functions. Additive models. Neural nets. Linear models. Decision trees, kNNs etc. SVM is more of a special interest, imho.
评论 #23039219 未加载
zetazzedabout 5 years ago
Interestingly, a top Kaggler (Ahmet) just posted a nice contest solution with SVMs for the TReNDS Neuroimaging contest: <a href="https:&#x2F;&#x2F;www.kaggle.com&#x2F;aerdem4&#x2F;rapids-svm-on-trends-neuroimaging" rel="nofollow">https:&#x2F;&#x2F;www.kaggle.com&#x2F;aerdem4&#x2F;rapids-svm-on-trends-neuroima...</a>
评论 #23035932 未加载
评论 #23035961 未加载
rangerranvirabout 5 years ago
Since, I am here. I would like to take a small feedback about the general structure of the website and how it feels.<p>If someone has a suggestion on how I can improve the user experience feel free to hop in and let me know.