TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

An Idiot’s guide to Support vector machines (2003) [pdf]

312 点作者 bladecatcher超过 6 年前

10 条评论

quantombone超过 6 年前
Back in around 2008, SVMs were all the rage in computer vision. We would use hand designed visual features and then a linear SVM on top. That was how object detectors were built (remember DPM?)<p>Funny how SVMs are just max-margin loss functions and we just took for granted that you needed domain expertise to craft features like HOG&#x2F;SIFT by hand.<p>By 2018, we use ConvNets to learn BOTH the features and the classifier. In fact, it’s hard to separate where the features end and the classifier begins (in a modern CNN).
评论 #18795572 未加载
评论 #18794977 未加载
评论 #18795232 未加载
评论 #18795586 未加载
评论 #18794828 未加载
评论 #18795778 未加载
评论 #18795378 未加载
abhgh超过 6 年前
If you need closer to a ELI5 version I recommend this - [1].<p>Disclaimer: written by me.<p>[1] <a href="https:&#x2F;&#x2F;blog.statsbot.co&#x2F;support-vector-machines-tutorial-c1618e635e93" rel="nofollow">https:&#x2F;&#x2F;blog.statsbot.co&#x2F;support-vector-machines-tutorial-c1...</a>
评论 #18797293 未加载
评论 #18796833 未加载
cultus超过 6 年前
I notice this doesn&#x27;t mention hinge loss, which is by far the simpler way of arriving at the SVM. Hinge loss is just max(0, 1- t*y), where y is the output of the linear model and t = +-1 is the label. Thus, it takes the common-sense approach of not penalizing losses that are far enough away from the decision boundary, and penalizing linearly after that.<p>An SVM is literally just a linear model with hinge loss instead of log loss (logistic regression) or squared loss (ordinary linear regression) in primal form. For apparently historical reasons, it is usually derived from the &quot;hard-margin&quot; SVM in dual form, motivating with trying to maximize the margin. This is complicated and not very intuitive.<p>This also causes people to conflate the kernel trick and the dual form, while in fact they have nothing to do with each other. You can use the kernel trick in primal svm just fine.<p>Stochastic gradient descent can also be used for primal methods, while it doesn&#x27;t work in the dual. That makes it much faster for large problems than the dual.
评论 #18796161 未加载
评论 #18799017 未加载
cultus超过 6 年前
There&#x27;s been some work on variational Bayesian formulations of SVMs in the last few years. These can give actual uncertainty estimates and do automatic hyperparameter tuning. This one in particular is very cool:<p><a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1707.05532.pdf" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1707.05532.pdf</a>
usgroup超过 6 年前
Clearly idiots are not what they use to be in my day ...
评论 #18795440 未加载
评论 #18799543 未加载
rusbus超过 6 年前
It&#x27;s interesting how quickly support vector machines went from the hot new thing to classify images to an afterthought after deep learning started having great results.
评论 #18794722 未加载
评论 #18798992 未加载
评论 #18795491 未加载
simonw超过 6 年前
Bullet point on page 2: &quot;Optimal hyperplane for linearly separable patterns&quot;<p>I think the author may be working from a very different definition of the word &quot;idiot&quot;.
评论 #18796251 未加载
iamwil超过 6 年前
I have a question!<p>In the pdf, it said that the optimization problem in SVMs have a nice property in that it was quadratic, which means that there&#x27;s a nice global minimum to go towards, and not lots of local minimum like in NN. That means, it seems SVMs won&#x27;t get stuck at a suboptimal solution.<p>Is that not a problem in DNNs now? Or is it that it&#x27;s such high dimensionality that local minima don&#x27;t stop the optimizer, because there&#x27;s always another way around the local minimum?
评论 #18797055 未加载
评论 #18799242 未加载
评论 #18797084 未加载
评论 #18797023 未加载
评论 #18797546 未加载
评论 #18798563 未加载
strikelaserclaw超过 6 年前
implementing a svm was my senior project in college. Brings back nightmares.
mistrial9超过 6 年前
compare to Tzotsos 2006 &quot;A SUPPORT VECTOR MACHINE APPROACH FOR OBJECT BASED IMAGE ANALYSIS&quot;