TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Illustrated FixMatch for semi-supervised learning

237 pointsby amitnessabout 5 years ago

10 comments

hadsedabout 5 years ago
The cold hard reality of machine learning is that most useful data isn&#x27;t readily available to just be collected. Semi-supervised and weakly supervised learning, data augmentation, multi-task learning, these are the things that will enable machine learning for the majority of companies out there who need to build datasets and potentially leverage domain expertise somehow to bootstrap intelligent features in their apps. This is great work in that direction for computer vision.<p>Even the giants are recognizing this fact and are leveraging it to great effect. Some keywords to search for good papers and projects: Overton, Snorkel, Snorkel Metal
评论 #22770071 未加载
jonponabout 5 years ago
Great summary! Reminds me a lot about Leon Bottou&#x27;s work on using deep learning to learn causal invariant representations. (Video: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=lbZNQt0Q5HA" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=lbZNQt0Q5HA</a>)<p>We can view the augmentations of the image as &quot;interventions&quot; forcing the model to learn an invariant representation of the image.<p>Although the blog post did not frame it as this type of problem (not sure if the paper did), I think it can definitely be seen as such and is really promising.
评论 #22771940 未加载
antipaulabout 5 years ago
I wish all papers were structured this way, by default.<p>That is, plenty of good diagrams, clear explanations and intuitions, no unnecessary mathiness.
评论 #22770669 未加载
评论 #22770770 未加载
评论 #22772047 未加载
manthideaalabout 5 years ago
I wonder if a two step process could work better than this, first use a variational autoencoder or simple an autoencoder then use it to train the labeled sampled.<p>In (1) there is a full example of using the two step strategy but using more labeled data to obtain 92% of accuracy. Someone can try changing the second part to use only ten labels for the classifying part and share results?<p>(1) <a href="https:&#x2F;&#x2F;www.datacamp.com&#x2F;community&#x2F;tutorials&#x2F;autoencoder-classifier-python" rel="nofollow">https:&#x2F;&#x2F;www.datacamp.com&#x2F;community&#x2F;tutorials&#x2F;autoencoder-cla...</a><p>Edited: I found a deep analysis in (2), in short for CIFAR 10 the VAE semi-supervised learning approach provides poor results, but the author has not used augmentation!<p>(2) <a href="http:&#x2F;&#x2F;bjlkeng.github.io&#x2F;posts&#x2F;semi-supervised-learning-with-variational-autoencoders&#x2F;" rel="nofollow">http:&#x2F;&#x2F;bjlkeng.github.io&#x2F;posts&#x2F;semi-supervised-learning-with...</a>
评论 #22775685 未加载
starpilotabout 5 years ago
I wish there was a way to augment data as easily for free text, and other business data. I always see these few-shot learning papers for images, I suspect because it&#x27;s easy to augment image datasets and because image-recognition is interesting to laypeople. The vast majority of data we deal with in business is text&#x2F;numerical which is much harder to use in these approaches.
评论 #22770897 未加载
评论 #22774778 未加载
fermienricoabout 5 years ago
I don&#x27;t know much about ML&#x2F;Deep-Learning and I have a burning question:<p>Say we have 10 images as a starting point. Then we create 10,000 images from those 10 images by adding noise, filters, flip them, skew them, distort them, etc. Isn&#x27;t the underlying data the same (or some formal definition of shannon information entropy)? Would that actually improve neural networks?<p>I&#x27;ve always wondered. Is it possible to generate infinite data and get almost perfect neural network accuracy?
评论 #22770923 未加载
评论 #22771393 未加载
评论 #22772679 未加载
评论 #22771104 未加载
评论 #22771002 未加载
评论 #22771219 未加载
评论 #22770635 未加载
评论 #22770852 未加载
评论 #22772839 未加载
edsykesabout 5 years ago
I had a read through this and I couldn&#x27;t really tell if there was something novel here?<p>I understand that perturbations and generating new examples from labelled examples is a pretty normal park of the process when you only have a limited number of examples available.
评论 #22770311 未加载
sireatabout 5 years ago
It is not the same thing but kind of reminds of my naive and obvious(meaning something that came up when drinking beer) idea of generating bunch of variations of your labeled data in cases when you do not have enough.<p>Let&#x27;s say you only have one image of dog, you generate bunch of color variations, sharpness adjustments, flips, transforms, etc. Voila you have 256 images of the same dog.<p>EDIT: I noticed that this is definitely a common idea as others have already pointed out.
master_yoda_1about 5 years ago
I am not sure how this article got ranked so high. I am suspicious about reading these article written by non experts. I would prefer to go to authentic sources and read the original paper. Most of the time information in these articles are misleading and wrong.
评论 #22771157 未加载
mattkrauseabout 5 years ago
Title is (slightly) wrong.<p>As the first paragraph says: &quot;In this post, we will understand the concept of FixMatch and also see it got 78% accuracy on CIFAR-10 with just 10 images.&quot;<p>Reporting the <i>best</i> performance on a method that deliberately uses just a small subset of the data is shady as heck.
评论 #22770695 未加载
评论 #22770744 未加载
评论 #22774452 未加载