TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: Label errors in benchmark ML test sets

2 pointsby anishathalyeabout 4 years ago

2 comments

anishathalyeabout 4 years ago
Hi HN! One of the authors here.<p>We found pervasive errors in the test sets of 10 of the most commonly used benchmark ML datasets, so we made labelerrors.com where anyone can examine the data labels. We think it’s neat to browse through the errors to get an intuitive sense of what kinds of things go wrong (e.g. completely mixed-up labels, like a frog being labeled “cat”, or situations where an image contains multiple things, like a bucket full of baseballs being labeled “bucket”), so that’s why we built this errors gallery. To our surprise, there are lots of errors, even in gold standard datasets like ImageNet and MNIST.<p>For those who want to dig into the details, we have a blog post here: <a href="https:&#x2F;&#x2F;l7.curtisnorthcutt.com&#x2F;label-errors" rel="nofollow">https:&#x2F;&#x2F;l7.curtisnorthcutt.com&#x2F;label-errors</a>, where we talk more about the study and the implications<p>Happy to answer any questions here!
anishathalyeabout 4 years ago
Also of interest might be cleanlab (<a href="https:&#x2F;&#x2F;github.com&#x2F;cgnorthcutt&#x2F;cleanlab" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;cgnorthcutt&#x2F;cleanlab</a>), the open-source software we used for initially identifying potentially mislabeled data.