TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Andrew Ng: Unbiggen AI

209 pointsby sbehereabout 3 years ago

13 comments

notsag-hnabout 3 years ago
I was going to interview at LandingAI. I was asked before the interview to install a spyware browser extension to monitor my traffic to detect if I was cheating during the interview. I respectfully declined and didn't have that interview.
评论 #30941458 未加载
评论 #30942275 未加载
评论 #30941634 未加载
评论 #30942127 未加载
评论 #30944778 未加载
whatever1about 3 years ago
My understanding is that they are trying to automate the data preparation steps that seasoned ML practitioners are doing anyway today.<p>The fact that he tries this in manufacturing makes the case stronger. In most manufacturing companies you do not have access to top ML talent.<p>You have Greg who knows python and recently visualized some production metrics.<p>If we could empower Greg with automated ML libraries that guide him in the data preparation steps in combination with precooked networks like autogluon, then manufacturing could become a huge beneficiary of the ML revolution.
评论 #30940571 未加载
评论 #30942657 未加载
itissidabout 3 years ago
That is the problem with generalization and cop outs like these. It&#x27;s no good to people in the field doing actual work where the devil is in the detail.<p>Big data is fairly important to a lot of things, for example I was listening to Tesla&#x27;s use of Deep net models where they mentioned that there were literally so many variations of Stop Signs that they needed to learn what was really in the &quot;tail&quot; of the distribution of Stop Sign types to construct reliable AI
评论 #30940392 未加载
评论 #30940394 未加载
评论 #30940941 未加载
a_square_pegabout 3 years ago
I’ve been wondering about the limits of data-centric approach – there seems to be this implicit notion that more data equals better performing ML or AI. I think it would be interesting to imagine a point of diminishing return on additional data if we consider that our ability to perceive is probably largely based on two parts - sensory input and knowledge. Note that I’m making an explicit distinction here on the difference between data and knowledge.<p>For instance, an English speaker and a non-English speaker may listen to someone speaking English and while the auditory signals received by both are the same, the meaning of the speech will only be perceived by the English speaker. When we’re learning a new language, it’s this ‘knowledge’ aspect that we’re enhancing in our brain, however that is encoded.<p>This knowledge part is what allows us to see what’s not there but should be (e.g. the curious incident of the dog in the night) and when the data is inconsistent (e.g. all the nuclear close calls). I’m really not sure how this ‘knowledge’ part will be approached by the AI community but feel like we’re already close to having squeezed out as much as we can from just the data side of things.<p>Somewhat related, we have a saying in Korean – ‘you see as much as you know’.
评论 #30941845 未加载
评论 #30943864 未加载
评论 #30941770 未加载
aj7about 3 years ago
“I once built a face recognition system using 350 million images.”<p>Did this make any of you a little queasy?
评论 #30941867 未加载
a-dubabout 3 years ago
data quality is important. every ai project i&#x27;ve worked on has started with visualizing the data and thinking about it.<p>it&#x27;s easy to get complacent and focus on building big datasets. in practice, looking at the data often reveals issues sometimes in data quality and sometimes scope of what&#x27;s in there (if you&#x27;re missing key examples, it&#x27;s simply not going to work).<p>most ml is actually data engineering.
atbpacaabout 3 years ago
Glad to see the term ML being used more often than AI in the comments as it looks like most &quot;AI&quot; models are trained for image classification. Having said that, the idea of &quot;doing more with less&quot; sounds interesting and I wonder what it means exactly. Does it mean taking a dataset of 50 images and to create 1000s of synthetic images from it?
评论 #30941431 未加载
DeathArrowabout 3 years ago
Pretty interesting. Mr. Ng claims that for some applications having a small set of quality data can be as good as using huge set of noisy data.<p>I wonder if, assuming the data is of highest quality, with minimal noise, having more data will matter for training or not. And if it matters, on what degree?
评论 #30940931 未加载
评论 #30942904 未加载
xiphias2about 3 years ago
I can imagine that customizing AI solutions in an automated way is quite important, but writing that as the next wave is probably an overstatement.<p>Of course few shot learning is important for models, but for example for Pathways it was already part of the evaluation.
kappiabout 3 years ago
For industrial application, there are already mature systems based on CV. For majority of those applications, there is no need for deep learning or multilayer CNN. Shocked to see Andrew Ng talking like a marketing guy.
leobgabout 3 years ago
What are some ML data annotation tools that guide you towards those data points where the model gets confused? I hear Prodigy does this. Any others?
评论 #30941939 未加载
TOMDMabout 3 years ago
Yeah that&#x27;d be great.<p>I also want cars that run on salt water.<p>I&#x27;m not saying that small data ai is equally impossible, but simply saying &quot;we should make this better thing&quot; isn&#x27;t enough.
评论 #30940306 未加载
评论 #30940409 未加载
评论 #30942058 未加载
评论 #30940858 未加载
评论 #30940284 未加载
tacosbaneabout 3 years ago
can we build an AI to detect that the AI goalposts keep getting moved?
评论 #30944010 未加载