TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Attention and Memory in Deep Learning and NLP

97 pointsby dennybritzover 9 years ago

3 comments

MrQuincleover 9 years ago
Early work on attention models is done by Itti, Koch, and Niebur. [1,2]. It&#x27;s called &quot;saliency&quot; and I think Denny would consider this more along the lines of what the concept of &quot;attention&quot; should be (considering his own words&#x2F;reservations in using the term). Koch is currently studying neural correlates, Itti is still working on this topic though. Niebur is into the neuroscience part of it (nematode expert).<p>There is a lot of neuroscientific work on attention, really a lot! Overt and covert attention. Microsaccades, very small eye movements, with already a bunch of possible functional roles. Almost everything we know about the brains of little kids is by studying where they look at and where they pay attention to.<p>Structure-wise attention models can be quite simple. The structure that is often seen is a WTA (winner-take-all) network with subsequent serial inhibition. The first winner is inhibited, so the next winner can come on stage. This is the same system as Baars has in his global workspace theory [3]. It is also the same method as in mundane RANSAC models [4]. That&#x27;s a workhorse of computer vision in which a consensus&#x2F;voting model can be used to have data points voting for higher-level structures. When one structure is detected, votes for it are removed, and the next most salient structure can be voted for.<p>[1] <a href="http:&#x2F;&#x2F;ilab.usc.edu&#x2F;bu&#x2F;" rel="nofollow">http:&#x2F;&#x2F;ilab.usc.edu&#x2F;bu&#x2F;</a><p>[2] <a href="http:&#x2F;&#x2F;cns-alumni.bu.edu&#x2F;~yazdan&#x2F;pdf&#x2F;Itti_etal98pami.pdf" rel="nofollow">http:&#x2F;&#x2F;cns-alumni.bu.edu&#x2F;~yazdan&#x2F;pdf&#x2F;Itti_etal98pami.pdf</a><p>[3] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Global_Workspace_Theory" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Global_Workspace_Theory</a><p>[4] <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;RANSAC" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;RANSAC</a>
andreykover 9 years ago
&quot;I consider the approach of reversing a sentence a “hack”. It makes things work better in practice, but it’s not a principled solution.&quot;<p>I had the same feeling about boundary box recommendations&#x2F;guesses that were used to speed up object recognition with Deep Learning fairly recently. Just as with a sliding box approach it is intuitive and works, but it also seems quite inelegant and like a better approach should be possible. Visual attention seems like it should work much better in the long term, so it is exciting the field has come to a point where it has been developed.
zappo2938over 9 years ago
I&#x27;m curious which people on Hacker New are interested in this field?
评论 #10834248 未加载