Luckily, this question has been asked and answered fairly recently!<p>Specifically <a href="https://news.ycombinator.com/item?id=5955043" rel="nofollow">https://news.ycombinator.com/item?id=5955043</a> from last week, complete with a response from Matt Cutts (<a href="https://news.ycombinator.com/item?id=5955374" rel="nofollow">https://news.ycombinator.com/item?id=5955374</a>):<p><i>> It's not that PG has a grudge against Google (or vice versa) or anything like that. I believe that search engine bots crawl Hacker News hard enough that PG blocks most crawling by bots. In the case of Google, he does allow us to crawl from some IP addresses, but it's true that Google isn't able to crawl/index every page on Hacker News.</i><p><i>> Here's a link where I answered the same question about three weeks ago: <a href="https://news.ycombinator.com/item?id=5837004" rel="nofollow">https://news.ycombinator.com/item?id=5837004</a> , so this isn't a new issue. In fact, PG has been blocking various bots since 2011 or so; <a href="https://news.ycombinator.com/item?id=3277661" rel="nofollow">https://news.ycombinator.com/item?id=3277661</a> is one of the original discussions about this.</i><p><i>> And to show this isn't a Google-specific issue, note that Bing's #1 result for the search [hacker news] is a completely different site, thehackernews.com: <a href="http://www.bing.com/search?q=hacker+news" rel="nofollow">http://www.bing.com/search?q=hacker+news</a> </i><p><i>> In general, I think PG's priority is to have a useful, interesting site for hackers. That takes precedence and is the reason why I believe PG blocks most bots: so that crawling doesn't overload the site.</i>