Should be an interesting research tool to see how often we use shitty commit messages when trying to get something fixed...<p><a href="https://github.com/search?utf8=%E2%9C%93&q=fix&type=Commits&ref=searchresults" rel="nofollow">https://github.com/search?utf8=%E2%9C%93&q=fix&type=Commits&...</a>
I'm curious how long it took to build their search index, and how they do it/what technologies/algorithms they are using. That's a lot of data!
Google put all GitHub data in BigQuery a while back [1]. What took them so long?<p>[1] <a href="https://cloud.google.com/bigquery/public-data/github" rel="nofollow">https://cloud.google.com/bigquery/public-data/github</a>