Google BigQuery has pure separation of storage and compute, which allows for Public Datasets [0] to exist in ready-to-query highly optimized format. Run SQL immediately!<p>BigQuery has a perpetual free query tier of 1 Terabyte per month ($5). In addition, you can a get $300 in Google Cloud credits for two months to do more work [1].<p>(work on Google Cloud and used to work on BigQuery)<p>Edit: spaces, how do they work?!?!<p>[0] <a href="https://cloud.google.com/bigquery/public-data/" rel="nofollow">https://cloud.google.com/bigquery/public-data/</a><p>[1] <a href="https://cloud.google.com/free-trial/" rel="nofollow">https://cloud.google.com/free-trial/</a>
For me, the killer dataset would be the Google Scholar data. That would just blow the whole scientometric space wide open. It would also be a nice introduction for researchers into BigQuery (calculate your own H-factor)
I'd highly recommend anyone who hasn't played with BigQuery yet to take a few moments to give it a shot. I was highly skeptical at first when my old company moved to it, but now I've come to love it as a platform.<p>What I love most:
[0] Exceptionally fast queries against large datasets.
[1] Very Cost Effective (although as others have called out, it can be accidentally misused resulting in a big bill).
[2] Non data-engineers can setup, use, and manage, with minimal difficulty.
[3] Gets you away from high-priced solutions like Vertica or Teradata.
[4] No management headaches like Redshift.<p>Downsides:
[0] Quotas can get annoying to work with.
[1] Not a ton of wrappers in a diverse set of languages.
[2] Not a ton of support with desktop SQL clients.
The complexity of the Stack Overflow Data Explorer was the only reason I never played around with SO Data.<p>I am looking through the tables now and there is certainly a lot of cool stuff that can be done! :)<p>Although, there appear to be a few tables with garbage data and only few rows, like posts_privilage_wiki and posts_wiki_placeholder.
Hey, I figured it out!<p><a href="https://bigquery.cloud.google.com/savedquery/809799891616:6bda226894244e1f82ed214cfd1c2af3" rel="nofollow">https://bigquery.cloud.google.com/savedquery/809799891616:6b...</a><p>There's a learning curve. Getting in without giving up your credit card isn't exactly intuitive. And Leslie isn't the most popular name in the US -- it's gender neutral name, so there are lots of rows.
I really hope GCP continues to do this with other types of syndicated public data feeds like American FactFinder, FRED, etc. They are missing a substantial audience that would see this as a killer app and compelling reason to move to GCP for this. The success of Enigma.io suggests there is a legit (although easily reproduced) business case here with minimal effort. Anyone who has ever worked to prep and integrate this type of data could now spend that time doing science.
I have two Google accounts and BigQuery has a problem if I want to use the not default one. It inhibited my curiosity a bit.<p>I had an idea to query the usage of string handling functions in C code bases, so I could do something like a manual linting around them.
Very cool. I love the idea of public data sets like these.<p>I wonder if AWS can swing somethings similar with Athena. They already have "requester pays" buckets for S3 so should be inline with that to have a similar offering for Athena connected to S3 resources.
Did StackOverflow agreed to offer the data or Google just took because they can?
Stackoverflow has their own query tool. Why would they give their data to Google?