TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Google Cloud Public Datasets now hosts EPA and OpenAQ air quality data

115 点作者 vgt将近 8 年前

4 条评论

panarky将近 8 年前
Important context:<p>1) US federal government has removed vast troves of public data produced by the EPA, OSHA and the Interior Department.<p>Source: <a href="https:&#x2F;&#x2F;www.washingtonpost.com&#x2F;politics&#x2F;under-trump-inconvenient-data-is-being-sidelined&#x2F;2017&#x2F;05&#x2F;14&#x2F;3ae22c28-3106-11e7-8674-437ddb6e813e_story.html" rel="nofollow">https:&#x2F;&#x2F;www.washingtonpost.com&#x2F;politics&#x2F;under-trump-inconven...</a><p>2) Google has now equipped Street View cars with air pollution sensors.<p>Source: <a href="https:&#x2F;&#x2F;environment.google&#x2F;projects&#x2F;airview&#x2F;" rel="nofollow">https:&#x2F;&#x2F;environment.google&#x2F;projects&#x2F;airview&#x2F;</a><p>Conclusion: If we can no longer rely on getting accurate data from federal agencies, private companies with a public mission will need to gather the data and make it available.
评论 #14515899 未加载
adorable将近 8 年前
Important note: so-called &quot;live&quot; air quality measurements are in reality never &quot;live&quot; due to the nature of the measuring stations (taking measures takes time) and the way data is compiled and shared by the monitoring agencies. As a result typical delays range from 1 to 6 hours, which means you end up using &quot;old&quot; data or signaling a peak when in reality the pollution peak is already over.<p>This is solved by using models that predict air quality levels down to the hour. One option is to use <a href="https:&#x2F;&#x2F;plume.io" rel="nofollow">https:&#x2F;&#x2F;plume.io</a>
vgt将近 8 年前
A not on Google Cloud Public Datasets:<p>- Public Datasets are updated almost weekly these days [0]<p>- They are hosted in Google BigQuery storage immediately accessible by Standard SQL and easily joined against other public or private datasets<p>- BigQuery has a perpetual free tier - 10GB of storage per month, 1TB of query per month. [1]<p>[0] <a href="https:&#x2F;&#x2F;cloud.google.com&#x2F;bigquery&#x2F;public-data&#x2F;" rel="nofollow">https:&#x2F;&#x2F;cloud.google.com&#x2F;bigquery&#x2F;public-data&#x2F;</a><p>[1] <a href="https:&#x2F;&#x2F;cloud.google.com&#x2F;free&#x2F;" rel="nofollow">https:&#x2F;&#x2F;cloud.google.com&#x2F;free&#x2F;</a><p>(work on G)
gaetanrickter将近 8 年前
I&#x27;m thinking of connecting these to public companies <a href="http:&#x2F;&#x2F;54.174.116.134&#x2F;recommend&#x2F;datasets&#x2F;index-hn02.html" rel="nofollow">http:&#x2F;&#x2F;54.174.116.134&#x2F;recommend&#x2F;datasets&#x2F;index-hn02.html</a> and running a few t-SNE and other clustering algo&#x27;s for visualizing the dataset.