TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

HEPCloud formation: handling troves of high-energy physics data

57 pointsby seycombiover 8 years ago

3 comments

boulosover 8 years ago
Disclosure: I work on Google Cloud, and helped out with this effort.<p>We&#x27;re hoping to provide some more &quot;How Fermilab did it&quot; later (maybe at NEXT?) but our blog post had a little more information [1]. End-to-end from &quot;Let&#x27;s do this!&quot; to Supercomputing was about 3 weeks (including some bugfixes to the GCE support in HTCondor).<p>For people asking, Fermilab put up about 500T into a GCS Regional bucket (into us-central1). We have great peering at Google, so I think this was streaming at about 100 Gbps from Fermilab to us.<p>As surmised, the science that we ran was lots of individual tasks, each one needing about 2 GB of RAM per vCPU, so during Supercomputing we right-sized the VMs using Custom Machine Types [2]. IIRC, several tasks needed up to 20 GB per task, so we sized PD-HDD to 20 GB per vCPU. All of this was read from GCS in parallel via gcsfuse [3]; we chose the Regional bucket for GCS to minimize cost per byte, and to maximize throughput efficiency (no reason to replicate elsewhere for this processing).<p>All the data after processing went <i>straight</i> back to Fermilab over that pipe. The output data size though was pretty small IIRC, and I don&#x27;t think we were ever much over 10Gbps on output.<p>HTCondor was used to submit work from Fermilab onto GCE directly (the submission &#x2F; schedd boxes were <i>at</i> Fermilab). and spun up Preemptible VMs. We used a mix of custom-32vCPU-64GB in us-central1-b and us-central1-c, as well as custom-16vCPU-32GB in us-central1-a and us-central1-f. You can see a graph of Fermilab&#x27;s monitoring here [4] when it was all setup. 160k vCPUs for $1400&#x2F;hr!<p>[Edit: Newlines, I always forget two newlines]<p>[1] <a href="https:&#x2F;&#x2F;cloudplatform.googleblog.com&#x2F;2016&#x2F;11&#x2F;Google-Cloud-HEPCloud-and-probing-the-nature-of-Nature.html" rel="nofollow">https:&#x2F;&#x2F;cloudplatform.googleblog.com&#x2F;2016&#x2F;11&#x2F;Google-Cloud-HE...</a><p>[2] <a href="https:&#x2F;&#x2F;cloud.google.com&#x2F;custom-machine-types&#x2F;" rel="nofollow">https:&#x2F;&#x2F;cloud.google.com&#x2F;custom-machine-types&#x2F;</a><p>[3] <a href="https:&#x2F;&#x2F;github.com&#x2F;GoogleCloudPlatform&#x2F;gcsfuse" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;GoogleCloudPlatform&#x2F;gcsfuse</a><p>[4] <a href="https:&#x2F;&#x2F;twitter.com&#x2F;googlecloud&#x2F;status&#x2F;798293201681457154" rel="nofollow">https:&#x2F;&#x2F;twitter.com&#x2F;googlecloud&#x2F;status&#x2F;798293201681457154</a>
dekhnover 8 years ago
DOE and CERN have done a really nice job with HEP, the jobs can basically flow to any compute backend without a lot of effort. Instead of saying &quot;we have a private cloud&quot; or &quot;we use the public cloud&quot;, they just say &quot;we use all the clouds&quot;.
评论 #13096854 未加载
hcrispover 8 years ago
How does the data get efficiently moved to the spun up cloud nodes? The article mentions parceling out jobs to nodes with independent storage, but how does that help data-intensive work? Are they doing something besides just copying data to the local storage of the cloud nodes? It seems like that would be a bottleneck.
评论 #13096825 未加载
评论 #13096465 未加载
评论 #13100246 未加载