TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Ask HN: Does a dynamically scaling cloud workstation exist somewhere?

7 pointsby eknsover 3 years ago
I frequently work with some &#x27;data science&#x27; projects that go from some MB to hundreds of GB.<p>Ideally I&#x27;d have a cloud terminal I&#x27;d connect to which could scale its RAM to fit my process RAM usage (and possibly scale up CPUs transparently too).<p>I know that you can scale up various cloud instances, but managing the runtime state is a problem. I&#x27;d like to avoid ever having to kill whatever processes I have running.<p>Something like Google&#x27;s Live Migration would also be a good match here, if it enabled migrating to a bigger machine type without rebooting, or without otherwise losing process state.<p>Ideally I&#x27;m looking for something that I could transparently scale up and down, and which I could always SSH into without having to manually start&#x2F;shutdown the instances.<p>Bonus points if GPUs could be added&#x2F;removed in the same manner.

1 comment

tgdnover 3 years ago
Have you looked into Spark? There are managed Spark options on AWS&#x2F;GCP (for example Databricks). Spark lets you do exactly what you are saying.<p>Define minimum&#x2F;maximum number of nodes, the machine capacity (RAM&#x2F;CPU) and let Spark handle the scaling for you.<p>It gives you a Jupyter-like runtime to work on possibly massive datasets. Spark is perhaps too much for what you&#x27;re looking for. Kubernetes could possibly be used with Airflow&#x2F;DBT possibly, for example for ETL&#x2F;ELT pipelines.
评论 #28855011 未加载