TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

MLflow v0.8.0 Features Improved Experiment UI and Deployment Tools

51 pointsby dmatrixover 6 years ago

3 comments

m_keover 6 years ago
I&#x27;m looking into switching over to using MLflow or Polyaxon for experiment management and tracking. We currently us a a custom built django app for experiment tracking and run experiments by hand on desktop workstations but we&#x27;re starting to move some of that over to GCP.<p>For people who have used either of the projects, what are your opinions and are there any hidden issues that you ran into?<p>Ideally we&#x27;d like to have a platform that makes it easy to schedule runs on the desktops or GCP depending on requirements and available resources. Seems like kubernetes might be the best option for that and it doesn&#x27;t look like MLflow supports it out of the box yet.
评论 #18504863 未加载
评论 #18505525 未加载
评论 #18505747 未加载
评论 #18507251 未加载
antisocialover 6 years ago
We are evaluating MLflow. I would like to know if there are any plans for making this an Apache project?
mlthoughts2018over 6 years ago
As an ML engineer, I’ve found MLFlow to be really a disastrously bad way to look at the problem. It’s something that managers or executives buy into without understanding it, and my team of engineers (myself included) have hated it.<p>There are many feature specific reasons, but the biggest thing is that reproduction of experiments needs to be synonymous with code review and the identically same version control system you use for other code or projects.<p>This way reproducibility is a genuine constraint on deployment and deployment of an experiment, whether just training a toy model, incorporating new data, or really launching a live experiment, is conditional on reproducibility and code review of the code, settings, runtime configs, etc., that embodies it totally.<p>This is much better solved with containers, so that both runtime details and software details are located in the same branch &#x2F; change set, and a full runtime artifact like a container can be built from them.<p>Then deployment is just whatever production deployment already is, usually some CI tool that explains where a container (built from a PR of your experiment branch for example) is deployed to run, along with whatever monitoring or probe tracking tools you already use.<p>You can treat experiments just like any other deployable artifact, and monitor their health or progress exactly the same.<p>Once you think of it this way, you realize that tools like ML Flow are <i>categorically</i> the wrong tool for the job, almost by definition, and they exist mostly just to foster vendor lock-in or support reliance on some commercial entity, in this case Databricks.
评论 #18512925 未加载
评论 #18507376 未加载
评论 #18507050 未加载