TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: Weave - actually measure engineering productivity

22 pointsby adchurch6 months ago
Hey HN,<p>We’re building Weave: an ML-powered tool to measure engineering output, that actually understands engineering output!<p>Why? Here’s the thing: almost every eng leader already measures output - either openly or behind closed doors. But they rely on metrics like lines of code (correlation with effort: ~0.3), number of PRs, or story points (slightly better at ~0.35). These metrics are, frankly, terrible proxies for productivity.<p>We’ve developed a custom model that analyzes code and its impact directly, with a far better 0.94 correlation. The result? A standardized engineering output metric that doesn’t reward vanity. Even better, you can benchmark your team’s output against peers while keeping everything private.<p>Although this one metric is much better than anything else out there, of course it still doesn&#x27;t tell the whole story. In the future, we’ll build more metrics that go deeper into things like code quality and technical leadership. And we&#x27;ll build actionable suggestions on top of all of it to help teams improve and track progress.<p>After testing with several startups, the feedback has been fantastic, so we’re opening it up today. Connect your GitHub and see what Weave can tell you: <a href="https:&#x2F;&#x2F;app.workweave.ai&#x2F;welcome" rel="nofollow">https:&#x2F;&#x2F;app.workweave.ai&#x2F;welcome</a>.<p>I’ll be around all day to chat, answer questions, or take a beating. Fire away!

10 comments

senko6 months ago
&quot;Hello Jane, please have a seat. We need to talk about your productivity. Yes, I know you helped the team through a crunch and delivered the new feature, which works flawlessly and is loved by our users. And our balance sheet is much healthier after you found that optimization that saves us $1mm&#x2F;year. We also appreciate that younger teammates look to you for guidance and learn a lot from you.<p>But you see, the AI scored your productivity at 47%, barely &quot;meets expectations&quot;, while we expect everyone to score at least 72%, &quot;exceeds expectations&quot;. How is that calculated? The AI is a state of the art proprietary model, I don&#x27;t know the details...<p>Anyways, we&#x27;ve got to design a Personal Improvement Plan for you. Here&#x27;s what our AI recommends. We&#x27;ll start with the TPS reports...&quot;
评论 #42198741 未加载
rkagerer6 months ago
How did you come up with those magic correlation numbers?<p>Is this generally just sniffing surface quality and quantity of written code, or is consideration given to how architecturally sound the system is built, whether the features introduced and their implementations make sense, how that power is exposed to users and whether the UI is approachable and efficient, user-feedback resulting from the effort, long-term sustainability and technical debt left behind (inadvertently or with deliberation), healthy practices for things like passwords &amp; sensitive data, etc?<p>I&#x27;m glad to see an effort at capturing better metrics, but my own feeling is trying to precisely measure developer productivity is like trying to measure IQ - it&#x27;s a flawed errand and all you wind up capturing is one corner of a larger picture. Your website shares zero information prior to login, and I&#x27;m looking forward to you elaborating a little more on your offering!<p>EDIT: Would also love to hear feedback from developers at the startups you tested at - did they like it and felt it better reflected their efforts during periods they felt productive vs. not, was there any initial or ongoing resistance &amp; skepticism, did it make managers more aware of factors not traditionally captured by the alternative metrics you mentioned, etc.
评论 #42198642 未加载
adchurch6 months ago
Our metric is approximately &quot;hours of work for an expert engineer.&quot; Here are some example open source PRs and their output metrics calculated by our algorithm:<p><a href="https:&#x2F;&#x2F;github.com&#x2F;PostHog&#x2F;posthog&#x2F;pull&#x2F;25056">https:&#x2F;&#x2F;github.com&#x2F;PostHog&#x2F;posthog&#x2F;pull&#x2F;25056</a>: 15.266 (Adds backend, frontend, and tests for a new feature)<p><a href="https:&#x2F;&#x2F;github.com&#x2F;microsoft&#x2F;vscode&#x2F;pull&#x2F;222315">https:&#x2F;&#x2F;github.com&#x2F;microsoft&#x2F;vscode&#x2F;pull&#x2F;222315</a>: 8.401 (Refactors code to use a new service and adds new tests)<p><a href="https:&#x2F;&#x2F;github.com&#x2F;facebook&#x2F;react&#x2F;pull&#x2F;27977">https:&#x2F;&#x2F;github.com&#x2F;facebook&#x2F;react&#x2F;pull&#x2F;27977</a>: 5.787 (Small change with extensive, high effort tests; approximately 1 day of work for expert engineer)<p><a href="https:&#x2F;&#x2F;github.com&#x2F;microsoft&#x2F;vscode&#x2F;pull&#x2F;213262">https:&#x2F;&#x2F;github.com&#x2F;microsoft&#x2F;vscode&#x2F;pull&#x2F;213262</a>: 1.06 (Mostly straightforward refactor; well under 1 day of work)
评论 #42198054 未加载
评论 #42197826 未加载
jaredsohn6 months ago
If you build something that doesn&#x27;t solve problems with impact to the business, your real productivity is zero. How does this account for that?<p><a href="https:&#x2F;&#x2F;blog.pragmaticengineer.com&#x2F;the-product-minded-engineer&#x2F;" rel="nofollow">https:&#x2F;&#x2F;blog.pragmaticengineer.com&#x2F;the-product-minded-engine...</a>
评论 #42203378 未加载
评论 #42199744 未加载
henning6 months ago
As soon as people know how the metric is calculated, they will game that metric and it will cease to be useful.
评论 #42198397 未加载
评论 #42197870 未加载
id006 months ago
Let me just ignore my natural distain to the whole thing (as a engineer and a manager)<p>&gt; We’ve developed a custom model that analyzes code and its impact directly...<p>This is a bold claim all things considering. Don&#x27;t you need to fine tune this model for every customer as their business metrics likely vastly different? How do you measure the impact of refactoing? What about regressions or design mistakes that surface themselves after months or even years?
评论 #42198795 未加载
评论 #42198819 未加载
jaredsohn6 months ago
I&#x27;m looking forward to developers setting up LLM prompts to make their code seem more complex and like it required more effort.
itsdrewmiller6 months ago
What do you see as the major threats to validity for your approach?
mg576 months ago
Pretty dumb to think you can infer effort from the code itself. You make one &quot;smart invocation&quot; to a remote microservice and replace 1000 lines of code!<p>The information for effort is not available at the code level - sorry to burst your bubble.
评论 #42198476 未加载
adambeecee6 months ago
Hey HN! I&#x27;m one of the co-founders of Weave, and I wanted to jump in here to share a bit more.<p>Building this has been a wild ride. The challenge of measuring engineering output in a way that’s fair and useful is something we’ve thought deeply about—especially because so many of the existing metrics feel fundamentally broken.<p>The 0.94 correlation is based on rigorous validation with several teams (happy to dive into the details if anyone’s curious). We’re also really mindful that even the best metrics only tell part of the story—this is why our focus is on building a broader set of signals and actionable insights as the next step.<p>Would love to hear your thoughts, feedback, or even skepticism—it’s all helpful as we keep refining the product.
评论 #42197766 未加载
评论 #42197855 未加载