I'm still having a hard time with coding agents. They are useful but also somehow immature hence dangerous. The other day I asked copilot with GPT4o to add docstrings to my functions in a long Python file. It did a good job on the first half. But when I looked carefully, I realized the second half of my file was gone. Just like that. Half of my file had been silently deleted, replaced by a single terrifying comment along the lines of "continue similarly with the rest of the file". I use Git of course so I could recover my deleted code. But I feel I still can't fully trust an AI assistant that will silently delete hundreds of lines of my codebase just because it is too lazy or something.
> So far the biggest limiting factor is remembering to use it. Even people I consider power users (based on their Claude token usage) agree with the sentiment that sometimes you just forget to ask Claude to do a task for you, and end up doing it manually. Sometimes you only notice that Claude could have done it, once you are finished. This happens to me an embarrassing amount.<p>Yea, this happens to me too. Does it say something about the tool?<p>It's not like we are talking about luddites who refuse to adopt the technology, but rather a group who is very open to use it. And yet sometimes, we "forget".<p>I very rarely regret forgetting. I feel a combination of (a) it's good practice, I don't want my skills to wither and (b) I don't think the AI would've been <i>that</i> much faster, considering the cost of thinking the prompt and that I was probably in flow.
> The most common thing that makes agentic code ugly is the overuse of comments.<p>I've seen this complaint a lot, and I honestly don't get it. I have a feeling it helps LLMs write better code. And removing comments can be done in the reading pass, somewhat forcing you to go through the code line by line and "accept" the code that way. In the grand scheme of things, if this were the only downside to using LLM-based coding agents, I think we've come a long way.
> The product manager he sits next to has shipped 130 PRs in the last 12 months. When we look for easy wins and small tasks for new starters, it’s harder now, because he’s always got an agent chewing through those in the background.<p>I'd be curious to hear more about this, whether from the author or from someone who does something similar. When the author says "background", does that literally mean JIRA tickets are being assigned to the agent, and it's spitting back full PRs? Is this setup practical?
I have a confession. I don’t really get what Claude Code is… It’s not a model, it’s not an editor with AI integrated… So what is it? It bugs me on the website, I click on it, read, still don’t get it.<p>I have a Claude console account, if you can call it that? It always takes me 3 times to get the correct email address because it does not work with passkeys or anything that lets me store credentials. I just added the api key to OpenWebUI. It’s nice and cheaper than a subscription for me even though I use it all day.<p>But I’m still confused. I just now clicked on “build with Claude”, it takes me to that page where I put in the wrong email address 3 times. And then you can buy credits.
"Making it easy to run tests with a single command. We used to do development & run tests via docker over ssh. It was a good idea at the time. But fixing a few things so that we could run tests locally meant we could ask the agent to run (and fix!) tests after writing code."<p>Good devops practices make AI coding easier!
Having linting/prettifying and fast test runs in Cursor is absolutely necessary. On a new-ish React Typescript project, all the frontier models insist on using outdated React patterns which consistently need to be corrected after every generation.<p>Now I only wish for an Product Manager model that can render the code and provide feedback on the UI issues. Using Cursor and Gemini, we were able to get a impressively polished UI, but it needed a lot of guidance.<p>> I haven’t yet come across an agent that can write beautiful code.<p>Yes, the AI don't mind hundreds of lines of if statements, as long as it works it's happy. It's another thing that needs several rounds of feedback and adjustments to make it human-friendly. I guess you could argue that human-friendly code is soon a thing of the past, so maybe there's no point fixing that part.<p>I think improving the feedback loops and reducing the frequency of "obvious" issues would do a lot to increase the one-shot quality and raise the productivity gains even further.
Good to see experiences from people rolling out AI code assistance at scale. For me the part that resonates the most is the ambition unlock. Using Brokk to build Brokk (a new kind of code assistant focused on supervising AI rather than autocompletes, <a href="https://brokk.ai/" rel="nofollow">https://brokk.ai/</a>) I'm seriously considering writing my own type inference engine for dynamic languages which would have been unthinkable even a year ago. (But for now, Brokk is using Joern with a side helping of tree-sitter.)
> You can see this in practice when you use Claude Code, which is pay-per-token. Our heaviest users are using $50/month of tokens. That’s a lot of tokens. I asked our CFO and he said he’d be happy to spend $100/dev/month on agents. To get 20% more productive that’s a bargain.<p>fwiw we interviewd the Claude Code team (<a href="https://www.latent.space/p/claude-code" rel="nofollow">https://www.latent.space/p/claude-code</a>) and they said that even within Anthropic (where Claude is free, we got into this a bit), the usage is $6/day so about $200/month. not bad! especially because it goes down when you under-use.
> I haven’t yet come across an agent that can write beautiful code.<p>o3 in codex is pretty close sometimes. I prefer to use it for planning/review but it far exceeds my expectations (and sometimes my own abilities) quite regularly.
As someone who really dislikes using Cursor, what does the HN hivemind think of alternatives? Is there a good CLI like Claude Code but for Gemini / other models? Is there a good Neovim plugin that gets the contextual agent mode right?
Even if you don't think AI will replace the job of software developer completely, there's no way compensation is going to stay at the current level when anyone can ship code.
>Our head of product is a reformed lawyer who taught himself to code while working here. He’s shipped 150 PRs in the last 12 months.<p>>The product manager he sits next to has shipped 130 PRs in the last 12 months.<p>this is actually horrifying, lol. i haven't even considered product guys going ham on the codebase