TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Ask HN: How have you integrated LLMs in your development workflow?

63 点作者 mjbale1168 个月前
Is it working for you? have you seen any performance increases?

42 条评论

Arch4858 个月前
I&#x27;ve tried it, and I don&#x27;t like it. There&#x27;s too much confabulation to be useful as code completion, and any task more complicated than that results in logic errors most of the time.<p>Code completion was fine without LLMs, and solving problems myself usually ends up being quicker than trying to coerce an LLM into doing it correctly and then verifying that the output is actually correct.<p>The one time I used an LLM in my workflow to good success was using ChatGPT to automatically create an enum of every EU&#x2F;EEA country and a switch statement over that enum. Those sort of &quot;grunt work&quot; tasks that don&#x27;t require any thinking, but a lot of typing, seem to be where LLMs shine.
评论 #41690285 未加载
aeze8 个月前
I use Cursor primarily with Claude 3.5 Sonnet. Overall a solid productivity increase depending on the task.<p>I have a few observations:<p>- I vastly prefer Cursor&#x27;s Copilot++ UX for autocomplete compared to GitHub&#x27;s in VSCode, which I used until a few months ago.<p>- The Composer multi-file editor (cmd+i) is easily its most powerful feature and what I use most often, even when I&#x27;m working on single files. It just works better for some reason.<p>- It&#x27;s far more effective working in popular stacks, eg. Typescript&#x2F;NextJS etc. It&#x27;s rarely a time-saver when working in Elixir, for example.<p>- In a similar vein, the less &#x27;conventional&#x27; your task or code is, the less useful it becomes.<p>- As the context increases, it gets noticeably less useful. I often find myself having to plan what context I want to feed it and resetting context often.<p>- It&#x27;s very effective at &#x27;translation&#x27; tasks, eg. converting a schema from one format to another. It&#x27;s much less effective at generating complex business logic.<p>- I only find it useful to generate code I confidently know how to write myself already. Otherwise, it doesn&#x27;t save me time. The times I&#x27;ve been tempted, it&#x27;s almost always bitten me.
评论 #41685740 未加载
评论 #41685999 未加载
评论 #41706483 未加载
troupo8 个月前
I have seen none across ChatGPT, Claude and Copilot.<p>They may be somewhat useful for some small function&#x2F;lookup, but:<p>- they will always, without fail, hallucinate non-existent functions, methods, types and libraries. Your job becomes a constant code review, and I&#x27;m not here to babysit an eager junior<p>- claude will always, without fail, rewrite and break your code, and any fixes you&#x27;ve done to Claude&#x27;s code. Even if you tell it not to. Again, you end up babysitting an eager junior<p>- fo any question outside of StackOverflow&#x27;s top-3 languages and top-10 libraries they are next to useless as they know nothing about anything<p>These are not &quot;AIs&quot; or code assistants or autocompletes. These are generic token prediction machines which don&#x27;t care if you write code or Harry Potter fanfics.<p>If you train them specifically on vast amounts of code and documentation, then maybe, just maybe, they may actually be useful
vander_elst8 个月前
Code autocompletion. I think it works around 60% of the cases, sometimes it saves some tipping. Performance increase: small. Additionally, I don&#x27;t learn the API as good as if I were to type it, considering moving back.
评论 #41685613 未加载
评论 #41644151 未加载
codeptualize8 个月前
Yes, I use copilot, Claude, and sometimes Zed with the built in Claude AI tooling.<p>It&#x27;s really useful for basic stuff, scripting, boilerplate, repetitive things, SQL. For example it&#x27;s great at converting types from different languages, or generate types from API reference docs.<p>You gotta be careful though as it is not perfect and if you do something &quot;not typical&quot; it will not work well. Also sometimes it is a bit outdated, produces code with different &quot;style&quot;, or just plain wrong stuff, so you have to tweak it a bit.<p>It&#x27;s not going to code for you yet (unless you do very basic stuff) but it&#x27;s a great tool to increase productivity. I do believe I move faster.
kleiba8 个月前
People who use code completion LLMs: is your company running a local install, or are they fine with your code getting shared with LLMs in the cloud?
评论 #41685749 未加载
评论 #41685761 未加载
评论 #41685718 未加载
评论 #41685702 未加载
评论 #41691533 未加载
评论 #41685719 未加载
nicbou8 个月前
I write more than I code, but I use ChatGPT a lot throughout the day.<p>I am integrating AI translations into my custom static site generator. I will test the outcome heavily before putting my name (and a big warning) on my translated content, but the early results look good. Getting it right is a lot harder than piping the page through ChatGPT. Everything needs to be translated from the UI strings to the URL slugs.<p>My work will no longer just benefit English-speaking immigrants, but also locals and immigrants who Google things in other languages. I am very excited about it.<p>I also use ChatGPT heavily for “what’s the word&#x2F;expression for” questions and other fuzzy language queries. As a non-native speaker, I want to know if a given expression is appropriate for the context, if it’s dated, if it’s too informal and if it’s dialect-appropriate.<p>I also use it for coding, but mostly because it’s faster than reading python’s docs. I ask it questions I know the answer to, hoping to find better approaches. So much happened since Python 2.7, and I don’t always know what to ask for.<p>On occasion I treat it like I treated my mom as a child. I ask it all sorts of questions about the world I observe around me. It’s amazing to get a short, clear answer to work from, instead of sifting through barely relevant, spammy search results. This is super helpful when getting to know new tech and figuring out what it actually is.<p>It’s just so damn cool to have a sort of real-life Hitchhiker’s Guide slash Pokédex in my pocket. These things appeared in the span of a year, and nobody seems impressed. Well, I am mad impressed.
评论 #41695050 未加载
danielbln8 个月前
I&#x27;ve been using Zed with Claude 3.5 for a few weeks now, and I find it incredibly useful. Being able to be in full control what goes into the context, my workflow is usually<p>- add files, tabs, terminal output and IDE diagnostics into the context via slash commands - feed in documentation or other remote web content, also via simple slash commands - activate workflow mode, which will help you edit the files instead of having to copy things around - then ask questions, ask for refactoring, ask for new features<p>Often I like to ask for the high level approach, and if I agree with it let it guide the implementation. It makes mistakes, so I always have to validate and test what it creates, but as you add more information into the context and the LLM has a good amount of stuff to work with, the output quality really improves significantly.<p>It&#x27;s a process though, and it takes time to get familiar with the workflow, build intuition when the LLM falls on its face, when you should try a different approach etc.<p>Generally I can echo what others have said, it works best if you already kind of know what you want to achieve and just use the LLM as an assistant that does the grunt work for you, documents the process, provides a rubber duck etc.<p>Generally, I would not want to work without an integrated LLM anymore, it provides that much value to my workflow. No panacea, no silver bullet, but when used right in the right circumstances it can be incredibly useful.<p>A secondary usecase for me is working on repositories where tasks and todos are structured in markdown files (even things like travel planning). Letting the LLM Guide you through todos and create a documentation trail through the process, identify important todos, carry along information as you go is wonderful, I would absolutely give that a try as well.
threeseed8 个月前
For common languages e.g. Python, Javascript it works reasonably well although it still seems to prefer older versions of libraries way too often. Which means you need to do refactoring straight out of the bat.<p>Whenever I use it with Rust, Golang, Scala etc it&#x27;s not worth the effort.
dm38 个月前
I find LLMs via Aider great for:<p>* code translation - e.g. convert a self-contained implementation of a numerical algorithm from one language to another and generate test cases and property tests which make sure the implementations are equivalent. The goal is to avoid having to proof read the generated code.<p>* one-off scripts - any task where code design doesn&#x27;t matter, the amount of code is limited to couple hundred lines (GPT-4o) and the result will be thrown away after use.<p>* API exploration - producing examples for APIs and languages I&#x27;m not fluent in. Reading reference documentation gives a better understanding, LLMs get the results out faster.
评论 #41685738 未加载
kutenai8 个月前
I use Tabnine - <a href="https:&#x2F;&#x2F;www.tabnine.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.tabnine.com&#x2F;</a> It supports multiple LLM&#x27;s and it has it&#x27;s own internal one, and it can be secure for a company and their codebase, unlike the generic OpenAI, Claude, etc.<p>I do have OpenAI keys and Claude, and I use them both (just as the mood fits, to see which works best).<p>I&#x27;m been coding for decades, so I&quot;m quite experienced. I find that an LLM is no substitute for experience, but it definitely helps with progress. I work regularly in a range of languages: Java, Javascript with Typescript, Javascript pure ESM, Python, SQL. It&#x27;s great to have a quick prototype tool.<p>One key takeaway - learning to &quot;drive the LLM&quot; is a skill by itself. I find that some people are &quot;hesitant&quot; to learn this, and they usually complain about how bad the LLM is at generating code.. but, in reality, they are bad at &quot;driving&quot; the LLM.<p>If I put you in an F1 car, the car would perform perfectly, but unless you had the skills to handle the car, you will not win any races.. might not get around the track one time..<p>Also, I&#x27;m in my 60&#x27;s so, this is all &quot;new&quot; tech. I&#x27;ve just never been afraid of &quot;new&quot; tech. I&#x27;d hate for some 30 year old hot-shot to show my up because they learned to master using that LLM Tool and I just blew it off as &quot;new tech&quot;.<p>Anyway, my $0.02
jimcollinswort18 个月前
I&#x27;ve been using the cursor editor (vs code clone with LLM) for 6 months and it definitely improves productivity. Biggest gains are for writing&#x2F;refactoring just a few lines at a time, where I can understand and test the new code easily.<p>Working with multiple products, Python and many different libraries and frameworks I have to constantly look up commands and usage. With the LLM I select a line or two, and ask &#x27;sort by region, exclude pre 2023&#x27; and it writes (or adds) the necessary Pandas calls to the dataframe. The LLM has my code, tools documentation and more as context, so I don&#x27;t have to say much to get the right code, but the questions are important, have to converse with a programmer mindset still.<p>It has almost completely replaced using Google for helping with code. Often I waste too much time in Slashdot looking at a dozen possible similar situations, but not what I need. The LLM immediately gives me something right in my own code. Usually have to give it some followup commands to tweak the code or take a different approach.
smokel8 个月前
The competition in AI editors is a bit silly at the moment. Everyone and their dog are &quot;building&quot; an AI assisted editor now by duct taping Ollama onto VS Code. I don&#x27;t like my data being sent to untrusted parties, so I cannot evaluate most of these. On top of that, the things keep evolving as well, and editors that I dismissed a few months ago, are now all of a sudden turning into amazing productivity boosters, thanks to developments in both models as well as in editor tricks.<p>My money is on Cursor [1], which does not stop to amaze me, and seems to get a lot of traction. The integration is very clever, and it is scary how it figures out what I intend to do. Then again, I&#x27;m probably doing mundane tasks most of the time. For the few bright moments in my day I tend to use ChatGPT, because most of my real problems are in application domains, not in code.<p>I am not a firm believer in forking large open-source projects, though, as it will take a lot of effort to keep up with future diversions. This makes me a bit wary of projects such as Cursor and Void [2]. Somebody needs deep pockets to sustainably surpass the popularity of VS Code. To point out just one problem with forking: VS Code works fine in Ubuntu, but Cursor does not work out of the box there. Having to disable the sandbox is a show-stopper for most.<p>In that respect, the extensions might be a safer bet, and I think Sourcegraph&#x27;s Cody and Continue are making the largest waves there. Hard to tell with so many waves.<p>[1] <a href="https:&#x2F;&#x2F;www.cursor.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;www.cursor.com&#x2F;</a><p>[2] <a href="https:&#x2F;&#x2F;voideditor.com&#x2F;">https:&#x2F;&#x2F;voideditor.com&#x2F;</a>
评论 #41686476 未加载
danmaz748 个月前
Until recently I was only using Copilot as an advanced autocomplete and to help writing tests, for which it&#x27;s pretty useful.<p>A couple weeks ago I had to create some classes (typescript) implementing a quite big spec for a file format to transfer quizzes between applications. I decided to try some more advanced tool, ending up with Cursor and continue.dev. I copied the spec (which are public on the web) into a text file and used them as context, together with a skeleton start for the main class I needed, and experimented with different engines and different prompts.<p>The end result is that I got a very good starting point for my code, saving me many hours. Surprisingly, for this task, the best result was generated by Gemini 1.5 pro.<p>Since then I&#x27;ve been trying to integrate these tools more into my day to day programming, with varying results, but I&#x27;m now convinced that, even with the limits of the current LLMs, this technology can have a much higher impact on programming with better harnessing, eg integrating it with compiler&#x2F;code analysis tools output and automated testing.
resonious8 个月前
Asking ChatGPT to give me boilerplate is a huge productivity win for me still. For example, I recently asked ChatGPT to give me a Terraform file that sets up API Gateway, Lambda, and DynamoDB. The script it gave me worked after only a couple of minor tweaks. Was up and running in about 15 minutes, whereas I&#x27;m pretty sure it&#x27;d be hours if I had to dredge through the docs myself.
cl428 个月前
100%. I usually use it for prototyping new features.<p>Two examples...<p>1. Recently wanted to build a Chrome plugin and never built one before. Used o1-preview to build it all for me.<p>2. Wanted to build a visualization of the world with color-coded maps using D3. Again, hadn&#x27;t used D3 much in the past... Claude basically wrote all the code for me and then I just had to make edits to fit my site&#x2F;template.
moltar8 个月前
One use cases I like a lot is copy paste the entire error from the console into the chat window. No explanation or preamble. Just raw output. Most of the time I get some useful leads out of it. It’s not always solving it entirely for me but even a hint in the right direction can save you hours of digging in the wrong direction.
tlarkworthy8 个月前
I wrote my own Notebook Assistent and I use it to quickly make data driven arguments. It is able to both code and read program outputs so it&#x27;s very good at pulling remote sources, then peeking at the result, doing the data wrangling and then displaying the output. I think having the LLM being able to read program state is a big upgrade for what you can achieve, it can do test driven development for example because it can read the test report. However, being a side project limits the fluidity of the UX, it&#x27;s open source and buildless though if you want to modify it.its very hackable by design. <a href="https:&#x2F;&#x2F;observablehq.com&#x2F;@tomlarkworthy&#x2F;robocoop" rel="nofollow">https:&#x2F;&#x2F;observablehq.com&#x2F;@tomlarkworthy&#x2F;robocoop</a>
kbrecordzz8 个月前
I have not, LLM is generative and can only create generic stuff and if your project contains lots of generic stuff, is it really interesting? I want to spend time on the parts I find interesting to think thoughts about with my human brain. Problems that have a quick answer in ChatGPT but a more interesting and personal answer after thinking about it for hours and trying 3 different versions of it before deciding on one. Then if I want to quickly test how to create an X11 window I might generate the code and work backwards from that. But… I’m not really in a hurry and if you aren’t you don’t really need LLM
atoav8 个月前
I haven&#x27;t except on one or two occasionally where I tried it to use me help figuring out how to refactor a bunch of functions quickly.<p>The downside of LLMs is that it doesn&#x27;t remove your need to know your code. And writing code yourself is a very good way to know it.
anonzzzies8 个月前
We have our own tooling; it&#x27;s very good and definitely increases productivity for most. One observation is that people who are bad at <i>reading</i> code seem often to be get negatively impacted vs writing themselves. So from our teams I think that code reading will be more valuable than writing, if it was not already. Just having a loop of telling the llm something, running it and then tell what&#x27;s wrong with the observed result is very slow vs reviewing the code and tell what&#x27;s wrong with it instead.<p>We also use cursor, copilot, claude dev, aider and plandex to see if anything is beating anything else and to get ideas from for our toolkit.
gbro3n8 个月前
ChatGPT o1 preview is producing amazing results for code and generally helpful with life. It&#x27;s not well integrated to many IDE tools yet (I use Cody a bit which has a wait list for o1 preview), but then I keep going back to just using the ChatGPT interface anyway because it feels cleaner to manually select and upload the relevant files attachments anyway. I&#x27;m sure that will change as integrations improve. Claude seems really popular on here but I wonder if that&#x27;s just because Sonnet is free and useful enough. I prefer the results of newer ChatGPT models currently and feel they are worth paying for.
paradite8 个月前
I built a simple desktop tool [1] to streamline my coding workflow with LLMs, like integrating with local code context, managing prompts and custom instructions, etc. It supports copy&#x2F;paste into ChatGPT&#x2F;Claude, and sending prompts via API.<p>Currently I estimate it writes about 70% to 80% code (including adding features to the tool itself), and saved me hours of work per week.<p>Lately I&#x27;ve been exclusively using API since it is cheaper than paying $20 monthly subscription.<p>[1] <a href="https:&#x2F;&#x2F;prompt.16x.engineer&#x2F;" rel="nofollow">https:&#x2F;&#x2F;prompt.16x.engineer&#x2F;</a>
forgotlastlogin8 个月前
I am currently working on a side project. I am not very familiar with the domain, and I am using a language I am not very proficient at. Claude has been a great help, discussing different approaches to a problem, sometimes writing (and explaining) a function or two. Or explaining an error.<p>On top of that, I have configured a chatgpt code reviewer on my github repos. Most of the comments I get from it on PRs are useless but from time to time it does spot a problem or suggest an improvement.
WD-428 个月前
Code completion. I’m about 50&#x2F;50 on whether it’s actually helpful or I just spend more time reading the completions to make sure they are correct instead of just learning and typing it out myself.
yko8 个月前
Writing documentation, amongst other things, has been a save for me.<p>I use it to understand new codebases quickly, create the documentation boilerplate for the code I&#x27;m working on that needs better docs, or update&#x2F;rewrite outdated ones.<p>When the codebase fits in the context window, it&#x27;s simple. But even if I&#x27;m working on a larger thing, it takes a bit of RAG-alike effort to build knowledge topology, and then it&#x27;s super easy to get docs on (the actual!) architecture, specific components, and all the way down to atomic function level.
评论 #41693983 未加载
lawls8 个月前
I enjoy using the Projects feature of Claude. You have a workspace to add knowledge (file&#x2F;image&#x2F;text) that you are able to update and maintain throughout multiple discussions.
评论 #41685904 未加载
andrei_says_8 个月前
I’ve tried but found it to be a distraction. Instead of focusing on the design and code now I have to focus on getting something out of a wiggling llm algorithm which I need to corner into producing what I need - and then troubleshoot the output.<p>It’s 10x easier and more direct to just do it myself, distilling my mental model into precise instructions.<p>Maybe it works better for languages with more boilerplate- I use Ruby which is quite terse.
imathew8 个月前
I haven&#x27;t liked any integrated tooling at all, but that&#x27;s not such a surprise since I generally disable autocomplete and autocorrect since, like tinsel, I find them distracting.<p>I do really like to use the plain web browser tools, though (currently claude), for generating boilerplate code that I then review and integrate carefully into my code. It has sped up my workflow in that way.
fullstackwife8 个月前
Like many others I&#x27;m working on my own tool [1] that helps me make progress in my hobby projects. An example outcome of using this tool is my js13k game [2] (the idea and majority of code). The tool is being developed using the tool itself (inception ;)).<p>This experience of developing such tool is invaluable, I doubt I would be able to learn so much about LLMs and AI service providers if I was using some off-the-shelf software. It also allows me to adjust it to my own needs, while the &quot;main stream&quot; tools usually try to please &quot;large enterprise customers&quot;, so as a regular user you end up living with annoying bugs&#x2F;quirks, that will never be fixed (because no large customer asks for it).<p>In my opinion the future of such tools is multi modal (both input and output). For my specific use case (hobby gamedev) the goal is to eventually become &quot;the idea guy&quot;, who occasionally jumps into the code to perform some fun programming task, and is not able to create assets(GFX&#x2F;SFX) themselves.<p>For example: when I&#x27;m using my tool, and I want to fix a problem like misaligned buttons, instead of typing an exhaustive prompt, I rather prefer to paste a screenshot, and formulate task with just few words (&quot;align those buttons&quot;).<p>[1] <a href="https:&#x2F;&#x2F;github.com&#x2F;gtanczyk&#x2F;genaicode&#x2F;">https:&#x2F;&#x2F;github.com&#x2F;gtanczyk&#x2F;genaicode&#x2F;</a><p>[2] <a href="https:&#x2F;&#x2F;js13kgames.com&#x2F;2024&#x2F;games&#x2F;monster-steps" rel="nofollow">https:&#x2F;&#x2F;js13kgames.com&#x2F;2024&#x2F;games&#x2F;monster-steps</a>
评论 #41688817 未加载
Lucasoato8 个月前
LLM code-aware chats: continue.dev, with Claude 3.5<p>Autocompletions: GitHub Copilot, still it’s better that Mistral small code models in my opinion
pftburger8 个月前
Vercel’s V0 has been amazing. I lean backend, and the whole mess of styling is a continual struggle. V0 gets it right 90% of the time<p>But like everything you have to spec it well<p>Then any top model basically for duplicating work.<p>“Here is component B, here is component A and test A. Produce a test for B following the same pattern”
mdp20218 个月前
LLMs can be revolutionary for approaching areas you do not know well (&quot;how do you do this with this library, provide short working code&quot;).<p>But you need to use Retrieval Augmented Generation (or bare LLMs will either invent the nonexistent or present solutions to non-proposed problems).
评论 #41690316 未加载
Aeolun8 个月前
I use Github Copilot at work (and at home).<p>I’m a huge fan of keeping things simple by being verbose, and Copilot takes all the drudgery away by filling out all the repeating patterns.<p>Once in a while when I can’t be bothered to figure out Typescript types, I ask ChatGPT to write one for me.
ramon1568 个月前
I only use it to get a sense of a project. Instead of looking for a similar project (and ending up in medium hell), I now ask claude some first steps. I can continue from there. It&#x27;s like helping out a friend get his project doing g
wseqyrku8 个月前
It&#x27;s great for refactorings and text manipulation -- things I used to do regex for can be done using gpt, though that depends on the scope so it is still not quite there to replace regex. Oh did I say it&#x27;s great at bash too?
评论 #41685707 未加载
JohnFen8 个月前
I haven&#x27;t. I tried them out for a good while, but didn&#x27;t find the cost&#x2F;benefit ratio of using them to be to my advantage.
palmfacehn8 个月前
It is fine for expanding CSVs, test data, static arrays and similar. Not comfortable using it as part of an editor or cloud services generally.
npteljes8 个月前
I use it as a personal assistant, and as a replacement of internet search. It has been working great for things like refactoring small parts of code, generating boilerplate or working out what an arcane error message can mean. I find it useful for comparing two technologies, and to have it give me a refresher about some topic.<p>On a personal assistant level, it&#x27;s been useful to have me remind of words that I have forgotten, or to have it rephrase a sentence in a different way, to suit some mood, etc.<p>Occasionally I have fun with it by having it answer in rhymes, or theatrically like a fortune teller, or like lyrics of some gangsta rap.<p>I don&#x27;t know about performance increases. What I notice most is that I&#x27;m less annoyed with it, than with the general state of the internet. The web pages and I have different goals: I want a specific piece of information, and they want me to load all their ads, affiliates and such in return, and to spend time on the page reading their drivel. It is also horrible to search for version-specific information, for example, I appreciate that it brings up the Ruby answer for the latest version, but I&#x27;m stuck on 2.5.0, and so, I need that specifically. LLMs are usually great at this. Orders of magnitude less fluff, rich text-only answer.
Smithalicious8 个月前
Right now I&#x27;m working on my first large clojure codebase, so I do a lot of &quot;what is the idiomatic way to do X?&quot; questioning, pasting in small functions and asking if there&#x27;s a good simplifying refactor etc. Before this I was writing a lot of numpy and doing a lot of math, so it was a lot of &quot;there&#x27;s what I want to calculate, how do I make numpy do that?&quot;.<p>I use Claude now that they&#x27;ve successfully made chatgpt too insufferable to use.<p>I haven&#x27;t tried any of the special AI editors since I don&#x27;t know if I would survive the traumatic injury of having emacs taken away from me.
Euphorbium8 个月前
Writes regexes for me.
Nullabillity8 个月前
I&#x27;d rather fucking die.
评论 #41685916 未加载