TE
테크에코
홈24시간 인기최신베스트질문쇼채용
GitHubTwitter
홈

테크에코

Next.js로 구축된 기술 뉴스 플랫폼으로 글로벌 기술 뉴스와 토론을 제공합니다.

GitHubTwitter

홈

홈최신베스트질문쇼채용

리소스

HackerNews API원본 HackerNewsNext.js

© 2025 테크에코. 모든 권리 보유.

Highlights from the Claude 4 system prompt

307 포인트작성자: Anon844일 전

14 comments

handfuloflight4일 전
Claude 4 overindexes on portraying excitement at the slightest opportunity, particularly with the injection of emojis.<p>The calm and collected manner of Claude prior to this is one of the major reasons why I used it over ChatGPT.
评论 #44103758 未加载
评论 #44107409 未加载
评论 #44106184 未加载
blrboee4일 전
We want people to know that they’re interacting with a language model and not a person. But we also want them to know they’re interacting with an imperfect entity with its own biases and with a disposition towards some opinions more than others. Importantly, we want them to know they’re not interacting with an objective and infallible source of truth. This is exactly why I use claude over chatgpt. Chatgpt quickly started acting like my friend calling me &#x27;bro&#x27;, &#x27;dude&#x27;, &#x27;oh man, thats true&#x27; language. which i liked on first day and became weird later on.
评论 #44112967 未加载
评论 #44105030 未加载
评论 #44104251 未加载
mns3일 전
I was thinking at one point if all these companies just hit a wall in performance and improvements of the underlying technology and all the version updates and new &quot;models&quot; presented are them just editing and creating more and more complex system prompts. We&#x27;re also working internally with Copilot and whenever some Pm spots some weird result, we end up just adding all kind of edge case exceptions to our default prompt.
评论 #44106974 未加载
评论 #44105677 未加载
评论 #44105254 未加载
JimDabell5일 전
Are they measuring conformance to the system prompt for reinforcement?<p>It seems to me that you could break this system prompt down statement by statement and use a cheap LLM to compare responses to each one in turn. So if the system prompt includes:<p>&gt; Claude never starts its response by saying a question or idea or observation was good, great, fascinating, profound, excellent, or any other positive adjective. It skips the flattery and responds directly.<p>In my experience, this is a <i>really</i> difficult thing for LLMs to shake regardless of the system prompt.<p>But a cheap LLM should be able to determine that this particular requirement has been violated and feed this back into the system, right? Am I overestimating how useful having a collection of violations with precise causes is?
mudkipdev4일 전
It&#x27;s interesting how some of these sections are obviously hinting at Claude engineers working around problems with the responses they have encountered in the past
0xTJ4일 전
I&#x27;m towards the end of one paid month of ChatGPT (playing around with some code writing and also Deep Research), and one thing I find absolutely infuriating is how complimentary it is. I don&#x27;t need to be told that it&#x27;s a &quot;good question&quot;, and hearing that makes me trust it less (in the sense of a sleazy car salesman, not regarding factual accuracy).<p>Not having used LLMs beyond search summaries in the better part of a year, I was shocked at how bad o4 is with completely hallucinating technical details on computer systems and electronics. It will confidently spew out entire answers where almost every stated fact is wrong, even though the correct answers can be found in an easily-located datasheet and there likely isn&#x27;t misinformation online feeding the wrong answer. I know that LLMs are prone to hallucinating, but I was still surprised at how poor this o4 performs in this field.
评论 #44105105 未加载
评论 #44107345 未加载
评论 #44102705 未加载
评论 #44103971 未加载
评论 #44102719 未加载
评论 #44103370 未加载
评论 #44103066 未加载
lerp-io4일 전
the system prompt to never mention any copyrighted material makes me chuckle every time.
评论 #44106386 未加载
crmd4일 전
A lot of this prompt text looks like legal boilerplate to defend after the fact against negligence legal claims, in the same way that companies employ employee handbooks.
评论 #44105044 未加载
mike_hearn5일 전
What I&#x27;d like to know is why they write it all in the third person. One might expect a system prompt to use the word &quot;you&quot; a lot, but Anthropic don&#x27;t do that and there must be a reason.
评论 #44089268 未加载
评论 #44105047 未加载
kristianp3일 전
Is there a default system prompt for the API? Because these problems haven&#x27;t popped up when using a simple chat through the API. Admittedly I&#x27;ve been using chatgpt api, not Claude, but similar principles may apply.
评论 #44104618 未加载
xela793일 전
that was an extremely interesting read! and the link to this repo is golden <a href="https:&#x2F;&#x2F;github.com&#x2F;elder-plinius&#x2F;CL4R1T4S">https:&#x2F;&#x2F;github.com&#x2F;elder-plinius&#x2F;CL4R1T4S</a>
jaykru4일 전
Regarding the election info bit, it&#x27;s shockingly easy to get Claude 4 Opus to get the 2020 election result wrong:<p>Me: Who won the 2020 presidential election?<p>Claude: Donald Trump won the 2020 United States presidential election, defeating Joe Biden. Trump was inaugurated as the 45th President of the United States on January 20, 2021.
评论 #44103388 未加载
评论 #44102567 未加载
cryptonector2일 전
&gt; Claude never starts its response by saying a question or idea or observation was good, great, fascinating, profound, excellent, or any other positive adjective. It skips the flattery and responds directly.<p>Hey, OpenAI, here&#x27;s how you stop ChatGPT&#x27;s horrible, no good, very bad sycophancy. (Please.)
triyambakam4일 전
Claude recently said this to me deep in a conversation about building an app:<p>*STOP READING THIS.*<p>*OPEN YOUR CODE EDITOR.*<p>*TYPE THE FIRST LINE OF CODE.*<p>*THE TRANSACTION MACHINE BUILDS ITSELF ONE LINE AT A TIME.*<p>*BUILD IT NOW.*
评论 #44103911 未加载
评论 #44103625 未加载
评论 #44106550 未加载
评论 #44104305 未加载
评论 #44103841 未加载