TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Is Grok Basically Just an OpenAI Wrapper?

82 点作者 bundie超过 1 年前

11 条评论

minimaxir超过 1 年前
An xAI engineer replied:<p>&gt; The issue here is that the web is full of ChatGPT outputs, so we accidentally picked up some of them when we trained Grok on a large amount of web data. This was a huge surprise to us when we first noticed it. For what it’s worth, the issue is very rare and now that we’re aware of it we’ll make sure that future versions of Grok don’t have this problem. Don’t worry, no OpenAI code was used to make Grok.<p><a href="https:&#x2F;&#x2F;twitter.com&#x2F;ibab_ml&#x2F;status&#x2F;1733558576982155274" rel="nofollow noreferrer">https:&#x2F;&#x2F;twitter.com&#x2F;ibab_ml&#x2F;status&#x2F;1733558576982155274</a>
评论 #38586722 未加载
评论 #38585352 未加载
评论 #38586338 未加载
评论 #38599670 未加载
评论 #38585288 未加载
andy_xor_andrew超过 1 年前
It&#x27;s far, far more likely that Grok was just trained on data that includes text generated by GPT.<p>This is super common in the open-source&#x2F;local AI area. Many models are trained on output from GPT. The better models will filter out anything that mentions GPT or OpenAI. Seems like Grok is not one of the better models.
评论 #38585587 未加载
评论 #38586812 未加载
评论 #38586650 未加载
评论 #38586738 未加载
sschueller超过 1 年前
I believe it is. I&#x27;ve had people send me screenshots where it refers to itself as chatGPT.
评论 #38585779 未加载
ronsor超过 1 年前
This isn&#x27;t really that newsworthy. Grok was probably trained on ChatGPT logs, just like the litany of other chat-oriented LLMs that are also open source. They still should&#x27;ve filtered out the OpenAI canned responses (there are datasets with that filter).
评论 #38585376 未加载
评论 #38585307 未加载
berkes超过 1 年前
What&#x27;s grok? (I dropped off Twitter&#x2F;X the moment the inmates started running that asylum again. It appears to have to do with X?)
评论 #38585139 未加载
评论 #38585324 未加载
评论 #38586933 未加载
评论 #38585133 未加载
LorenDB超过 1 年前
I had self hosted Llama 2 call itself Bard the other day, so I&#x27;d take this with a grain of salt - or maybe even the whole shaker.
alsodumb超过 1 年前
A simple explanation is that Grok probably used GPT3.5 or GPT4 API to generate synthetic data, most likely for RLHF.
hn_throwaway_99超过 1 年前
I feel like comments defending Twitter (I&#x27;m still calling it that until they change their domain) are half missing the point. True, it&#x27;s not necessarily an &quot;OpenAI <i>API</i> wrapper&quot;, but the fact that it was trained on ChatGPT&#x27;s logs basically means it&#x27;s still an &quot;OpenAI wrapper&quot; of sorts, and it&#x27;s going to be inferior in nearly every way (I always thought a huge problem with LLMs going forward was the risk that they would be &quot;contaminated&quot; with non-human training data).<p>I can&#x27;t think of any reason anybody would use Grok over ChatGPT besides political tribe signalling.
评论 #38585556 未加载
评论 #38585308 未加载
评论 #38585222 未加载
评论 #38585910 未加载
simonw超过 1 年前
Grok is a new model. Its training data has seen enough examples of OpenAI generated content that it occasionally spits out text like this that makes it look like it&#x27;s by OpenAI.<p>This is common issue across all sorts of other alternative models too. It&#x27;s not particularly surprising.
评论 #38585221 未加载
qarl超过 1 年前
I wonder why they decided not to filter-out references to OpenAI and ChatGPT.
评论 #38585810 未加载
gafage超过 1 年前
It makes no sense to me that one would train a chatbot on chatgpt conversations and not filter strings that literally say &quot;openai&quot; and &quot;chatgpt&quot;. Extreme incompetence.
评论 #38585331 未加载
评论 #38585235 未加载