TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

GPT-2: 6-Month Follow-Up

160 点作者 xcodevn超过 5 年前

12 条评论

gambler超过 5 年前
<i>&quot;Cornell University is studying human susceptibility to digital disinformation generated by language models.&quot;</i><p><i>&quot;The Middlebury Institute of International Studies Center on Terrorism, Extremism, and Counterterrorism (CTEC) is exploring how GPT-2 could be misused by terrorists and extremists online.&quot;</i><p><i>&quot;The University of Oregon is developing a series of “bias probes” to analyze bias within GPT-2.&quot;</i><p>But apparently no university studies the social and economic impact of using terabytes of public data to train algorithms that for all practical reasons end up being inaccessible to an average person.<p>If things go on the way they&#x27;re going right now, in 20 years millions of people will be &quot;mechanical turked&quot;. Most of information processing tools will be mediated exclusively through companies like Google and Amazon. They will be less like normal tools (e.g. word processors) and more like systems you have to be a part of. Can you imagine the levels of inequality involved? The hyper-centralization of power? <i>This</i> is the foremost challenge presented by AI, not some hypothetical nonsense involving terrorists using a text generator.<p>And it&#x27;s not like there aren&#x27;t any solutions. Douglas Engelbart, for example, pointed out a great way of introducing technology into society without screwing most of the society over:<p><a href="http:&#x2F;&#x2F;dougengelbart.org&#x2F;content&#x2F;view&#x2F;138" rel="nofollow">http:&#x2F;&#x2F;dougengelbart.org&#x2F;content&#x2F;view&#x2F;138</a><p>We kind of followed his vision for a while, with good results, but AI seems to be going in an entirely different direction.
评论 #20752555 未加载
评论 #20754104 未加载
评论 #20753948 未加载
评论 #20754042 未加载
评论 #20752121 未加载
revel超过 5 年前
The OpenAI approach to managing the release of the larger dataset strikes me as totally flawed and upside down. The biggest concern the team seem to have is that the fully trained GPT2 model will be used to spread propaganda and misinformation. They also imply that the biggest hurdle to training a similar model is money needed to pay for the training resources.<p>The problem with this approach is that the users most likely to be malicious users of GPT2 are state actors. China, for example, <i>already</i> spends millions on an immense propaganda factory. Money is not a serious obstacle for a state. Given that other research entities are, by the sound of things, already far along with development of similar models it seems unlikely that China and the US don&#x27;t already have functional models internally.<p>On the other hand, legitimate business and research is clearly hamstrung by withholding the full model. What we have is the maximum degree of inconvenience and the minimum degree of security. It feels almost perfectly analogous to ban on liquids in airports. The motivation for that ban was that existing security measures couldn&#x27;t detect liquids, but simply announcing a ban was to be enforced didn&#x27;t change the fact that liquids were undetectable. Instead millions of travelers were pointlessly inconvenienced at great cost.<p>Release the kraken already!
评论 #20754077 未加载
评论 #20753962 未加载
评论 #20753834 未加载
minimaxir超过 5 年前
For finetuning GPT-2 on custom text, my gpt-2-simple package (<a href="https:&#x2F;&#x2F;github.com&#x2F;minimaxir&#x2F;gpt-2-simple" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;minimaxir&#x2F;gpt-2-simple</a>) gets close to going OOM when finetuning the 345M model, even on a 16GB VRAM server GPU. <i>Doubling</i> the size of the model with the 774M model might cause it to not work at all, so I’ll need to test.<p>Of course, the default output from the model might be sufficient, although it’ll take twice as long to generate text compared to the 345M which is slow even on a GPU.<p>How exactly the large GPT-2 models are deployed is a mystery I really wish was open-sourced more.
评论 #20752765 未加载
评论 #20749230 未加载
评论 #20750852 未加载
评论 #20753883 未加载
The_rationalist超过 5 年前
&lt;rant&gt; Are there any real use case for GPT-2? Does it solve any problem? I&#x27;ve read almost all state of the art leaderboards of all Nlp tasks of paperswithcode.com and truth is except text generation, openAI has not one state of the art, they are not even visible in leaderboards. OpenAI is maybe the AI research center with the biggest funding and comparatively to other well known (Microsoft, Facebook, Google or even zalando..) they are the ones with the least results.<p>From my observations most SOTAs come from chineses researchers by far, followed by deepmind.<p>BTW isn&#x27;t that a sad truth that not even one of all major AI actors has a draft of an AGI architecture, something comparable to CYC or opencog. <a href="https:&#x2F;&#x2F;wiki.opencog.org&#x2F;w&#x2F;CogPrime_Overview" rel="nofollow">https:&#x2F;&#x2F;wiki.opencog.org&#x2F;w&#x2F;CogPrime_Overview</a><p>Two other observations I would like to share: Many important NLP tasks have almost nobody publicly working on them it seems, on paperswithcode.com or NLP-progress (from github) some tasks have only one or two papers... And many others have not evolved since 2016. Most of the time it seems trivial to beat the old state of the art, just use BERT or XLnet on a task where nobody applied it before and hop, free state of the art for you! Yet researchers don&#x27;t seems to chase those low hanging, high returns fruits. Also researchers seems to work a lot in isolation, many new generic improvements like new optimizers (RAdam for example) and new activation functions (Swish) allow to beat most of older state of the art on almost all task just by using them. Yet researchers will take years before using them because of an absurd inertia. Also unlike an open source program, BERT and XLnet have very low response and activity on github despite major open issues... &lt;&#x2F;rant&gt;
评论 #20749860 未加载
评论 #20750902 未加载
评论 #20750233 未加载
评论 #20753346 未加载
评论 #20749680 未加载
gambler超过 5 年前
Hopefully someone will make a working demo of it, like Adam King did for 345M. People should be able to experiment with this stuff without relying on the hype of press releases:<p><a href="https:&#x2F;&#x2F;medium.com&#x2F;@VictorBanev&#x2F;interrogating-gpt-2-345m-aaff8dcc516d" rel="nofollow">https:&#x2F;&#x2F;medium.com&#x2F;@VictorBanev&#x2F;interrogating-gpt-2-345m-aaf...</a><p>Not sure why open AI doesn&#x27;t do this themselves. That fully aligns with their stated mission.
评论 #20751512 未加载
评论 #20750298 未加载
评论 #20751136 未加载
评论 #20756945 未加载
zitterbewegung超过 5 年前
I was able to take all of Donald Trumps tweets and using GPT2 to make a program that would mimic his tweets.<p>I found that it might be very effective. I have the test at<p><a href="https:&#x2F;&#x2F;docs.google.com&#x2F;forms&#x2F;d&#x2F;1p7tlobl5y5plBCu_enK4KawR7B8_4Yyb-wCUh6vr9A0&#x2F;edit" rel="nofollow">https:&#x2F;&#x2F;docs.google.com&#x2F;forms&#x2F;d&#x2F;1p7tlobl5y5plBCu_enK4KawR7B8...</a><p>I got the information from trumptwitterarchive.com<p>I also explored creating a system that could recognize fake tweets from real ones and I believe I got 94% accuracy. It was a Bayes classifier but I think I have to double check my work.
评论 #20751068 未加载
lucidrains超过 5 年前
Hmm, no mention of Megatron in their timeline? <a href="https:&#x2F;&#x2F;nv-adlr.github.io&#x2F;MegatronLM" rel="nofollow">https:&#x2F;&#x2F;nv-adlr.github.io&#x2F;MegatronLM</a>
评论 #20753107 未加载
rovyko超过 5 年前
&gt;As part of our staged release strategy, our current plan is to release the 1558M parameter model in a few months, but it’s plausible that findings from a partner, or malicious usage of our 774M model, could change this.<p>This seems naive but I think it&#x27;s a misdirection. Of course the model will have malicious users. Propaganda teams started testing its integration as soon as it was released. It&#x27;s likely that OpenAI is counting on this for insights into HOW the model can be used maliciously. It&#x27;s also possible that the model results have inherent trackable markers and OpenAI can later say that X% of social media posts were made using this model.<p>So what are the positive applications, aside from prettifying data like sports and weather reports?<p>Even with Skyrim&#x27;s 800+ books, you frequently ran into the same book. Imagine libraries filled with plausible text that hides nuggets of lore seeded by developers. Along with more realistic text-to-speech this can allow games to support a large diversity of NPCs that have true radiant dialogue and sound more realistic than &quot;I saw a mudcrab the other day&quot;.<p>With some modifications, I think models like this can outweigh even their nefarious applications:<p>Defense against text decomposition analysis. The model can be used to obfuscate writing patterns that can reveal a person&#x27;s identity, either by randomizing form or standardizing it. Take your post and run it through the formatter to get the same idea and intent, but in a style that can&#x27;t be traced to your other writing. Or you reform it into style of Ernest Hemmingway, like thousands of others.<p>Realtime plausible deniability encryption. Messages in a monitored chat can look like mundane conversation but contain encrypted messages. This would require the model accept seeds and work partially in reverse to diff two sets of text to reveal the hidden message.<p>In it&#x27;s current form it doesn&#x27;t look like it can do any of those things, but there&#x27;s the potential.
baalimago超过 5 年前
Even if GPT-2 were released, very very few would have the hardware to run it because of gpu ram running out (and doing some sort of load-unload system would make training times unfeasibly long). And those who have the hardware to run it, has probably already made a version of their own or reasons not to. So I&#x27;m wondering if this GPT-2 hype is a genuine concern of openai, or if it&#x27;s mostly a PR flex to say &#x27;Look at us, we made a good model!&#x27;.<p>As an example, look here by Nvidia <a href="https:&#x2F;&#x2F;devblogs.nvidia.com&#x2F;training-bert-with-gpus&#x2F;" rel="nofollow">https:&#x2F;&#x2F;devblogs.nvidia.com&#x2F;training-bert-with-gpus&#x2F;</a> who made GPT-2 8B, which is ~5 times as large as GPT-2.
wyldfire超过 5 年前
Are there any applications for the GPT-2 models beyond text synthesis? Inference, question-answering, NER detection&#x2F;disambiguation, anything like this?
评论 #20749275 未加载
评论 #20749243 未加载
评论 #20751168 未加载
brentsch超过 5 年前
I&#x27;m curious about the &quot;fine-tuning based detection&quot; mentioned in the report (&quot;Fine-tunes a language model to &#x27;detect itself&#x27;... over a range of available settings&quot;). Does anyone know good articles&#x2F;papers (or have an off-the-top tl;dr) to get a high-level grasp of &quot;self-detection&quot; for generative models?
评论 #20753653 未加载
lxe超过 5 年前
Anyone wired a &quot;talktotransformer&quot;-style system to this one yet? Would like to see how it works without going through the steps of setting it up.<p>EDIT: Looks like <a href="https:&#x2F;&#x2F;talktotransformer.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;talktotransformer.com&#x2F;</a> already uses the 774M one!
评论 #20752488 未加载