TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

OpenAI Releases Largest GPT-2 Text Generation Model

266 pointsby epoch_100over 5 years ago

15 comments

vicayaover 5 years ago
Just like various chat bot incarnations, it&#x27;s reasonably OK at generating generic sentences without real understanding. I tried several attempts with this leading fragment: &quot;Is borrow checking unique to Rust? Probably not. But&quot;. The results are not pretty (the best one is as follows). Not impressed at all with this giant model :|<p>&quot;Is borrow checking unique to Rust? Probably not. But the Rust library is written by and for Rust programmers, so if you&#x27;re looking for a library that&#x27;s designed for the language you already know, this might be your first choice.<p>The library is still in a fairly early stage. It does not support the standard libraries.<p>If you&#x27;re interested in reading about how the library works, the docs are up on GitHub for reference!<p>How can I get this for my project?<p>Get it here: <a href="https:&#x2F;&#x2F;github.com&#x2F;paulp&#x2F;rustbinder" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;paulp&#x2F;rustbinder</a><p>I&#x27;ve just read your blog post and want to use it, what can I do?<p>If you&#x27;re looking for a project to use this library for, check out the docs for the &quot;RustBinder project&quot;.<p>What are the current limitations and how will future changes affect this library?<p>This is still a work in progress. The library currently does not support the standard Rust library. There are a few work in progress&quot;
评论 #21460825 未加载
评论 #21461298 未加载
评论 #21459956 未加载
评论 #21460615 未加载
评论 #21461096 未加载
评论 #21461856 未加载
rm_-rf_slashover 5 years ago
At a credibility score of 6.91&#x2F;10, many people will rightly judge that the full GPT-2 model will remain insufficient for malicious use in creating fake news.<p>However, even the smaller models are already good enough for spamming&#x2F;trolling&#x2F;astroturfing. It doesn’t take a Shakespearean soliloquy to convince people of a point. Just enough of a flood of short 1-3 sentence pro&#x2F;con comments on a forum can drastically affect the perceived public opinion of an issue. Those comments can then spur real people to reply, which could result in an ultimately organic but directed propaganda vector. Propaganda directors will carefully craft something for people to look at, and the GPT-2 bots will move people’s eyes in that direction.<p>You can see the same happen on r&#x2F;subsimulatorgpt2, where the longer titles and prompts and replies eventually sprawl into incoherence, but the shorter sentences from the finetuned bots in the comments section are effectively indistinguishable from the kinds of short comments you would find on their respective subreddits.<p>Or in other words, the malicious uses for GPT-2 won’t be a tidal wave, but a flash flood.
评论 #21456345 未加载
评论 #21455213 未加载
krickover 5 years ago
Wow, some samples are frighteningly good. I was impressed by previous models and I don&#x27;t know if I&#x27;m just lucky this time, but... wow. Can anybody who is not into climbing even tell this is all fake?<p><i>Jain Kim is an experienced climber.</i><p>In 2006, she became the first woman from Korea to climb all five 8,000 meters (24,064 ft) peaks in the Swiss alpine ski run Alps in 24 hours. In 2009, she made history again by setting the record for the fastest time to climb an 8,000 meter peak with a team from China and South Korea.<p>She made the first ascent of 8,832-meter K2 in China, the second highest mountain in the world, in 2009 and the third highest mountain in Europe. She also is the first female Korean to summit a world-class peak.<p>During her two years as a mountaineering professor at Sogang University in Korea, she established two new routes in the Yalu River area. The first of these routes is a 3,547-meter peak named K2 on Mount Long in China. Her second route is on the same mountain, called the Lomonosov Ridge, at 3,632 meters.
评论 #21458862 未加载
评论 #21460804 未加载
评论 #21458751 未加载
评论 #21458804 未加载
评论 #21458772 未加载
评论 #21459800 未加载
clmntover 5 years ago
We (Hugging Face) added it to Write With Transformers if you want to try the text generation capabilities of the model: <a href="https:&#x2F;&#x2F;transformer.huggingface.co&#x2F;doc&#x2F;gpt2-xl" rel="nofollow">https:&#x2F;&#x2F;transformer.huggingface.co&#x2F;doc&#x2F;gpt2-xl</a>
评论 #21461354 未加载
epoch_100over 5 years ago
Paper: <a href="https:&#x2F;&#x2F;d4mucfpksywv.cloudfront.net&#x2F;papers&#x2F;GPT_2_Report.pdf" rel="nofollow">https:&#x2F;&#x2F;d4mucfpksywv.cloudfront.net&#x2F;papers&#x2F;GPT_2_Report.pdf</a><p>Code: <a href="https:&#x2F;&#x2F;github.com&#x2F;openai&#x2F;gpt-2" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;openai&#x2F;gpt-2</a>
评论 #21456025 未加载
rfhjtover 5 years ago
Prompt: &quot;Real things don&#x27;t exist unconditionally and things that exist unconditionally are not real. However the reality has an essense. It is&quot;<p>Response: &quot;an actual thing, and it is not the thing to which we attach meaning. It is not real because it is not a thing. And therefore, it does not possess the qualities that are inherent in all real things.&quot;<p>Just wow. Sure, there are a few logical mistakes here, but this response serves as a good prompt for my bio-GPT. In other words, we usually need some starting points or hints for analysis and discovering these hints is non trivial because whatever we can think of is not very new to us. This GPT just gave me an answer that smells like a serious wisdom and I&#x27;ll surely dig in that direction to see if this idea has any substance.<p>Edit: what&#x27;s happening here is while I can&#x27;t ask this model to give me a short and conscise summary on a topic, I can still interrogate this model and find out what it&#x27;s seen in the training set. I can&#x27;t possibly read all these books in the training set, but now I can rapidly navigate in the multidimensional meaning space: I tell it where to start and it says what it seems in close proximity to my prompt. This is a breakthru.
评论 #21462159 未加载
hint23over 5 years ago
You can try it at: <a href="http:&#x2F;&#x2F;textsynth.org" rel="nofollow">http:&#x2F;&#x2F;textsynth.org</a>
评论 #21459701 未加载
评论 #21457761 未加载
评论 #21456705 未加载
评论 #21456925 未加载
评论 #21456766 未加载
评论 #21457313 未加载
评论 #21461337 未加载
评论 #21457329 未加载
评论 #21459164 未加载
评论 #21459589 未加载
评论 #21459449 未加载
buboardover 5 years ago
&gt; (CTEC) found that extremist groups can use GPT-2 for misuse, specifically by fine-tuning GPT-2 models on four ideological positions: white supremacy, Marxism, jihadist Islamism, and anarchism. CTEC demonstrated that it’s possible to create models that can generate synthetic propaganda for these ideologies<p>I wonder how they tested that
评论 #21456056 未加载
评论 #21457924 未加载
评论 #21455654 未加载
评论 #21461057 未加载
chaz6over 5 years ago
Surely we are not far off models capable of submission-quality essays that will enable a new generation of cheating.
评论 #21455909 未加载
评论 #21456215 未加载
评论 #21456180 未加载
评论 #21458516 未加载
评论 #21457993 未加载
评论 #21461824 未加载
rfhjtover 5 years ago
Prompt: The coming global recession is a real possibility and&quot;<p>Response: &quot;The coming global recession is a real possibility and the Fed is playing games, creating artificial market conditions to make a recovery seem possible in the short-term. The Fed has an option to change its monetary policies but it will not make the problem go away, so it is in their best interest to pretend it won&#x27;t happen.&quot;<p>Change and to however and you&#x27;ll get another stereotype opinion. It really just composes pieces of texts it&#x27;s seen around the prompt, but it does this really well.<p>Most of the news agencies can now fire most of their monkey typewriters: this GPT will outperform them on every metric.
k8siover 5 years ago
Omfg can we stop making these things bigger PLEASE<p>Like, who cares??<p>* What I mean is, text gen models are big enough. We need controllable text generation; like, so it can talk about a specific THING sensibly. Rather than spew statistically plausible nonsense.
oaskmutboardover 5 years ago
I think this could make a great Tinder feature to suggest chat lines.
评论 #21458393 未加载
评论 #21460714 未加载
gerashover 5 years ago
Sampling realistic text from large pretrained models is non-trivial. I came across this paper in one of ACL 2019 workshops:<p><a href="https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1904.09751.pdf" rel="nofollow">https:&#x2F;&#x2F;arxiv.org&#x2F;pdf&#x2F;1904.09751.pdf</a>
ionwakeover 5 years ago
Sorry for asking but is there an example output and an example input?
评论 #21456012 未加载
评论 #21455823 未加载
评论 #21455478 未加载
评论 #21455039 未加载
评论 #21455327 未加载
490d0aff0ee8over 5 years ago
Tangent rant.<p>I&#x27;m skimming over some of the code at <a href="https:&#x2F;&#x2F;github.com&#x2F;openai&#x2F;gpt-2&#x2F;blob&#x2F;master&#x2F;src&#x2F;model.py" rel="nofollow">https:&#x2F;&#x2F;github.com&#x2F;openai&#x2F;gpt-2&#x2F;blob&#x2F;master&#x2F;src&#x2F;model.py</a> and I can&#x27;t help but feel frustrated at how unreadable this stuff is.<p>1. Why is it acceptable to have single-letter variable names everywhere?<p>2. There&#x27;s little to almost no documentation in the code itself. It&#x27;s unclear what the parameters of any given function mean.<p>3. There are magic constants everywhere.<p>4. Function names are so terse... (&quot;gelu&quot;, &quot;attn&quot;)
评论 #21457038 未加载
评论 #21456400 未加载
评论 #21456791 未加载
评论 #21456605 未加载
评论 #21456798 未加载
评论 #21456385 未加载
评论 #21456426 未加载
评论 #21456793 未加载
评论 #21456391 未加载
评论 #21457203 未加载
评论 #21456622 未加载