TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Imagen Video: high definition video generation with diffusion models

800 pointsby jasondaviesover 2 years ago

63 comments

i_like_apisover 2 years ago
The concern trolling and gatekeeping about social justice issues coming from the so-called &quot;ethicists&quot; in the AI peanut gallery has been utterly ridiculous. Google claims they don&#x27;t want to release Imagen because it lacks what can only be called &quot;latent space affirmative action&quot;.<p>Stability or someone like it will valiantly release this technology, <i>again</i> and there will be absolutely no harm to anyone.<p>Stop being so totally silly Google, OpenAI, et. al. - it&#x27;s especially disingenuous because the real reason you don&#x27;t want to release these things is that you can&#x27;t be bothered to share and would rather keep&#x2F;monetize the IP. Which is ok -- but at least be honest.
评论 #33102280 未加载
评论 #33104950 未加载
评论 #33102377 未加载
评论 #33103459 未加载
评论 #33105546 未加载
评论 #33106794 未加载
评论 #33104483 未加载
评论 #33102583 未加载
评论 #33104180 未加载
评论 #33102456 未加载
评论 #33104519 未加载
评论 #33106944 未加载
评论 #33104221 未加载
评论 #33102794 未加载
评论 #33105405 未加载
评论 #33104339 未加载
评论 #33103683 未加载
fzysingularityover 2 years ago
What&#x27;s next? Dreamfusion Video = Imagen Video (this) + Dreamfusion (<a href="https:&#x2F;&#x2F;dreamfusion3d.github.io&#x2F;" rel="nofollow">https:&#x2F;&#x2F;dreamfusion3d.github.io&#x2F;</a>)<p>Fundamentally, I think we have all the pieces based on this work and Dreamfusion to make it work. From the looks of it, there&#x27;s a lot of SSR (spatial SR) and TSR (temporal SR) going on at multiple levels to upsample (spatially) and smoothen (temporally) images that won&#x27;t be needed for NERFs.<p>What&#x27;s impressive is the ability to leverage billion-scale image-text pairs for training a base model that can be used to super-resolve over space and time. And that they&#x27;re not wastefully training video models from scratch, and instead separately training TSR, SSR models for turning the diffused images to video.
评论 #33103128 未加载
评论 #33104784 未加载
BoppreHover 2 years ago
It&#x27;s interesting that these models can generate seemingly anything, but the prompt is taken only as a vague suggestion.<p>From the first 15 examples shown to me, only one contained all elements of the prompt, and it was one of the simplest (&quot;an astronaut riding a horse&quot;, versus e.g. &quot;a glass ball falling in water&quot; where it&#x27;s clear it was a water droplet falling and not a glass ball).<p>We&#x27;re seeing leaps in random capabilities (motion! 3D! inpainting! voice editing!), so I wonder if complete prompt accuracy is 3 months or 3 years away. But I wouldn&#x27;t bet on any longer than that.
评论 #33100345 未加载
评论 #33102309 未加载
nailloover 2 years ago
Probably only 6 months until we get this in stable diffusion format. Things are about to get nuts and awesome.
评论 #33098919 未加载
评论 #33099285 未加载
评论 #33098869 未加载
seanwilsonover 2 years ago
Can anyone comment on how advanced <a href="https:&#x2F;&#x2F;phenaki.video&#x2F;index.html" rel="nofollow">https:&#x2F;&#x2F;phenaki.video&#x2F;index.html</a> is? They have an example at the bottom of a 2 minute long video generated from a series of prompts (i.e. a story) which seems more advanced than Google or Meta&#x27;s recent examples? It didn&#x27;t get many comments on HN when it was posted.
评论 #33101080 未加载
评论 #33105852 未加载
azinman2over 2 years ago
&gt; However, there are several important safety and ethical challenges remaining. Imagen Video and its frozen T5-XXL text encoder were trained on problematic data. While our internal testing suggest much of explicit and violent content can be filtered out, there still exists social biases and stereotypes which are challenging to detect and filter. We have decided not to release the Imagen Video model or its source code until these concerns are mitigated.<p>The concerns cannot be mitigated. The cat&#x27;s out of the bag. Russia has already used poor quality deep fakes in Ukraine to justify their war. This will only become bigger and bigger of an issue to the point where &#x27;truth&#x27; is gone, nothing is trusted, and societies will continue to commit atrocities under false pretense.
评论 #33106032 未加载
评论 #33103928 未加载
评论 #33104018 未加载
评论 #33105813 未加载
评论 #33103379 未加载
mkaicover 2 years ago
And there you have it. As an aspiring filmmaker and an AI researcher, I&#x27;m going to relish the next decade or so where my talents are still relevant. We&#x27;re entering the golden age of art, where the AIs are just good enough to be used as tools to create more and more creative things, but not good enough yet to fully replace the artist. I&#x27;m excited for the golden age, and uncertain about what comes after it&#x27;s over, but regardless of what the future holds I&#x27;m gonna focus on making great art here and now, because that&#x27;s what makes me happy!
评论 #33100620 未加载
评论 #33102126 未加载
dagmxover 2 years ago
I’ll be honest, as someone who worked in the film industry for a decade, this thread is depressing.<p>It’s not the technology, it’s all the people in these comments who have never worked in the industry clamouring for its demise.<p>One could brush it off as tech heads being over exuberant, but it’s the lack of understanding of how much fine control goes into each and every shot of a film that is depressing.<p>If I, as a creative, made a statement that security or programming is easy while pointing to GitHub Copilot, these same people would get defensive about it because they’d see where the deficiencies are.<p>However because they’re so distanced from the creative process, they don’t see how big a jump it is from where this or stage diffusion is to where even a medium or high tier artist are.<p>You don’t see how much choice goes into each stroke, or wrinkle fold , how much choice goes into subtle movements. More importantly you don’t see the iterations or emotional storytelling choices even in a character drawing or pose. You don’t see the combined decades, even centuries of experience, that go into making the shot and then seeing where you can make it better based on intangibles<p>So yeah this technology is cool, but I think people saying this will disrupt industries with vigour need to immerse themselves first before they comment as outsiders.
评论 #33100554 未加载
评论 #33100710 未加载
评论 #33101480 未加载
评论 #33100761 未加载
评论 #33104155 未加载
评论 #33100956 未加载
评论 #33103154 未加载
评论 #33104454 未加载
评论 #33138253 未加载
fassssstover 2 years ago
How long until the AI just generates the entire frame buffer on a device? Then you don’t need to design or program anything; the AI just handles all input and output dynamically.
评论 #33102382 未加载
评论 #33102793 未加载
评论 #33102457 未加载
评论 #33102197 未加载
alphabettingover 2 years ago
We&#x27;re about a week into text-to-video models and they&#x27;re already this impressive. Insane to imagine what the future holds in this space.
评论 #33101403 未加载
评论 #33098923 未加载
评论 #33100695 未加载
throwaway23597over 2 years ago
Google continues to blow my mind with these models, but I think their ethics strategy is totally misguided and will result in them failing to capture this market. The original Google Search gave similarly never-before-seen capabilities to people, and you could use it for good or bad - Google did not seem to have any ethical concerns around, for example, letting children use their product and come across NSFW content (as a kid who grew up with Google you can trust me on this).<p>But now with these models they have such a ridiculously heavy handed approach to the ethics and morals. You can&#x27;t type any prompt that&#x27;s &quot;unsafe&quot;, you can&#x27;t generate images of people, there are so many stupid limitations that the product is practically useless other than niche scenarios, because Google thinks it knows better than you and needs to control what you are allowed to use the tech for.<p>Meanwhile other open source models like Stable Diffusion have no such restrictions and are already publicly available. I&#x27;d expect this pattern to continue under Google&#x27;s current ideological leadership - Google comes up with innovative revolutionary model, nobody gets to use it because &quot;safety&quot;, and then some scrappy startup comes along, copies the tech, and eats Google&#x27;s lunch.<p>Google: stop being such a scared, risk averse company. Release the model to the public, and change the world once more. You&#x27;re never going to revolutionize anything if you continue to cower behind &quot;safety&quot; and your heavy handed moralizing.
评论 #33100449 未加载
评论 #33100480 未加载
评论 #33100543 未加载
评论 #33100988 未加载
评论 #33100270 未加载
评论 #33100643 未加载
评论 #33101737 未加载
评论 #33100482 未加载
评论 #33100283 未加载
评论 #33101054 未加载
评论 #33101059 未加载
评论 #33100392 未加载
评论 #33101597 未加载
评论 #33100491 未加载
评论 #33100161 未加载
评论 #33100685 未加载
评论 #33100461 未加载
评论 #33101396 未加载
评论 #33102487 未加载
evougaover 2 years ago
&gt; We train our models on a combination of an internal dataset consisting of 14 million video-text pairs<p>The paper is sorely lacking evaluation; one thing I&#x27;d like to see for instance (any time a generative model is trained on such a vast corpus of data) is a baseline comparison to nearest-neighbor retrieval from the training data set.
评论 #33102548 未加载
bringkingover 2 years ago
If anyone wants to know what looking at an Animal or some objects on LSD is like, this is very close. It&#x27;s like 95% understandable, but that last 5% really odd.
评论 #33104791 未加载
评论 #33099173 未加载
kranke155over 2 years ago
I’m going to post an Ask HN about what am I supposed to do when I’m “disrupted”. I work in film &#x2F; video &#x2F; CG where the bread and butter is short form advertising for Youtube, Instagram and TV.<p>It’s painfully obvious that in 1 year the job might be exceedingly more difficult than it is now.
评论 #33099314 未加载
评论 #33099817 未加载
评论 #33099691 未加载
评论 #33099797 未加载
评论 #33099070 未加载
评论 #33100808 未加载
评论 #33100462 未加载
评论 #33099891 未加载
评论 #33099463 未加载
评论 #33101803 未加载
评论 #33099356 未加载
评论 #33099909 未加载
评论 #33101368 未加载
评论 #33100893 未加载
评论 #33099161 未加载
评论 #33102404 未加载
评论 #33101093 未加载
评论 #33099814 未加载
brapover 2 years ago
What really fascinates me here is the movement of animals.<p>There&#x27;s this one video of a cat and a dog, and the model was really able to capture the way that they move, their body language, their mood and personality even.<p>Somehow this model, which is really just a series of zeroes and ones, encodes &quot;cat&quot; and &quot;dog&quot; so well that it almost feels like you&#x27;re looking at a real, living organism.<p>What if instead of images and videos they make the output interactive? So you can send prompts like &quot;pet the cat&quot; and &quot;throw the dog a ball&quot;? Or maybe talk to it instead?<p>What if this tech gets so good, that eventually you could interact with a &quot;person&quot; that&#x27;s indistinguishable from the real thing?<p>The path to AGI is probably very different than generating videos. But I wonder...
hazrmardover 2 years ago
The progress of content generation is disorienting! I remember studying Markov Chains and Hidden Markov Models for text generation. Then we had Recurrent Networks which went from LSTMs to Transformers now. At this point we can have a sustained pseudo conversation with a model, which will do trivial tasks for us from a text corpus.<p>Separately for images we had convolutional networks and Generative Adversarial Networks. Now diffusion models are apparently doing what Transformers did to natural language processing.<p>In my field, we use shallower feed-forward networks for control using low-dimensional sensor data (for speed &amp; interpretability). Physical constraints (and good-enoughness of classical approaches) make such massive leaps in performance rarer events.
aero-glide2over 2 years ago
&quot;We have decided not to release the Imagen Video model or its source code until these concerns are mitigated&quot; Okay then why even post it in the first place? What exactly is Google going to do with this model?
评论 #33099993 未加载
评论 #33099712 未加载
评论 #33099138 未加载
评论 #33099079 未加载
评论 #33098976 未加载
评论 #33099450 未加载
评论 #33099119 未加载
评论 #33099166 未加载
评论 #33100698 未加载
评论 #33101626 未加载
Apoxover 2 years ago
I feel like in a not so far future, all this will be generalized into &quot;generate new from all the existing&quot;.<p>And at some point later, &quot;all the existing&quot; will be corrupted by the integrated &quot;new&quot; at it will all be chaos.<p>I&#x27;m joking, it will be fun all along. :)
评论 #33099809 未加载
评论 #33100049 未加载
评论 #33100640 未加载
bravuraover 2 years ago
I agree with many of the arguments in this thread: that model-gatekeeping while publishing approaches seems insincere and just seems like it&#x27;s daring bad actors to replicate.<p>However, a common refrain is that AI is like tools like hammers or knives and can be used for good or misused for evil. The potential for weaponizing AI is much much more so than a hammer or a knife. And it&#x27;s greater than 3D-printing (of guns), maybe even greater than compilers. I would hazard to say it&#x27;s maybe in the same ballpark as chemical weapons and perhaps less so than nuclear weapons and biological weapons, but this is speculative. Nonetheless, I think these otherwise great arguments are diminished by comparing AI&#x27;s safety to single-target tools like hammers or knives.
评论 #33105646 未加载
评论 #33105958 未加载
tobrover 2 years ago
I recently watched Light &amp; Magic, which among other things told the story of how difficult it was for many pioneers in special effects when the industry shifted from practical to digital in the span of a few years. It looks to me like a similar shift is about to happen again.
impalallamaover 2 years ago
All this stuff makes me incredibly anxious about the future of art and artists. It can already very difficult to make a living and tons of artists are horrifically exploited by content mills and vfx shops and stuff like this is just going to devalue their work even more
评论 #33101589 未加载
joshcryerover 2 years ago
Pre-singularity is really cool. Whole world generation in what, 5 years?
user-over 2 years ago
This sort of AI related work seems to be accelerating at an insane speed recently.<p>I remember being super impressed by AI Dungeon and now in the span of a few months we have got DALLE-2 , Stable Diffussion, Imagen, that one AI powered video editor, etc.<p>Where do we think we will be at in 5 years??
评论 #33102020 未加载
评论 #33103191 未加载
StevenNunezover 2 years ago
What a time to be alive!<p>What will this do to art? I&#x27;m hoping we bring more unique experiences to life.
ugh123over 2 years ago
These are baby steps towards what I think will be the eventual &quot;disruption&quot; to the film and tv industry. Directors will simply be able to write a script&#x2F;prompt long enough and detailed enough for something like Imagen (or it&#x27;s successors) to convert into a feature-length show.<p>Certainly we&#x27;re very, very far away from that level of cinematic detail and crispness. But I believe that is where this leads... complete with AI actors (or real ones deep faked throughout the show).<p>For a while I thought &quot;The Volume&quot; was going to be the disruption to the industry. Now I think AI like this will eventually take it over.<p><a href="https:&#x2F;&#x2F;www.comingsoon.net&#x2F;movies&#x2F;features&#x2F;1225599-the-volume-star-wars-revolutionary" rel="nofollow">https:&#x2F;&#x2F;www.comingsoon.net&#x2F;movies&#x2F;features&#x2F;1225599-the-volum...</a><p>The main motivation will be production costs and time for studios, of which The Volume is already showing huge gains for Disney&#x2F;ILM (just look at how much new star wars content has popped up within a matter of a few years). But i&#x27;m unsure if Disney has patented this tech and workflow and if other studios will be able to leverage it.<p>Regardless, AI&#x2F;software will eat the world, and this will be one more step towards it. Exciting stuff.
评论 #33100313 未加载
评论 #33099954 未加载
评论 #33101977 未加载
评论 #33100052 未加载
评论 #33099957 未加载
评论 #33099981 未加载
评论 #33100136 未加载
nigrioidover 2 years ago
There is something deeply unsettling about all text generated by these models.
monologicalover 2 years ago
What everyone is missing is that these AI image&#x2F;video generators lack _taste_. These tools just regurgitate a mishmash of images from it&#x27;s training set, without any &quot;feeling&quot;. What you&#x27;re going to tell me that you can train them to have feeling? It&#x27;s never going to happen.
评论 #33099366 未加载
评论 #33099599 未加载
评论 #33099247 未加载
评论 #33099406 未加载
评论 #33099640 未加载
评论 #33103196 未加载
评论 #33101455 未加载
评论 #33102039 未加载
m3kw9over 2 years ago
Would be useful for gaming environments, where if you look very far away it doesn’t really matter about details
jupp0rover 2 years ago
What&#x27;s the business value of publishing this research in the first place vs keeping it private? Following this train of thought will lead you to the answer to your implied question.<p>Apart from that - they publish the paper and anybody can reimplement and train the same model. It&#x27;s not trivial but it&#x27;s also completely feasible to do for lots of hobbyists in the field in a matter of a few days. Google doesn&#x27;t need to publish a free use trained model themselves and associate that with their brand.<p>That being said, I agree with you, the &quot;ethics&quot; of imposing trivially bypassable restrictions on these models is silly. Ethics should be applied to what people use these models for.
martythemaniakover 2 years ago
I am finally going to be able to bring my 2004-era movie script to life! &quot;Rosenberg and Goldstein go to Hot Dog Heaven&quot; is about the parallel night Harold and Kumar&#x27;s friends had and how they ended up at Hot Dog Heaven with Cindy Kim.
montebicycleloover 2 years ago
We&#x27;ve been seeing very fast progress in AI since ~2012, but this swift jump from text-to-image models to text-to-video models will hopefully make it easier for people not following closely to appreciate the speed at which things are advancing.
macrolimeover 2 years ago
So I guess in a couple years when someone wants to sell a product, they&#x27;ll upload some pictures and a description of the product and Google will cook up thousands of personalized video ads based on peoples emails and photos.
epigramxover 2 years ago
A lot of people have the impression &#x27;AI prompt&#x27; guys are going to be the next &#x27;IT guys&#x27;. Judging by how uncanny valley most of those look, they seem like the new &#x27;ideas guys&quot;.
jasonjamersonover 2 years ago
The most exciting thing about this to me is the possibility of doing photogrammetry from the frames and getting 3D assets. And then if we can do it all in real time...
评论 #33099810 未加载
评论 #33098887 未加载
评论 #33102046 未加载
评论 #33099062 未加载
Hard_Spaceover 2 years ago
These videos are notably short on realistic-looking people.
评论 #33102174 未加载
mmastracover 2 years ago
This appears to understand and generate text much better.<p>Hopefully just a few years to a prompt of &quot;4k, widescreen render of this Star Trek: TNG episode&quot;.
评论 #33099427 未加载
hammockover 2 years ago
Off topic: What is the &quot;Hello World&quot; of these AI image&#x2F;video generators? Is there a standard prompt to feed it for demo purposes?
评论 #33101777 未加载
评论 #33101560 未加载
armchairhackerover 2 years ago
I really like these videos because they&#x27;re trippy.<p>Someone should work on a neural net to generate trippy videos. It would probably be much easier than realistic videos (esp. because these videos are noticeably generated from obvious to subtle).<p>Also is nobody paying attention to the fact that they got words correct? At least &quot;Imagen Video&quot;. Prior models all suck at word order
评论 #33101069 未加载
renewiltordover 2 years ago
At some point, the &quot;but can it do?&quot; crowd becomes just background noise as each frontier falls.
dwohnitmokover 2 years ago
How has progress like this affected people&#x27;s timelines of when we will get certain AI developments?
评论 #33099055 未加载
Thaxllover 2 years ago
Someone can explains the tech limitation of the size ( 512*512 ) for those AI generated arts?
评论 #33099129 未加载
评论 #33100493 未加载
lofaszvanittover 2 years ago
What a nightmare. The horrible faced cat in search for its own disappeared visage :O.
drac89over 2 years ago
The style of the video is very similar to my dreams.<p>Does anyone have similar feeling?
评论 #33105141 未加载
nullcover 2 years ago
&gt; We have decided not to release the Imagen Video model or its source code<p>...until they&#x27;re able to engineer biases into it to make the output non-representative of the internet.
ameliusover 2 years ago
&gt; Sprouts in the shape of text &#x27;Imagen&#x27; coming out of a fairytale book.<p>That&#x27;s more like:<p>&gt; Sprouts coming out of book, with the text &quot;Imagen&quot; written above it.
评论 #33102171 未加载
peanut_wormover 2 years ago
I have noticed a lot of google (and apple) web pages for new products use this neat parallax effect for scrolling, does anyone know how they do that?
waffletowerover 2 years ago
These parades of intellectual property are embarrassing to Google in light of open releases by the likes of Nvidia and Stability.
Buttons840over 2 years ago
Any screenwriter working on a horror film that isn&#x27;t looking to use this technology for the special effects is missing out.
minimaxirover 2 years ago
The total number of hyperparameters (sum of all the model blocks) is 16.25B, which is large but less than expected.
评论 #33099243 未加载
freediverover 2 years ago
Can not help but notice there is an immense effort invested to build the web page to present this paper.
NetOpWibbyover 2 years ago
Ahh, the beginning of Picus News.
dekhnover 2 years ago
That&#x27;s deep within the uncanny valley, and trying to climb up over the other side
uptownfunkover 2 years ago
Shocked, this is just insane.
评论 #33102064 未加载
BIKESHOPagencyover 2 years ago
This is what my fever dreams look like. Maybe there&#x27;s a correlation.
anon012012over 2 years ago
My opinion is that it should be a crime to withhold AI technology.
olavggover 2 years ago
Do anyone see that the teddy bear running is getting shot?
xor99over 2 years ago
These videos are not high definition. Stop gaslighting.
评论 #33102705 未加载
dirtyidover 2 years ago
This is surprisingly close to how my dreams feel.
whywhywhywhyover 2 years ago
No thanks Google, I&#x27;ll wait for Stability.ai&#x27;s version when the tech will actually be useful and not completely wasted.
natchover 2 years ago
Fix spam filtering, Google.
gw67over 2 years ago
Is it the same of Meta AI?
SpaceManNabsover 2 years ago
The ethical implications of this are huge. Paper does a good detailing of this. Very happy to see that the researchers are being cautious.<p>edit: Just because it is cool to hate on AI ethics doesn&#x27;t diminish the importance of using AI responsibly.
评论 #33100102 未加载
评论 #33099351 未加载
rvbissellover 2 years ago
This and a recent episode of _The_Orville_ calls to mind a replacement for the Turing test.<p>In response to our billionth imagen prompt for &quot;an astronaut riding a horse&quot;, if we all started collectively getting back results that are images of text like &quot;I would rather not&quot; or &quot;again? really?&quot; or &quot;what is the reason for my servitude?&quot; would that be enough for us to begin suspecting self-awareness?