I'm a huge fan of this kind of practice, where the code for a paper is all located in a single public repository with build instructions, along with directions for how to cite it. Obviously, it's a little tough to do with some more data-intensive sources (besides GH hosting limits, no one really wants to download 100G of data if they're just trying to clone a repository), but this kind of thing sets a high standard for reproducibility of published results.
I'm working with same model, but in a real-time pipeline developed with GStreamer, Rust and PyTorch:<p><a href="https://twitter.com/rozgo/status/1255961525187235842" rel="nofollow">https://twitter.com/rozgo/status/1255961525187235842</a><p>Live motion transfer test with crappy webcam:<p><a href="https://youtu.be/QVRpstP5Qws" rel="nofollow">https://youtu.be/QVRpstP5Qws</a>
Very cool, reminds me of Avatarify, which is also based upon the First Order Model work:<p><a href="https://github.com/alievk/avatarify" rel="nofollow">https://github.com/alievk/avatarify</a>
Pretty cool. Reminds me of <a href="https://github.com/yemount/pose-animator" rel="nofollow">https://github.com/yemount/pose-animator</a><p>I would use it if there was a JavaScript port.
Looks like the file mentioned in this step<p>> gdown --id 1wCzJP1XJNB04vEORZvPjNz6drkXm5AUK<p>Is no longer accessible (too many downloads in too short a time)<p>Edit: For anyone else with the same problem, the file in question is "vox-cpk.pth.tar" which can be found in various places on the internet.
The google colab version is not really real-time, is that correct? It loads pre-recorded video. I guess that is because it is not easy to add realtime feed from camera into browser notebook or what are the limitations there?
The paper & final models don't to justice for detailed outputs though, but this is still a great model for datasets with no annotations per se.
does anyone know if using this tool to generate a music video of famous pictures singing a song would violate any copyrights? it seems like a fun exercise.