I always wanted a markdown-like video editor. I want to chop up my videos and make notes about what is in them.<p>Text based document would be so much easier for this than big clunky Premiere.
Any thought on providing an option to make a super cut that produces a desired output?<p>Ie `videogrep --input *.mp4 --produce "I am a robot"` and will find all the pieces it needs to produce the desired output?
This is awesome! I’ve considered building something nearly identical over the years, as I’ve definitely used VTT files to aid in searching for content to edit, but never did because getting all the FFmpeg stuff to work made my head hurt. I’m so glad someone else has done the hard work for me and that it’s been documented so well!<p>Love this.
If anyone else decides to give this a try on video files with multiple audio tracks, there doesn't seem to be an easy way to tell it to select a certain track.<p>I got it working by manually adding `-map 0:2` (`2` being the trackid I'm interested in) when calling ffmpeg.<p>You'll have to make that edit in both `videogrep/transcribe.py` as well as `moviepy/audio/io/readers.py`.<p>And I'm not sure how easy adding real support for that would be, considering that moviepy doesn't currently have a way to support it (<a href="https://github.com/Zulko/moviepy/issues/1654" rel="nofollow">https://github.com/Zulko/moviepy/issues/1654</a>)
Exciting!<p>Back in 2011-12, my MFA (poetry) thesis project was a sort of poetic ~conversation between myself, and (selected) poems generated by a program I wrote, using transcripts of Glenn Beck's TV show.<p>I really, <i>really</i> wanted to be able to generate video ~performances of the generated poem in each pair for my thesis reading (and for evolving the project beyond the thesis). I have to imagine videogrep could support that in some form, at least if I had the footage. (Not that I want to re-heat that particular project at this point).<p>Great work.
This is very cool! I wonder if Videogrep works better with videos sourced from Youtube (consistent formats, framerates, bitrates) compared to arbitrary sources.<p>I've used ffmpeg before to chop video bits and merge them before. Mixed results. It'd struggle to cut at exact frames or the audio would go out of sync or the frame rate would get messed up.<p>I gave up and decided to tackle the problem on the playback side. Like players respect subtitle srt/vtt files, I wish there were a "jumplist" format (like a playlist but "intra-file") that you could place alongside video/audio files, and players would automatically play the media as per markers in the file, managing any prebuffering, etc. for smooth playback.<p>For a client project, I did this with the Exoplayer lib on Android, which kinda already has an "evented" playback support where you can queue events on the playback timeline. A "jumplist" file is a simple .jls CSV file with the same filename as the video file.<p>Each line contains:
<start-time>,<end-time>,<extra-features><p>"extra-features" could be playback speed, pan, zoom, whatever.<p>Code parses the file and queues events on the playback timeline (On tick 0 jump to first <start-time>, on each <end-time> go to next <start-time>).<p>I set it up to buffer the whole file aggressively, but that could be improved. Downside may be that more data is downloaded than is played. Upside is that multiple people can author their own "jumplist" files without time consuming re-encode of media.
Resounding notions of Object Oriented Ontology[1] in Cinema[2][3] here, which is very much about pick out & possibly stitching together key items from film.<p>> "<i>All of the elements of a shot’s mise en scène, all of the non-relational objects within the film frame, are figures of a sort. The figure is the likeness of a material object, whether that likeness is by-design or purely accidental. A shot is a cluster of cinematic figures, an entanglement. Actors and props are by no means the only kinds of cinematic figures—the space that they occupy and navigate is itself a figure"</i><p>And the words they say, as seen here.<p>[1] <a href="https://en.wikipedia.org/wiki/Object-oriented_ontology" rel="nofollow">https://en.wikipedia.org/wiki/Object-oriented_ontology</a><p>[2] <a href="https://larvalsubjects.wordpress.com/2010/05/03/cinema-and-object-oriented-ontology/" rel="nofollow">https://larvalsubjects.wordpress.com/2010/05/03/cinema-and-o...</a><p>[3] <a href="https://sullivandaniel.wordpress.com/2010/05/02/film-theory-as-a-form-of-procrastination/" rel="nofollow">https://sullivandaniel.wordpress.com/2010/05/02/film-theory-...</a>
Hi Sam, I'm big fan of your work! Coincidently, I just made a simple POC video editor by editing text using this speech to text model <a href="https://huggingface.co/facebook/wav2vec2-large-960h-lv60-self" rel="nofollow">https://huggingface.co/facebook/wav2vec2-large-960h-lv60-sel...</a>. It might be cool to integrate into your Videogrep tool, it also works offline with CPU or GPU, and gives you timestamps for word or character level.<p><a href="https://twitter.com/radamar/status/1528660661097467904" rel="nofollow">https://twitter.com/radamar/status/1528660661097467904</a>
Very nice project!<p>Would be cool if it didn't need a .srt file, but that it would scan the audio for a search prase.<p>Edit: Never mind, I see that you can create transcriptions using vosk!
This needs a function where you can give it a string and it goes and finds the longest matches from the database then builds a video that says what the string says.<p>Also it would be fun if it output a kdenlive project file, so you could easily tweak the boundaries or clip orders.
Enjoying the serendipity of finding the right tool at the right time: <a href="https://twitter.com/xn/status/1528845032438083584" rel="nofollow">https://twitter.com/xn/status/1528845032438083584</a>
great project! since it relies heavily on subtitle files, and as an alternative to generating your own, which websites would you recommend to find subtitles for videos which are not on youtube i.e. movies and series? preferably ones with ratings systems similar to guitar tabs websites - I can envisage a musical similarity in the variance and quality of user-submitted content e.g. timing, volume, tone, punctuation, expression, improvisation, etc. since I doubt many are composed from the actual scripts. I have never used vosk so am also wondering whether it would be quicker and more reliable than filtering and spot checking say a few subtitle files per video
Super niche but this would be great to build a comprehensive clip archive of the genovaverse.<p>Search by text, generated videos of frequent phrases and other meme-worthy sayings from the Sith Lord.<p>What do you think about that DALE?
somewhat related <a href="https://news.ycombinator.com/item?id=31209924" rel="nofollow">https://news.ycombinator.com/item?id=31209924</a>
WTF is a supercut.
...OK apparently it means cutting a number of parts from the source video containing a given spoken text and joining them together again. Still not sure why you would call that a supercut.
Has Zuckerberg <i>deliberately</i> had work/make-up done to look like his own avatar might in some sort of 'metaverse' world? I can't be alone in thinking a lot of those clips look more like gameplay footage than photography?
The YouTube video linked in that post (<a href="https://www.youtube.com/watch?v=nGHbOckpifw" rel="nofollow">https://www.youtube.com/watch?v=nGHbOckpifw</a>) is probably the most hilarious thing I'll see this week. Thank you for sharing that.<p>Also is Zuckerberg for real? Half of those snippets look like it is a NPC from from a game. ¯\_(ツ)_/¯