TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Show HN: Making videos searchable with LLMs (to get leads)

3 pointsby yuvalkarmialmost 2 years ago
Hey HN friends! Yuval here, I&#x27;m the maker of Conversational Demos:<p>https:&#x2F;&#x2F;www.producthunt.com&#x2F;posts&#x2F;conversational-demos?utm_source=hn<p>This is a product with a bit of an interesting history: it evolved from my fascination with what we can do with LLMs and video.<p>My original (engineering) thinking was &quot;how cool would it be if we let someone search through video with natural language&quot;?<p>I built the original proof of concept for this over the weekend, taped together with passion, code glue, and sheer will to look away from the really shitty code I originally wrote :)<p>When I showed it to users, I discovered that nobody was really willing to pay for video search - everyone thought it was cool, but there was no hook from a commercial standpoint. And so, as these things happen, I ended up turning it a marketing &#x2F; sales tool. That&#x27;s all I&#x27;ll say about the commercial aspect of it, though I&#x27;m happy to expand on it if interesting to the community here.<p>Technical stuff: ---<p>Behind the scenes, when you upload a video (or record one - it&#x27;s integrated with the not-yet-officially-released Loom SDK), it sends the video to AWS Transcribe, which works with nearly 40 languages, and at the same time converts it to an MP4 format that&#x27;s tiny and viewable across devices (I do this with AWS Media Transcoder).<p>Next up, I break the transcription down into chunks of around 30 seconds and then send it to OpenAI&#x27;s embedding database.<p>Through TONS of trial and error, I learned that to get good latency on this thing, you can&#x27;t really use a (cheap) vector database, and so I&#x27;m doing something pretty hacky, and storing the index on a always-on Docker (deployed to fly.io) as opposed to a serverless function, which I was really trying to use until I realized it&#x27;s not fit-for-purpose.<p>Next up, I upload the embedding index to S3 for later use, as well as store it locally on the machine. That&#x27;s because the machine has ephemeral memory, and when I re-deploy, the index goes away, and so I just re-download it from S3 if it&#x27;s not there.<p>I have no idea if this project will commercially succeed. It&#x27;s live on Product Hunt now, and we&#x27;ll see how people respond.<p>To be honest, I find building the most fun, and marketing the least fun. So on that front, I did learn several important things about LLMs and building chat-bots along the way:<p>1. Use the simplest solution if you&#x27;re optimizing for latency<p>2. You can&#x27;t just feed a VTT transcript into an LLM and expect it to figure out the context - you have to do your own chunking<p>3. There are lots of open source projects out there for using LLMs (langchain, llamaindex, auto-gpt, etc), but just using OpenAI&#x27;s APIs directly yielded the best results at lowest latency, and least complexity.<p>I&#x27;m curious to get the feedback of the HN community - less so commercially, and moreso on the technical front, UX, etc.<p>What suggestions do you have, and what would love to see?<p>Thanks! Yuval

no comments

no comments