TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Launch HN: Sieve (YC W22) – Pluggable APIs for Video Search

71 pointsby mvoodarlaover 3 years ago
Hi HN, we’re Mokshith and Abhi from Sieve (<a href="https:&#x2F;&#x2F;sievedata.com" rel="nofollow">https:&#x2F;&#x2F;sievedata.com</a>). We’re building an API that lets you add video search to internal tools or customer applications, instantly. Sieve can process 24 hours of video in less than 10 minutes, and makes it easy to search video by detected objects &#x2F; characteristics, motion data, and visual similarity. You can use our models out of the box, or plug-in your own model endpoints into our infrastructure. (&#x27;Model&#x27; here means any software that produces output given an image.)<p>Every industry from security, to media, supply chain, construction, retail, sports, and agriculture is being transformed by video analytics—but setting up the infrastructure to process video data quickly is difficult. Having to deal with video ingestion pipelines, computer-vision model training, and search functionality is not pretty. We’re building a platform that takes care of all of this so teams can focus on their domain-expertise, building industry-specific software.<p>We met in high school, and were on the robotics team together. It was our first exposure to computer vision, and something we both deeply enjoyed. We ended up going to UC Berkeley together and worked on computer vision at places like Scale AI, Niantic, Ford, NVIDIA, Microsoft, and Second Spectrum. We were initially trying to solve problems for ourselves as computer vision developers but quickly realized the unique problems in video having to do with cost, efficiency, and scale. We also realized how important video would be in lots of verticals, and saw an opportunity to build infrastructure which wouldn’t have to be rebuilt by a fullstack dev at any company again.<p>Let’s take the example of cloud software for construction which might include tons of features from asset trackers to rental management and compliance checks. It doesn’t make sense for a construction software company to build their own video processing for telematics—the density and scale of video make this a difficult task. A single 30 FPS camera generates over 2.5M frames within a day of recording. Imagine this across thousands of cameras and many weeks of footage—not to mention the actual vertical-specific software they’re building for end users.<p>Sieve takes care of everything hard about processing and searching video. Our API allows you to process and search video with just two API calls. We use filtering, parallelization, and interpolation techniques to keep costs low, while being able to process 24 hours of video in under 10 minutes. Users can choose from our pre-existing set of models, or use their own models with our video processing engine. Our pricing can range anywhere from $0.08-$0.45 per minute of video processed based on the models clients are interested in and usage volume. Our FAQ page (<a href="https:&#x2F;&#x2F;sievedata.com&#x2F;faq" rel="nofollow">https:&#x2F;&#x2F;sievedata.com&#x2F;faq</a>) explains these factors in more detail.<p>Our backend is built on serverless functions. We split each video into individual chunks which are processed in parallel and passed through multiple layers of filters to determine which chunks are “important”. We’re able to algorithmically ignore parts of video which are static, or change minimally, and focus on the parts that contain real action. We then run more expensive models on the most “important” parts of video, and interpolate results across frames to return information to customers at 30 FPS granularity. Our customers simply push signed video URLs to our platform, and this happens automatically. You can then use our API to query for intervals of interest.<p>We haven’t built an automated sign up flow yet because we&#x27;re focused on building out the core product for now. But we wanted to give all of you the chance to try Sieve on your own videos for free, so we&#x27;ve set up a special process for HN users. Try it out here: <a href="https:&#x2F;&#x2F;sieve-data.notion.site&#x2F;Trying-Sieve-s-Video-Search-4bd7754bb04d468fb1a6c98225f68ccb" rel="nofollow">https:&#x2F;&#x2F;sieve-data.notion.site&#x2F;Trying-Sieve-s-Video-Search-4...</a>. We&#x27;ll email you a personal, limited-access API key.<p>Here&#x27;s a video demo of using our dashboard to do video search: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=_uyjp_HGZl4" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=_uyjp_HGZl4</a><p>We’d love to hear what you think about the product and vision, and ideas on how we can improve it. Thanks for taking the time to read this, we’re grateful to be posting here :)

5 comments

plasmaover 3 years ago
Neat project, you may want to consider adding audio processing too (eg sound detected) as part of the video.<p>You could go deeper and compare samples of audio that could be uploaded separately (eg, siren sounds), check out MFCC processing <a href="https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Mel-frequency_cepstrum#Applications" rel="nofollow">https:&#x2F;&#x2F;en.wikipedia.org&#x2F;wiki&#x2F;Mel-frequency_cepstrum#Applica...</a> to do Shazam-style audio comparison.<p>I wonder too if you process it on a per-frame basis or you can take series of frames too (eg analyze the last 5 seconds of frames) to detect things like a &quot;hand wave&quot;.
评论 #30184292 未加载
jensneuseover 3 years ago
Hey, that&#x27;s a really nice product. I love that it&#x27;s API first? Would you be interested in adding it to our API hub? We&#x27;re currently in private beta and would love to have your API on it. Why? It&#x27;s the best developer experience to integrate APIs and we&#x27;d love to send our users your way. Here&#x27;s a link if you&#x27;re interested: <a href="https:&#x2F;&#x2F;hub.wundergraph.com&#x2F;" rel="nofollow">https:&#x2F;&#x2F;hub.wundergraph.com&#x2F;</a>
评论 #30183608 未加载
applgo443over 3 years ago
This is really cool!<p>I have a few questions<p>1. Is there really no other competitor or company that tried to tackle this problem in the past? It feels like a really common usecase and someone must&#x27;ve done something about it!<p>2. Do you have a fixed set of words that the user should use to query? I&#x27;m an AI researcher&#x2F;practitioner who worked in this area. It&#x27;s super difficult to search for tail objects in images&#x2F;text.<p>3. Why API first?
评论 #30184742 未加载
评论 #30186176 未加载
PanManover 3 years ago
Like the concept, but especially for live feeds, seems expensive? at $3456&#x2F;month&#x2F;camera ? Is this not your use-case (it&#x27;s the first one on your use-cases page)? Or am I missing something? Congrats on the launch!
评论 #30184974 未加载
Neanderover 3 years ago
This is cool. Something tangential I&#x27;ve always wanted from video apps and APIs is the ability to highlight any video online. To just bracket cool clips in long videos and then be able to view and later edit those highlights together rapidly.