TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

The Rise of Synthetic Audio Deepfakes

183 pointsby ajay-dalmost 5 years ago

12 comments

eigenvaluealmost 5 years ago
My all-time favorite audio deepfake is Nobel prize winner Milton Friedman reading the lyrics to the 50 Cent track &quot;PIMP&quot;. It really captures Friedman&#x27;s tell-tale cadence and idiosyncratic lilt: <a href="https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=4mUYMvuNIas" rel="nofollow">https:&#x2F;&#x2F;www.youtube.com&#x2F;watch?v=4mUYMvuNIas</a>
yaloginalmost 5 years ago
There is already a large problem with political ads cherry picking and slicing audio and video to cheat viewers. I really worry that deep fakes will take it to another level completely. I fully expect the current administration to eagerly adopt it if available.
评论 #23967291 未加载
评论 #23966499 未加载
评论 #23970051 未加载
kharakalmost 5 years ago
This might be paranoid. But I&#x27;ve established a protocol with some people in my life. Should someone with my voice ever contact them and ask for money (because emergency bla bla), nothing is to be done until a passphrase is mentioned. It&#x27;s only a matter of time, until someone gets significant voice data and related contact numbers and proceeds with using those voices to train a model. Afterwards, that model will be used to real-time fake the original voice in a scamming attempt.
评论 #23969674 未加载
paul7986almost 5 years ago
Recently a friend changed her number and told me via text. Before adding her number i asked her a question that she and only I would know like who sat next to you at the old office.<p>Think im going to keep doing this type of verification. It may annoy friends and family, but not sure how a hacker could ever know such small details between you and another.
评论 #23966663 未加载
blueblistersalmost 5 years ago
There is an annual challenge for synthetic voice detection, ASVSpoof, that evaluates submissions on different types of attacks to speaker verification systems: text to speech, voice conversions and replayed attacks.<p>The conclusion from the 2019 evaluation [1]: <i>known</i> synthetic deep fakes are fairly easy to detect using simple models with very low error rates (even high-fidelity techniques with Wavenet vocoders).<p>[1]: ASVspoof 2019: Future Horizons in Spoofed and Fake Audio Detection speech (<a href="https:&#x2F;&#x2F;www.isca-speech.org&#x2F;archive&#x2F;Interspeech_2019&#x2F;pdfs&#x2F;2249.pdf" rel="nofollow">https:&#x2F;&#x2F;www.isca-speech.org&#x2F;archive&#x2F;Interspeech_2019&#x2F;pdfs&#x2F;22...</a>)
nshmalmost 5 years ago
&gt; Deepfake technology is not sophisticated enough to mimic an entire phone call with someone.<p>With modern voice conversion technology it is perfectly possible actually.
评论 #23965991 未加载
评论 #23965390 未加载
phjesusthatguy3almost 5 years ago
Audio &quot;deepfakes&quot; have been worked on much longer than ones for video, although video deepfakes have the added issue of deep-faking synchronized audio. Today&#x27;s consumers don&#x27;t seem to be bothered by video deepfakes if they play to the beliefs of the audience.
motohagiographyalmost 5 years ago
Useful example is how the Joe Biden Burisma phone call that bubbled up through Russian media was fabricated. I pulled it apart with ffmpeg and there were a number of artifacts that showed editing and splicing.<p>If you&#x27;re handy with ffmpeg and python, you can assess their veracity pretty easily. Of course, if I were on a political ratf&#x27;ing team, I&#x27;d use the same tools to add those artifacts to a copy of an offending (real but off message) stream and amplify the distribution of that fake-faked version with a debunking press release handy, so YMMV. While the Biden thing wasn&#x27;t a deepfake directly, (shallow fake?) we&#x27;re going to see tons of actual deepfakes around the election.<p>IMO, elections are no longer between candidates, they are a war on truth for domination of the narrative - office is the effect. A campaign that focuses on what happens once the war is over is daydreaming about the future and distracted from the present and this will lose them key battles. For this reason, I think deepfakes are going to be the biggest weapon in campaign arsenals for the near future. Interesting times.
评论 #23965930 未加载
评论 #23968746 未加载
sidthekidalmost 5 years ago
The images of spectrogram analysis between the real and fake voices seemed to be distinguishable by the human eye. Can a image model be trained to detect fake voice spectrograms based on pitch and tone choppiness?
评论 #23967269 未加载
评论 #23967217 未加载
seesawtronalmost 5 years ago
This makes me wonder how would one go about adding an authentication key to audio? We have seen in the past encryption for text shared via email and watermarks embedded in images but I haven&#x27;t come across something for audio. Happy to hear if someone has worked in this field.
评论 #23965977 未加载
leptoniscoolalmost 5 years ago
How long does it take to train the model?
inasioalmost 5 years ago
Next milestone I&#x27;m waiting for: Trump audio to Sarah Cooper video on the fly