TE
TechEcho
Home24h TopNewestBestAskShowJobs
GitHubTwitter
Home

TechEcho

A tech news platform built with Next.js, providing global tech news and discussions.

GitHubTwitter

Home

HomeNewestBestAskShowJobs

Resources

HackerNews APIOriginal HackerNewsNext.js

© 2025 TechEcho. All rights reserved.

Algorithmic Music Generation With Recurrent Neural Networks [video]

67 pointsby rndnalmost 10 years ago

8 comments

crucialfelixalmost 10 years ago
The funny thing is that the only really good ones are the first few where they claim its just random noise. The later ones just sound like a crappy radio.<p>With images the technique works because we like looking at the dense artifacts. Millions of dog heads essentially copy and pasted onto any appendage that looks like it should be a head. It looks like drugs and overwhelms in the same way.<p>If you just take white noise and throw it through a tuned filter bank (which is in essence and in final effect what they are doing here) then you just get crappy audio.<p>The more standard and successful use of NN in composition is the use it on pitch series and compositional forms. Like feeding it all of Beethoven and then getting it to generate similar compositions. That&#x27;s been going on for decades. You can do it with that kind of data.<p>But the thing about pop and electronic music is that the easily machine observable elements are not very interesting. Listen to the 4&#x2F;4 kick and snare pattern in the video. Its boring as hell. (Other tracks can be just a kick and snare and they are amazing and we celebrate them as classics and play them for 20 years. Machines will never understand why)<p>What&#x27;s great and essential are things like spatial relationships between elements in the mix: how does the surge of the compressed synth&#x2F;guitar cause the beat to tumble outward and stir you up ? after a series of peaks in a synth melody then the next time it pulls back creating a space that pulls at your heart strings. you create a negative space that the listener goes into. playing with listeners expectations based on what songs, conventions and tropes they already know and respond to.
评论 #9770663 未加载
评论 #9771064 未加载
评论 #9771061 未加载
评论 #9770493 未加载
mpdehaan2almost 10 years ago
I&#x27;m not 100% up to speed on my AI, but this sounds about like what you&#x27;d get with random variations on a signal, where the neural net is the &quot;which sounds like X&quot; filter, and picks one of the two to survive. But that would be using both some form of a genetic algorithm (details TBD?) and the neural net as the checker. But is it?<p>If they aren&#x27;t doing it that way, I&#x27;d be interested in hearing how it&#x27;s evolving the signal in that given direction - and also how that filter works (what libraries does it use?).<p>Sounds like it hit some sort of local maxima, so this system won&#x27;t ever produce the original song, but something a percentage of the way toward it.<p>I&#x27;m a bit more interested in algorithmic composition, but this could be interesting if trying to blend genres. For a long time I&#x27;ve wanted to build a program that could produce essentially an infinite song morphing between genres with lots of tunable parameters.
msamwaldalmost 10 years ago
It would be interesting to know how novel those sequences are (obviously, the outcome would be far less impressive if what we hear is basically a looped, noisy sample of a song that already exists).
m-i-lalmost 10 years ago
Not much information in the video on how this was achieved, but a quick search for &quot;gruv algorithmic music generation&quot; returns the following: <a href="https:&#x2F;&#x2F;sites.google.com&#x2F;site&#x2F;anayebihomepage&#x2F;cs224dfinalproject" rel="nofollow">https:&#x2F;&#x2F;sites.google.com&#x2F;site&#x2F;anayebihomepage&#x2F;cs224dfinalpro...</a> . Extract:<p><i>We compare the performance of two different types of recurrent neural networks (RNNs) for the task of algorithmic music generation, with audio waveforms as input (as opposed to the standard MIDI). In particular, we focus on RNNs that have a sophisticated gating mechanism, namely, the Long Short-Term Memory (LSTM) network and the recently introduced Gated Recurrent Unit (GRU). Our results indicate that the generated outputs of the LSTM network were significantly more musically plausible than those of the GRU.</i>
leaveyoualmost 10 years ago
Another promising field is RNN applied to TED talks: youtube.com&#x2F;watch?v=-OodHtJ1saY
acdalmost 10 years ago
I think it would sound better if we thought the neural network to play notes and music theory.
anentropicalmost 10 years ago
This is not music. Music is not simply organised sound. Music is a cultural practise.
评论 #9772862 未加载
durbinalmost 10 years ago
Source code link?