Jump to content
IGNORED

you know all this neural net nonsense? you can do it with audio too apparently...


lala

Recommended Posts

im not sure how it would all work, apparently its a LOT of work, listening to millions of songs (?), then you can maybe feed it white noise and it will make music?

 

 

or zoom in on a track (slow down?) and have only taught it amen breaks and it starts to sketch out or something.

 

 

 

 

Wheres Autechre when you need them

Link to comment
Share on other sites

could do it with spectral/FFT synthesis? It'd probably sound like low quality mp3s though.

 

 

ie. feed it spectrograph images of certain sounds and try to get it to extrapolate from there.. not sure if it'd be useful but probably worth a try

Link to comment
Share on other sites

 

Can this be done on audio? video?

Yes. To make a video you can run the code on each individual frame of the video then stitch them together afterwards. But there are more efficient ways discussed in this thread. The best resource for learning about this is here: https://github.com/graphific/DeepDreamVideo

If you wish to make one of those zoom-into-an-image-really-far gifs like this one then you should follow the guide here: (TODO: guide link)

To perform this on audio you need to really know what you are doing. Audio works better with RNNs than CNNs. You will need to create a large corpus of simple music to train your RNN on.

https://www.reddit.com/r/deepdream/comments/3cawxb/what_are_deepdream_images_how_do_i_make_my_own/

 

Link to comment
Share on other sites

Yeah I wanna hear what this would sound like so bad. Any guess on how soon we'd get to hear an example? Surely someones gonna give it a go.

 

Even after reading the thread on the image versions of this, I still don't really get whats going on so can't really judge how difficult it'd be.

Link to comment
Share on other sites

very...?

 

 

you have to let the net listen to millions of songs apparently. its harder than images according to one article i read. even if we could somehow get the software and/or be able to run it..

Link to comment
Share on other sites

A DNN is just meant to recognize patterns but not output it. All they're doing with images is amplifying certain levels of abstraction to see what the neurons are picking out. With audio, what you'd start out with is a bunch of single notes at layer 1 abstraction, and then at say, layer 12 you'd have the complete mozart symphony or whatever, which means you'd have to feed it the entire symphony (or enough to be unique in the system) for those neurons to fire and recognize the pattern. It's the same thing as seeing a dog or whatever in an image. A neural net is basically a set of constraints that trigger only when the right inputs are presented, from simple and short scale to big complex objects.

 

So with audio, you'd end up with different length "audio clips" either of single notes, chord progressions or entire chorus's or even entire tracks, blended together or in a row depending on the code. I don't know anything about programming audio software, so it's a little hard to see how to output the results. Either samples or some kind of synthesis generator that is able to output exactly what one wants?

 

See this is where I'm really interested in DNN's cause, it would be nice if the computer could have an understanding of music theory, and beyond that a taste of what's modern or what's old, and then be able to output original music that way. In theory all these abstract thoughts humans have should be just networks of neurons etc. You're kind of doing a randomizing effect otherwise, unless it stumbles upon some kind of music theory by accident IDK

Link to comment
Share on other sites

Yeah ^ thats why when you run the DD you get the output numbers going up in resolution through the inception layers.

 

would it be simple notes or would it be more like looking at the wave from afar and then zooming in to it, with realtime being the final inception layer?

Link to comment
Share on other sites

I have no clue. I imagine there's many ways to go about it. I mean you could do pure audio, or you could "convert to midi" type and work with the notes. I'm not sure what would be the best thing to feed it or what would be the best way to output it when you amplify layers to render them etc. Would be cool to input autechre as pure audio and then have it put those harmonies and frequencies etc back together in various layers but don't ask me how to do it ;e

 

Just to add I *think* you should be able to have abstraction layers for several dimensions like frequency, harmony, amplitude etc, which means it's not just a waveform but different aspects of audio.

Link to comment
Share on other sites

Guest skibby
lala, on 09 Jul 2015 - 2:18 PM, said:

im not sure how it would all work, apparently its a LOT of work, listening to millions of songs (?), then you can maybe feed it white noise and it will make music?

 

they already did it, it's called jazz

Jazz_Silhouette.jpg

Link to comment
Share on other sites

evidently the audio here was generated by neural net stuff

 

Isn't that battle music for FF7 on the PC before it recently got 'fixed' in the steam release ?
Link to comment
Share on other sites

 

evidently the audio here was generated by neural net stuff

 

Isn't that battle music for FF7 on the PC before it recently got 'fixed' in the steam release ?

 

No, it's the latest Liturgy album.

Link to comment
Share on other sites

  • 3 weeks later...

Archived

This topic is now archived and is closed to further replies.

  • Recently Browsing   0 members

    • No registered users viewing this page.
×
×
  • Create New...

Important Information

We have placed cookies on your device to help make this website better. You can adjust your cookie settings, otherwise we'll assume you're okay to continue.