Skip to content

Deep neural audiovisual network

Notifications You must be signed in to change notification settings

skumar2891/AVGAN

 
 

Repository files navigation

AVGAN

Deep neural audiovisual network

To combine the sound and image input streams the concatenation was used,after the two transformed soundimages were concatenated,as well. After the sound and image concatenation the GAN network starts the training while the generator tries to generate better images and the discriminator judges the output. After the training the network has to do the reverse way to display deconvoluted images. The last step is de deconcatenation of the sound images and the retransformation into a .wav file, so sound can be played.

For more information read the paper: https://github.com/markus-weiss/AVGAN/blob/master/Forschungsprojekt%20(2).pdf

AVGAN - on google Colab

https://colab.research.google.com/drive/1bxOf8m8MZAOuuih5zz3Wk3taASdY-U2t

The presentation

https://github.com/markus-weiss/AVGAN/blob/master/Generative%20Modelle%20f%C3%BCr%20audio-visuelle%20Formen.pptx.pdf

A GIF from the frist training with the dcgan

https://github.com/markus-weiss/AVGAN/blob/master/AVGAN_Paper.pdf

About

Deep neural audiovisual network

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 95.7%
  • Python 4.3%