Kajiyu / LLLNet

Keras Implementation of "Look, Listen and Learn" Model

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LLLNet

About

This is a Keras implementation of "Look, Listen and Learn" Model on the research by R. Arandjelovic and A. Zisserman, at DeepMind. This model can get cross-modal features between audios and images.

Core Concept

Audio-visual correspondence task (AVC)

Different Point from Original Model

  • SqueezeNet is used for visual CNN. Model Figure

About

Keras Implementation of "Look, Listen and Learn" Model


Languages

Language:Jupyter Notebook 86.7%Language:Python 13.3%