sayakpaul / Multimodal-Entailment-Baseline

This repository shows how to implement a basic model for multimodal entailment.

Home Page:https://keras.io/examples/nlp/multimodal_entailment/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Multimodal-Entailment-Baseline

This repository shows how to implement baseline models for multimodal entailment. One of these models looks like so:

High-resolution version is available here.

These models use the multimodal entailment dataset introduced here. This repository is best followed along with this blog post on keras.io: Multimodal entailment. The blog post goes over additional details, thought experiments, notes, etc.

A fun fact

The accompanying blog post marks the 100th example on keras.io.

About the notebooks

  • Multimodal entailment.ipynb: Shows how to train the model shown in above figure.
  • multimodal_entailment_attn.ipynb: Shows how to train a similar model with cross-attention (Luong style).
  • text_entailment.ipynb: Uses only text inputs to train a BERT-based model for the enatailment problem.

Acknowledgements

Thanks to the ML-GDE program for providing GCP credits.

Thanks to Nilabhra Roy Chowdhury who worked on preparing the image data.

About

This repository shows how to implement a basic model for multimodal entailment.

https://keras.io/examples/nlp/multimodal_entailment/

License:Apache License 2.0


Languages

Language:Jupyter Notebook 100.0%