batra-mlp-lab / visdial

[CVPR 2017] Torch code for Visual Dialog

Home Page:https://arxiv.org/abs/1611.08669

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Pytorch starter code

shubhamagarwal92 opened this issue · comments

Hi,

Do you guyz plan to release starter code in pytorch for the challenge? visdial-rl does provide some insights but is tailored more for Visual Dialog Agents as described in the paper "Learning Cooperative Visual Dialog Agents with Deep Reinforcement Learning"

Hi, visdial-rl is tailored more towards the two-agent setup but does support Visual Dialog training and evaluation i.e. just the Answering agent (A-Bot). The training and evaluation can be used with A-Bot alone to do supervised training on the VisDial 0.5 and 0.9 datasets with evaluation on the standard VisDial metrics.

Though the codebase supports just HRE (Hierarchical recurrent encoder and decoder) in hre.py, it can still serve as a starting point for building your own models for visual dialog.

Hi @shubhamagarwal92, visdial-rl will have some enhancements - we would add a section in _README_describing the usage of that codebase as a starter code for Visual Dialog Challenge. Please watch out for announcements on Discord about this soon. Thanks!