1429904852 / R-GCN

[ACMMM 2022] Learning from Different text-image Pairs: A Relation-enhanced Graph Convolutional Network for Multimodal NER

Home Page:https://dl.acm.org/doi/10.1145/3503161.3548228

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

R-GCN

Code and data for "Learning from Different text-image Pairs: A Relation-enhanced Graph Convolutional Network for Multimodal NER" (ACMMM 2022)

Overview

  • In this paper, we propose to leverage the external matching relations between different (text, image) pairs to improve the performance on the MNER task.

  • We design a Relation-enhanced Graph Convolutional Network (R-GCN) to model inter-modal relations and intra-modal relations simultaneously.

Setup

Dependencies

+ python 3.6
+ tensorflow 1.14.0
+ numpy 1.14.5

Download and preprocess the datasets

Because the image features after processing is very large, you can download them via the link Baidu Netdist. It should be noted that the path of the data is consistent with the file tree.

├── /data/
│  ├── /twitter2015/
│  │  │  ├── /images2015_feature/	        // the image feature for each image
│  │  │  ├── /twitter2015_images/	        // the original image
│  │  │  ├── /img2img_sim_topk_vec_2015/	// the top-k image feature for each image
│  │  │  ├── /img2text_sim_topk_vec_2015/	// the top-k image feature for each sentence
│  │  ├── train_2.txt
│  │  ├── valid_2.txt
│  │  ├── test_2.txt
│  ├── /twitter2017/
│  │  │  ├── /images2017_feature/
│  │  │  ├── /twitter2017_images/
│  │  │  ├── /img2img_sim_topk_vec_2017/
│  │  │  ├── /img2text_sim_topk_vec_2017/
│  │  ├── train_2.txt
│  │  ├── valid_2.txt
│  │  ├── test_2.txt

Usage

  • Train

You can use the folowing command to train R-GCN on the MNER task:

python trainer.py --config_path=config/tw15_config.json
python trainer.py --config_path=config/tw17_config.json
  • Test

You can use the folowing command to test R-GCN on the MNER task:

python test.py --config_path=config/tw15_config.json
python test.py --config_path=config/tw17_config.json

Citation

@inproceedings{10.1145/3503161.3548228,
    author = {Zhao, Fei and Li, Chunhui and Wu, Zhen and Xing, Shangyu and Dai, Xinyu},
    title = {Learning from Different Text-Image Pairs: A Relation-Enhanced Graph Convolutional Network for Multimodal NER},
    year = {2022},
    isbn = {9781450392037},
    publisher = {Association for Computing Machinery},
    address = {New York, NY, USA},
    url = {https://doi.org/10.1145/3503161.3548228},
    doi = {10.1145/3503161.3548228},
    booktitle = {Proceedings of the 30th ACM International Conference on Multimedia},
    pages = {3983–3992},
    numpages = {10},
    location = {Lisboa, Portugal},
    series = {MM '22}
}

If the code is used in your research, please cite our paper.

About

[ACMMM 2022] Learning from Different text-image Pairs: A Relation-enhanced Graph Convolutional Network for Multimodal NER

https://dl.acm.org/doi/10.1145/3503161.3548228

License:Apache License 2.0


Languages

Language:Python 80.8%Language:Jupyter Notebook 19.1%Language:Shell 0.1%