Danyache / skoltech_image_cap

Полное собрание кода для задач X-Ray captioning

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Image Captioning with Attention : How to Increase Speed and Quality of X-Ray Diagnostics

Here we have three main architectures implemented.

In the Show-Attend-Tell folder is the implementation of https://arxiv.org/pdf/1502.03044v1.pdf paper and it's versions combined with GPT-2 models.

In the On-the-Automatic-Generation-of-Medical-Imaging-Reports is the implementation of https://arxiv.org/pdf/1711.08195.pdf paper.

In the Transformer-Based-Generation folder we have Transformer-based implementation using the fairseq github.

All the models are implemented using the CheXNet weights. The weights that are used are at DGX at /raid/data/cxr14-2/DenseNet121_aug4_pretrain_WeightBelow1_1_0.829766922537.pkl. If you dont use the DGX server you should change the weights path in the following files:

  • Transformer-Based-Generation/preprocess/preprocess_images_cxr.py
  • Show-Attend-Tell/models.py
  • On-the-Automatic-Generation-of-Medical-Imaging-Reports/models.py

About

Полное собрание кода для задач X-Ray captioning


Languages

Language:Jupyter Notebook 96.0%Language:Python 3.9%Language:Shell 0.1%