Skip to content

Latest commit

 

History

History
73 lines (52 loc) · 6.27 KB

README.md

File metadata and controls

73 lines (52 loc) · 6.27 KB

Image Captioning for Interpretable Automatic Report Generation

Context

In this project, we implemented a model for Image Captioning for Interpretable Automatic Report Generation. We decided to combine a Res-NET model with XAI component together with a Bi-LSTM for explainable report generation.

We decided to use multi-modality for three main reasons :

  • Multi-modal pre-training : significant progress but most of the time in general domain (using : MS-COCO)
  • Vision and Language : most used information in clinical domain
  • Multi-purpose joint representations of vision and language demonstrated effectiveness for a series of downstream tasks (i.e. diagnosis classification, medical image-report retrieval, medical visual question answering, radiology report generation)

In this repository we report the code for (1) the model submitted to WIRN 2024, corresponding and (2) to some preliminary experiments we did in training a multimodal model


(1) WIRN 2024 Model

Our contribution entails proposing a modified version of MedViLL for improved performance and interpretability, inspired by the approach outlined by Biswas et al. (2020). By integrating eXplainable Artificial Intelligence (XAI) instead of a separate model like Mask R-CNN, we enrich the decoder's representation with both global and region-specific image details, aiming to enhance report generation accuracy and mitigate noise introduced by separate models. This approach not only streamlines the network's parameters but also broadens its applicability across various models and tasks, eliminating the need for retraining or fine-tuning additional components like Mask R-CNN.

Medvill : Acknowledgements

The Medvill architecture is a single BERT-based model that learns unified contextualized vision-language (VL) representation for both Vision Language Understanding (VLU) and Vision Language Generation (VLG). Here are the relevant references :

  • Moon, J. H., Lee, H., Shin, W., Kim, Y. H., & Choi, E. (2022). Multi-modal understanding and generation for medical images and text via vision-language pre-training. IEEE Journal of Biomedical and Health Informatics, 26(12), 6070-6080.
  • https://github.com/SuperSupermoon/MedViLL/

Dependencies & Python Version

Usage

Data used

We used the MIMIC-CXR-JPG dataset which is a dataset wholly derived from MIMIC-CXR, providing JPG format files derived from the DICOM images and structured labels derived from the free-text reports Original Dataset Reference:

  • Johnson, A., Lungren, M., Peng, Y., Lu, Z., Mark, R., Berkowitz, S., & Horng, S. (2024). MIMIC-CXR-JPG - chest radiographs with structured labels (version 2.1.0). PhysioNet. https://doi.org/10.13026/jsn5-t979.

(2) Preliminary Model

In this folder, we have our preliminary model and results of a multimodal model following this architecture Screenshot 2023-09-21 at 11 20 54

Dependencies & Python Version

  • The requirements.txt file contains all the libraries needed to set up your environment and be able to run the code.
  • This project requires Python 3.10.6 version to be run.

Usage

  • The config folder contains the configuration file that allows changing the experiment's hyperparameters.
  • In the src folder, you can find the folder with the data
    • All the code to run the model is contained in the src folder. In particular, by setting the image model as pre-trained in the configuration file (the weights are saved in the trained_models folder in the file image_model_512.pt) and by setting the pretrained variable for the combined model to False, it is possible to train it by running the main.py file. We could not load weights for the combined model because the file is too heavy, but we loaded the json file with the generated reports in the trained_models folder.
  • In the data folder, you can find the folder with the data
  • In the output folder, the output is saved

Data used

References

  • Xue, Y., Xu, T., Rodney Long, L., Xue, Z., Antani, S., Thoma, G. R., & Huang, X. (2018). Multimodal recurrent model with attention for automated radiology report generation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part I (pp. 457-466). Springer International Publishing.
  • Moon, J. H., Lee, H., Shin, W., Kim, Y. H., & Choi, E. (2022). Multi-modal understanding and generation for medical images and text via vision-language pre-training. IEEE Journal of Biomedical and Health Informatics, 26(12), 6070-6080.
  • Ramirez-Alonso, G., Prieto-Ordaz, O., López-Santillan, R., & Montes-Y-Gómez, M. (2022). Medical report generation through radiology images: an Overview. IEEE Latin America Transactions, 20(6), 986-999.
  • Johnson, A., Lungren, M., Peng, Y., Lu, Z., Mark, R., Berkowitz, S., & Horng, S. (2024). MIMIC-CXR-JPG - chest radiographs with structured labels (version 2.1.0). PhysioNet. https://doi.org/10.13026/jsn5-t979.
  • Johnson AE, Pollard TJ, Berkowitz S, Greenbaum NR, Lungren MP, Deng CY, Mark RG, Horng S. MIMIC-CXR: A large publicly available database of labeled chest radiographs. arXiv preprint arXiv:1901.07042. 2019 Jan 21.
  • Goldberger, A., Amaral, L., Glass, L., Hausdorff, J., Ivanov, P. C., Mark, R., ... & Stanley, H. E. (2000). PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation [Online]. 101 (23), pp. e215–e220.