[go: nahoru, domu]

Skip to content

Latest commit

 

History

History
103 lines (75 loc) · 3.84 KB

README.md

File metadata and controls

103 lines (75 loc) · 3.84 KB

X-ray Segmentation with detectron2 (image-captioining pretraining)


Sep 8, 2022

  • First release of the project.

In this project, asd.

Dependency installation

1. Installation & Running the repository

The code was successfully built and run with these versions:

pytorch-gpu 1.10.1
cudatoolkit 11.3.1
opencv 4.6.0.66
scikit-learn 1.0.2
transformers 4.20.1
detectron2 0.6cuda113
numpy 1.22.4

Note: You can also create the environment I've tested with by importing _environment.yml_ to conda.

2. Installation & Running using Docker

A. Install docker from https://docs.docker.com/engine/install/ubuntu/ (or select other distributions based on your OS)
B. Install nvidia-docker from https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#docker
C. Build docker image
    docker build -t xray_segmentation .
D. Run docker container (to avoid copying huge dataset for captioning, its better to mount the dataset directory to container)
    docker run -d -t -v /path/to/CANDIDPTX_files/on/host:/X-ray_segmentation/captioning/data/dataset --gpus all xray_segmentation or <image_id>
E. Execute the container
    docker exec -it <container_id> <command>
F. Run the code
    For segmentation: python segmentation/train.py --data_path="segmentation/data"
    For image captioning: python captioning/train.py
        The runing code will ask you to enter the path to the dataset & the path to the segmentation data and also number of workers for training transformer model, So you enter them like below:
            captioning/data/dataset
            segmentation/data
            0 or any other number based on your need
            (If it asks you twice, just enter the same values again)
    

Preprocessing Data

The image captioning transformer model and CNN both is trained on CANDID-PTX dataset, for this aim you can preprocess the data using dataloader in captioning/data/transformer_dataloader.py (for transformer) and captioning/data/cnn_dataloader.py (for CNN)

The Segmentaion model can be trained on both ChestX-Det dataset that you can download from here and CANDID-PTX dataset that you have to recieve access to it from here and CANDID-PTX dataset paper can be found here As detectron2 is implemented to train the segmentation model, the data is better to be in coco style, for this aim you can use the dataloader file in segmentation/data/dataloader for both datasets to prepare them in coco style and load for training process. Also for verification of your annotations in coco style, you can use the visualtization file in segmentation/data/data_visualize.py

You should place the data in the following structure:

/segmentation/
  data/
    CANDID_PTX/
        annotations/
        images/

    ChestX_Det/
        annotations/
        images/

Training

This code is written using detectron2. You can train the captioning models with running train.py and cnn_train.py in /captioning/ directory for transformer and CNN model respectively.
You can also config the model and training parameters in ./configuration.py.

...

Reference

@inproceedings{feng2021curation,
  title={Curation of the CANDID-PTX Dataset with Free-Text Reports},
  author={Sijing Feng, Damian Azzollini, Ji Soo Kim, Cheng-Kai Jin, Simon P. Gordon, Jason Yeoh, Eve Kim, Mina Han, Andrew Lee, Aakash Patel, Joy Wu, Martin Urschler, Amy Fong, Cameron Simmers, Gregory P. Tarr, Stuart Barnard, Ben Wilson},
  booktitle={Radiology: Artificial Intelligence},
  year={2021}
}