[go: nahoru, domu]

Skip to content

This repository enables you to test X-ray segmentation utlising detectron 2. It also has a new pretraining section that use the professional reports to enhance the pretraining of models with image captioning task.

Notifications You must be signed in to change notification settings

AmirSh15/X-ray_segmentation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

X-ray Segmentation with detectron2 (image-captioining pretraining)


Sep 8, 2022

  • First release of the project.

In this project, asd.

Dependency installation

1. Installation & Running the repository

The code was successfully built and run with these versions:

pytorch-gpu 1.10.1
cudatoolkit 11.3.1
opencv 4.6.0.66
scikit-learn 1.0.2
transformers 4.20.1
detectron2 0.6cuda113
numpy 1.22.4

Note: You can also create the environment I've tested with by importing _environment.yml_ to conda.

2. Installation & Running using Docker

A. Install docker from https://docs.docker.com/engine/install/ubuntu/ (or select other distributions based on your OS)
B. Install nvidia-docker from https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html#docker
C. Build docker image
    docker build -t xray_segmentation .
D. Run docker container (to avoid copying huge dataset for captioning, its better to mount the dataset directory to container)
    docker run -d -t -v /path/to/CANDIDPTX_files/on/host:/X-ray_segmentation/captioning/data/dataset --gpus all xray_segmentation or <image_id>
E. Execute the container
    docker exec -it <container_id> <command>
F. Run the code
    For segmentation: python segmentation/train.py --data_path="segmentation/data"
    For image captioning: python captioning/train.py
        The runing code will ask you to enter the path to the dataset & the path to the segmentation data and also number of workers for training transformer model, So you enter them like below:
            captioning/data/dataset
            segmentation/data
            0 or any other number based on your need
            (If it asks you twice, just enter the same values again)
    

Preprocessing Data

The image captioning transformer model and CNN both is trained on CANDID-PTX dataset, for this aim you can preprocess the data using dataloader in captioning/data/transformer_dataloader.py (for transformer) and captioning/data/cnn_dataloader.py (for CNN)

The Segmentaion model can be trained on both ChestX-Det dataset that you can download from here and CANDID-PTX dataset that you have to recieve access to it from here and CANDID-PTX dataset paper can be found here As detectron2 is implemented to train the segmentation model, the data is better to be in coco style, for this aim you can use the dataloader file in segmentation/data/dataloader for both datasets to prepare them in coco style and load for training process. Also for verification of your annotations in coco style, you can use the visualtization file in segmentation/data/data_visualize.py

You should place the data in the following structure:

/segmentation/
  data/
    CANDID_PTX/
        annotations/
        images/

    ChestX_Det/
        annotations/
        images/

Training

This code is written using detectron2. You can train the captioning models with running train.py and cnn_train.py in /captioning/ directory for transformer and CNN model respectively.
You can also config the model and training parameters in ./configuration.py.

...

Reference

@inproceedings{feng2021curation,
  title={Curation of the CANDID-PTX Dataset with Free-Text Reports},
  author={Sijing Feng, Damian Azzollini, Ji Soo Kim, Cheng-Kai Jin, Simon P. Gordon, Jason Yeoh, Eve Kim, Mina Han, Andrew Lee, Aakash Patel, Joy Wu, Martin Urschler, Amy Fong, Cameron Simmers, Gregory P. Tarr, Stuart Barnard, Ben Wilson},
  booktitle={Radiology: Artificial Intelligence},
  year={2021}
}




About

This repository enables you to test X-ray segmentation utlising detectron 2. It also has a new pretraining section that use the professional reports to enhance the pretraining of models with image captioning task.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published