Clone the repo
git clone https://github.com/ai-in-health/PromptNet
Install dependencies
conda create -n PromptNet python==3.9
conda activate PromptNet
# install a proper version of PyTorch
# see https://pytorch.org/get-started/previous-versions/
pip install pytorch>=1.10.1 torchvision>=0.11.2 torchaudio>=0.10.1 pytorch-cuda==11.8 -c pytorch -c nvidia
# install the rest dependencies
pip install -r requirement.txt
which should install in about 5 mins. We can run the code under torch==2.0.1
and torchvision==0.15.2
. Other versions may work.
The root of this repo/
mimic_cxr
├── annotation.json # MIMIC_CXR dataset
├── images
│ ├── p10
│ ├── .
│ ├── .
│ └── p19
finetune
├── annotation.json # Downstream dataset
├── images
│ ├── patient 1
│ ├── .
│ ├── .
pretrained
├── resnet
│ └── pytorch_model.bin
│
decoder_config
│ └── decoder_config.pkl # configuration file for model
│
model_weights
│ └── few-shot.pth
│ └── fully-supervised.pth
│
prompt
│ └── prompt.pt
│
└── ...
The data
folder contains the following subfolders:
- The
mimic_cxr
folder contains MIMIC-CXR image data and corresponding annotation (eg., image id, report etc...). - The
finetune
folder contains the image data and corresponding annotation (eg., image id, report etc...) of the downstream dataset. - The
pretrained
folder contains the initialized weights for our encoder which will download automatically when you run the code. - The
decoder_config
,model_weights
andprompt
folders contain the configuration file of the model, its weights, and the support prompt in the inference stage, respectively.
Here is an example of running command:
export dataset=mimic-cxr
python PromptNet.py \
--image_dir='./${dataset}/images' \
--json_path='./${dataset}/annotation.json' \
--dataset=${dataset} \
--max_seq_length=100 \
--threshold=10 \
--bs=32
Here is an example of running command:
export dataset=downstream_dataset
python PromptNet.py \
--train_mode=fine-tuning \
--image_dir='./${dataset}/images' \
--json_path='./${dataset}/annotation.json' \
--dataset=${dataset} \
--max_seq_length=100 \
--threshold=10 \
--bs=32 \
--prompt='./prompt/prompt.pt' \
--prompt_load=yes \
--random_init=no \
--weight_path='./model_weights/${model_weights}' \
Here is an example of running command:
export dataset=downstream_dataset
python test.py \
--dataset='$dataset' \
--max_seq_length=100 \
--threshold=10 \
--epochs=10 \
--bs=16 \
--load='./your_model_weights'
You could specify $dataset
to load your own corpus. Our method can be tested within 10 mins on a single V100 GPU
- To evaluate report generation, ensure that your system has installed JAVA. Here is an example:
- Download from the official website (https://www.java.com/en/download/manual.jsp) to obtain, e.g., jdk-8u333-linux-x64.tar.gz
- Unzip the file by running
tar -zxvf jdk-8u333-linux-x64.tar.gz
, and you will see the jre folder - Write the following lines to ~/.bashrc:
echo "export JRE_HOME=path/to/jre" >> ~/.bashrc
echo "export PATH=${JRE_HOME}/bin:$PATH" >> ~/.bashrc
- Activate the settings by running
source ~/.bashrc
- See if the java has been installed:
java -version
- You should install packages
pycocoevalcap
andpycocotools
(included inrequirement.txt
). - When calculating the
SPICE
metric, the code will try to automatically download two filesstanford-corenlp-3.6.0.jar
andstanford-corenlp-3.6.0-models.jar
, and save them to ${pycocoevalcapPath}/spice/lib/. If you encounter a network issue, you can prepare these two files by yourself:- Download a zip file from https://nlp.stanford.edu/software/stanford-corenlp-full-2015-12-09.zip
- Unzip it to get the above two files
- Run
pip show pycocoevalcap
to see where the package has been installed - Move the two files to ${pycocoevalcapPath}/spice/lib/
- To evaluate report generation, you should install the
stanfordcorenlp
package (included inrequirement.txt
), and download stanford-corenlp-4.5.2. The following is an example. Note that we setcorenlp_root = data/stanford-corenlp-4.5.2
inconfigs/__init__.py
.
wget https://nlp.stanford.edu/software/stanford-corenlp-4.5.2.zip --no-check-certificate
wget https://nlp.stanford.edu/software/stanford-corenlp-4.5.2-models-german.jar --no-check-certificate
wget https://nlp.stanford.edu/software/stanford-corenlp-4.5.2-models-french.jar --no-check-certificate
unzip stanford-corenlp-4.5.2.zip -d data/
mv stanford-corenlp-4.5.2-models-german.jar data/stanford-corenlp-4.5.2/
mv stanford-corenlp-4.5.2-models-french.jar data/stanford-corenlp-4.5.2/
rm stanford-corenlp-4.5.2.zip
If you encounter any problems when using the code, or want to report a bug, you can email [email protected]. Please try to specify the problem with details so we can help you better and quicker!