Skip to content

Latest commit

 

History

History
111 lines (101 loc) · 6.96 KB

README.md

File metadata and controls

111 lines (101 loc) · 6.96 KB

Multitask Deep Learning Reconstruction and Localization of Lesions in Limited Angle Diffuse Optical Tomography

This is the software repository "MultiNet" for our paper solving problem of breast cancer lesion reconstruction and localization in novel deep spatial attention based paradigms.

Motivation

Diffuse optical tomography (DOT) leverages near-infrared light propagation through tissue to assess its optical properties and identify abnormalities. DOT image reconstruction is an ill-posed problem due to the highly scattered photons in the medium and the smaller number of measurements compared to the number of unknowns. Limited-angle DOT reduces probe complexity at the cost of increased reconstruction complexity. Reconstructions are thus commonly marred by artifacts and, as a result, it is difficult to obtain an accurate reconstruction of target objects, e.g., malignant lesions.Reconstruction does not always ensure good localization of small lesions. Furthermore, conventional optimization-based reconstruction methods are computationally expensive, rendering them too slow for real-time imaging applications. Our goal is to develop a fast and accurate image reconstruction method using deep learning, where multitask learning ensures accurate lesion localization in addition to improved reconstruction. We apply spatial-wise attention and a distance transform based loss function in a novel multitask learning formulation to improve localization and reconstruction compared to single-task optimized methods. Given the scarcity of real-world sensor-image pairs required for training supervised deep learning models, we leverage physics-based simulation to generate synthetic datasets and use a transfer learning module to align the sensor domain distribution between in silico and real-world data, while taking advantage of cross-domain learning. Applying our method, we find that we can reconstruct and localize lesions faithfully while allowing real-time reconstruction. We also demonstrate that the present algorithm can reconstruct multiple cancer lesions. The results demonstrate that multitask learning provides sharper and more accurate reconstruction.

Architecture

The overall architecture of the deep residual attention model. (A) Transfer learning network used for data distribution adaptation between in silico measurements and real data collected from the probe. (B) The generation of in-silico training data pairs (ground truth image and their corresponding forward projection measurements) approximating our probe's specification. (C) Architecture of the proposed deep residual attention model. The squeeze and excitation block (inset) squeezes a given feature map along the channels and excites spatially.

Keywords

Diffuse optical tomography, image reconstruction, deep learning, tissue estimation, lesion localisation, multitask learning, transfer learning, handheld probe.

Citation

@ARTICLE{9557304,
  author={Ben Yedder, Hanene and Cardoen, Ben and Shokoufi, Majid and Golnaraghi, Farid and Hamarneh, Ghassan},
  journal={IEEE Transactions on Medical Imaging}, 
  title={Multitask Deep Learning Reconstruction and Localization of Lesions in Limited Angle Diffuse Optical Tomography}, 
  year={2022},
  volume={41},
  number={3},
  pages={515-530},
  doi={10.1109/TMI.2021.3117276}}

Table of contents

  1. Contribution
  2. Installation
  3. Usage
  4. Questions?

Contribution

  • We investigate the benefit of multitask deep learning on the quality of DOT reconstruction and breast lesion localization.
  • We leverage a deep spatial-wise attention network to adaptively re-weight features and attend to the most important ones.
  • We introduce a distance transform loss to improve lesion localization.
  • We present the first work that tests DL-based DOT reconstruction generalization on real patient data.
  • We conduct experiments to assess the trade-offs between network characteristics, scanning probe design, and performance.

Installation

git clone https://github.com/haneneby/MultiNet.git  
cd MultiNet
conda env create -f requirement.yml --name MultiNet
conda activate MultiNet

Usage

For quick hints about commands:

cd MultiNet
python3 MultiNet.py -h

Training

export CUDA_VISIBLE_DEVICES=0 #or change to your GPU config
mkdir myoutput
cd MultiNet
python3 MultiNet.py --epochs 100 --outputfolder ../myoutput.      

This will show something like:

[FuseNet++.py:100 -          initializer() ] Writing output in /dev/shm/MultiNet/MultiNet/../myoutput
[FuseNet++.py:101 -          initializer() ] Logging directory /dev/shm/MultiNet/MultiNet/../myoutput
[FuseNet++.py:104 -          initializer() ] CONF::		 epochs -> 100
[FuseNet++.py:104 -          initializer() ] CONF::		 lr -> 0.0001
[FuseNet++.py:104 -          initializer() ] CONF::		 seed -> 2
[FuseNet++.py:104 -          initializer() ] CONF::		 device -> gpu
[FuseNet++.py:104 -          initializer() ] CONF::		 batchsize -> 64
[FuseNet++.py:104 -          initializer() ] CONF::		 beta -> 0.2
[FuseNet++.py:104 -          initializer() ] CONF::		 checkpoint -> None
[FuseNet++.py:104 -          initializer() ] CONF::		 datasetdirectory -> ./data/data_samples/
[FuseNet++.py:104 -          initializer() ] CONF::		 outputfolder -> ../resu
[FuseNet++.py:104 -          initializer() ] CONF::		 checkpointdirectory -> .
[FuseNet++.py:104 -          initializer() ] CONF::		 mode -> train
[FuseNet++.py:104 -          initializer() ] CONF::		 outputdirectory -> /dev/shm/MultiNet/MultiNet/../myoutput
...
Epoch 1/100
  16/4539 [..............................] - ETA: 34:23 - loss: 1.4238

This will train the network and save output in myoutput. Examples of outputs are presented in Figures

Evaluation

For evaluation, put all your test measurments in a folder and set it path as an argument. Examples are available under data_samples. Then run the following command:

mkdir myoutput
cd MultiNet
python3 MultiNet.py  --datasetdirectory testdatadir --outputfolder ../myoutput  --mode test

The results will be saved output in myoutput.

Questions?

Please create a new issue detailing concisely, yet complete what issue you encountered, in a reproducible way.