Skip to content

Commit

Permalink
upd paper
Browse files Browse the repository at this point in the history
  • Loading branch information
juliusge committed Jul 17, 2024
1 parent aaa09fe commit 7f6ec6b
Showing 1 changed file with 9 additions and 9 deletions.
18 changes: 9 additions & 9 deletions paper/paper.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,34 +27,34 @@ affiliations:
- name: Hertie Institute for AI in Brain Health, University of Tübingen, Tübingen, Germany
index: 1

date: 20 June 2024
date: 20 July 2024
bibliography: paper.bib
---

<!-- ![Fundus Image Toolbox Icon](../icon.svg) -->

# Summary
The Fundus Image Toolbox is an open source Python suite of tools for working with retinal fundus images. It includes quality prediction, fovea and optic disc center localization, blood vessel segmentation, image registration, and fundus cropping functions. It also provides a collection of useful utilities for image manipulation and image-based PyTorch models. The toolbox has been designed to be flexible and easy to use, thus helping to speed up research workflows. All tools can be installed as a whole or individually, depending on the user's needs. \autoref{fig:example} illustrates main functionalities.
The Fundus Image Toolbox is an open source Python suite of tools for working with retinal fundus images. It includes quality prediction, fovea and optic disc center localization, blood vessel segmentation, image registration, and fundus cropping functions. It also provides a collection of useful utilities for image manipulation and image-based PyTorch models. The toolbox is designed to be flexible and easy to use, thus helping to speed up research workflows. All tools can be installed as a whole or individually, depending on the user's needs. \autoref{fig:example} illustrates the main functionalities.
Find the toolbox at [https://github.com/berenslab/fundus_image_toolbox](https://github.com/berenslab/fundus_image_toolbox).

# Statement of need
In ophthalmic research, retinal fundus images are often used as a resource for studying various eye diseases such as diabetic retinopathy, glaucoma and age-related macular degeneration. Consequently, there is a large amount of research on machine learning for fundus image analysis. However, many of the works do not publish their source code, and very few of them provide ready-to-use open source preprocessing tools for the community.
In ophthalmic research, retinal fundus images are often used as a resource for studying various eye diseases such as diabetic retinopathy, glaucoma and age-related macular degeneration. Consequently, there is a large amount of research on machine learning for fundus image analysis. However, many of the works do not publish their source code, and very few of them provide ready-to-use open source preprocessing tools to the community.

The Fundus Image Toolbox has been developed to address this need within the medical image analysis community. It offers a comprehensive set of tools for automated processing of retinal fundus images, covering a wide range of tasks (see Tools). The methods all accept paths to images, standard image types (e.g. images loaded with PIL, cv2, or matplotlib) and batches thereof and where possible, image batches are efficiently processed as such. This allows the tools to be seamlessly combined into a processing pipeline. The quality prediction and localization models have been developed by the authors and allow for both prediction and retraining while the other main functionalities are based on state-of-the-art methods from the literature that are applicable for inference. By providing an interface for these tasks, the toolbox aims to facilitate the development of new algorithms and models in the field of fundus image analysis. AutoMorph is the closest related work [@zhou2022], which provides a distinct and smaller set of tools for fundus image processing.
The Fundus Image Toolbox was developed to address this need within the medical image analysis community. It offers a comprehensive set of tools for automated processing of retinal fundus images, covering a wide range of tasks (see Tools). The methods all accept paths to images, standard image types (e.g., images loaded with Pillow, OpenCV, or Matplotlib) and batches thereof and where possible, image batches are efficiently processed as such. This allows the tools to be seamlessly combined into a processing pipeline. The quality prediction and localization models have been developed by the authors and allow for both prediction and retraining, while the other main functionalities are based on state-of-the-art methods from the literature that are applicable for inference. By providing an interface for these tasks, the toolbox facilitates the development of new algorithms and models in the field of fundus image analysis. AutoMorph is the closest related work [@zhou2022], which provides a distinct and smaller set of tools for fundus image processing.

# Tools
The main functionalities of the Fundus Image Toolbox are:

- Quality prediction (\autoref{fig:example}a). We trained an ensemble of ResNets and EfficientNets on the combined DeepDRiD and DrimDB datasets [@deepdrid;@drimdb] to predict the gradeability of fundus images. Both datasets are publicly available. The model ensemble achieved an accuracy of 0.78 and an area under the receiver operating characteristic curve of 0.84 on a DeepDRiD test split and 1.0 and 1.0 on a DrimDB test split.
- Fovea and optic disc localization (\autoref{fig:example}b). Prediction of fovea and optic disc center coordinates using a multi-task EfficientNet model. We trained the model on the combined ADAM, REFUGE and IDRID datasets [@adam;@refuge;@idrid], which are publicly available. On our test split, the model achieved a mean distance to the fovea and optic disc targets of 0.88 % of the image size. This corresponds to a mean distance of 3,08 pixels in the 350 x 350 pixel images used for training and testing.
- Vessel segmentation (\autoref{fig:example}c). Segmentation of blood vessels in a fundus image using an ensemble of FR-U-Nets. The ensemble achieved an average Dice score of 0.887 on the test split of the FIVES dataset [@koehler2024].
- Registration (\autoref{fig:example}d). Alignment of a fundus photograph to another fundus photograph of the same eye using SuperRetina: A keypoint-based deep learning model that produced registrations of at least acceptable quality in 98.5 % of the cases on the test split of the FIRE dataset [@liu2022].
- Circle crop. Fastly center fundus images and crop to a circle [@fu2019].
- Fovea and optic disc localization (\autoref{fig:example}b). Prediction of fovea and optic disc center coordinates using a multitask EfficientNet model. We trained the model on the combined ADAM, REFUGE and IDRID datasets [@adam;@refuge;@idrid], which are publicly available. On our test split, the model achieved a mean distance to the fovea and optic disc targets of 0.88 % of the image size. This corresponds to a mean distance of 3.08 pixels in the 350 x 350 pixel images used for training and testing.
- Vessel segmentation (\autoref{fig:example}c). The segmentation method produces a mask of blood vessels in a fundus image using an ensemble of FR-UNets. The ensemble achieved an average Dice score of 0.887 on the test split of the FIVES dataset [@koehler2024].
- Registration (\autoref{fig:example}d). Two fundus images of the same eye can be aligned using SuperRetina. The deep learning based model detects key points on the vessel trees of the two images and matches them. This results in a registered version of the second image that is aligned with the first. SuperRetina produced registrations of at least acceptable quality in 98.5 % of the cases on the test split of the FIRE dataset [@liu2022].
- Circle cropping. The OpenCV-based implementation fastly crops the circular background from a fundus image. The circle is further resized to touch the edges of the image, which centers the fundus [@fu2019].


![Examples for main functionalities of the Fundus Image Toolbox. (a.) Fovea and optic disc localization. (b.) Quality prediction. (c.) Vessel segmentation. (d.) Registration.\label{fig:example}](fig3.pdf){ width=100% }

# Acknowledgements
We thank Ziwei Huang for reviewing the package. This project was supported by the Hertie Foundation. JG received funding through the Else Kröner Medical Scientist Kolleg "ClinbrAIn: Artificial Intelligence for Clinical Brain Research”. The authors thank the International Max Planck Research School for Intelligent Systems (IMPRS-IS) for supporting SM.
We thank Ziwei Huang for reviewing the package and Murat Seçkin Ayhan for inspiring development of the quality prediction model. This project was supported by the Hertie Foundation. JG received funding through the Else Kröner Medical Scientist Kolleg "ClinbrAIn: Artificial Intelligence for Clinical Brain Research”. The authors thank the International Max Planck Research School for Intelligent Systems (IMPRS-IS) for supporting SM.

# References

0 comments on commit 7f6ec6b

Please sign in to comment.