This is the official implementation of RAUNE-Net, "RAUNE-Net: A Residual and Attention-Driven Underwater Image Enhancement Method", Wangzhen Peng, Chenghao Zhou, Runze Hu, Jingchao Cao, Yutao Liu*, International Forum of Digital Multimedia Communication(IFTC), 2023. Click here to view our poster demonstration.
Underwater image enhancement (UIE) poses challenges due to distinctive properties of the underwater environment, including low contrast, high turbidity, visual blurriness, and color distortion. In recent years, the application of deep learning has quietly revolutionized various areas of scientific research, including UIE. However, existing deep learning-based UIE methods generally suffer from issues of weak robustness and limited adaptability. In this paper, inspired by residual and attention mechanisms, we propose a more reliable and reasonable UIE network called RAUNE-Net ([rɔ:ni net]) by employing residual learning of high-level features at the network's bottle-neck and two aspects of attention manipulations in the down-sampling procedure. Furthermore, we have collected and created two datasets specifically designed for evaluating UIE methods, which contains different types of underwater distortions and degradations. The experimental validation demonstrate that our method obtains promising objective performance and consistent visual results across various underwater image test-sets compared to other eight UIE methods.
We used 1 dataset for training and 4 full-reference datasets and 3 non-reference datasets for testing. You can download these datasets by clicking links below.
- LSUI3879
- 3879 pairs of images randomly selected from LSUI [13].
- 🔗GoogleDrive
- LSUI400
- 400 pairs of images remained in LSUI [13].
- 🔗GoogleDrive
- UIEB100
- 100 pairs of images randomly selected from UIEB [8].
- 🔗GoogleDrive
- EUVP_Test515
- 515 pairs of testing images from EUVP [6].
- 🔗GoogleDrive
- OceanEx
- We gathered 40 high-quality underwater images from the official website of NOAA Ocean Exploration. Then, we applied CycleGAN to add underwater distortions and degradation styles to these images, making them the samples to be enhanced, while keeping the original high-quality images as reference images.
- 🔗GoogleDrive
- U45
- 45 real-world images from [10].
- 🔗GoogleDrive
- RUIE_Color90
- 90 real-world images from RUIE [12], which contains 3 types of color cast (i.e., greenish, blue-greenish, and bluish).
- 🔗GoogleDrive
- UPoor200
- UPoor200 is a dataset we collected and created, which consists of 200 real-world underwater images with poor visual quality. It includes distortions such as blue-green color cast, low lighting, blurriness, noise, and other types of distortions.
- 🔗GoogleDrive
we compared RAUNE-Net with eight other UIE methods (i.e., UT-UIE [13], SyreaNet [17], WaterNet [8], UGAN [3], FUnIE-GAN [6], UWCNN [7], SGUIE-Net [14], Cycle-GAN [9]).
We provide pretrained models for testing RAUNE-Net
, WaterNet
, UGAN
, and FUnIE-GAN
. Please click 🔗this link to download them. After the download is complete, please unzip and put pretrained weights into the pretrained
folder. Please see 👉here to organize the directory structure for testing.
Epoch | Dataset | PSNR | SSIM |
---|---|---|---|
85 | LSUI400 | 26.812 | 0.876 |
90 | EUVP_Test515 | 26.331 | 0.845 |
95 | UIEB100 | 22.751 | 0.879 |
90 | OceanEx | 22.728 | 0.876 |
Epoch | Dataset | PSNR | SSIM |
---|---|---|---|
85 | LSUI400 | 26.688 | 0.874 |
90 | EUVP_Test515 | 25.285 | 0.833 |
80 | UIEB100 | 22.279 | 0.868 |
50 | OceanEx | 22.132 | 0.887 |
Epoch | Dataset | PSNR | SSIM |
---|---|---|---|
85 | LSUI400 | 25.117 | 0.846 |
85 | EUVP_Test515 | 23.636 | 0.805 |
90 | UIEB100 | 21.368 | 0.825 |
60 | OceanEx | 22.436 | 0.822 |
Epoch | Dataset | PSNR | SSIM |
---|---|---|---|
85 | LSUI400 | 23.272 | 0.818 |
95 | EUVP_Test515 | 24.077 | 0.794 |
85 | UIEB100 | 19.614 | 0.813 |
50 | OceanEx | 20.448 | 0.855 |
Single Epoch Test:
python test_raunenet.py \
--name test \
--checkpoint_dir pretrained \
--epoch 95 \
--data_dir <PATH_TO_YOUR_TEST_DATASET> \
--test_name <YOUR_TESTSET_NAME> \
--num_down 2 \
--num_blocks 30 \
--result_dir results/raunet_test01
Multiple Epochs Test:
bash scripts/test_raunenet.sh
# name: test
# checkpoint_dir: pretrained
# num_down: 2
# num_blocks: 30
# use_att_up: false
Multiple Epochs Test and Assessment:
bash scripts/test_assess_raunenet.sh
# name: test
# checkpoint_dir: pretrained
# num_down: 2
# num_blocks: 30
# use_att_up: false
Single Epoch Test:
python test_waternet.py \
--name test \
--checkpoint_dir pretrained \
--data_dir <PATH_TO_YOUR_TEST_DATASET> \
--epoch 95 \
--test_name <YOUR_TESTSET_NAME> \
--result_dir results/waternet01
Multiple Epochs Test:
bash scripts/test_waternet.sh
# name: test
# checkpoint_dir: pretrained
Multiple Epochs Test and Assessment:
bash scripts/test_assess_waternet.sh
# name: test
# checkpoint_dir: pretrained
Single Epoch Test:
python test_ugan.py \
--name test \
--checkpoint_dir pretrained \
--data_dir <PATH_TO_YOUR_TEST_DATASET> \
--epoch 95 \
--test_name <YOUR_TESTSET_NAME> \
--result_dir results/ugan01
Multiple Epochs Test:
bash scripts/test_ugan.sh
# name: test
# checkpoint_dir: pretrained
Multiple Epochs Test and Assessment:
bash scripts/test_assess_ugan.sh
# name: test
# checkpoint_dir: pretrained
Single Epoch Test:
python test_funiegan.py \
--name test \
--checkpoint_dir pretrained \
--data_dir <PATH_TO_YOUR_TEST_DATASET> \
--epoch 95 \
--test_name <YOUR_TESTSET_NAME> \
--result_dir results/funiegan01
Multiple Epochs Test:
bash scripts/test_funiegan.sh
# name: test
# checkpoint_dir: pretrained
Multiple Epochs Test and Assessment:
bash scripts/test_assess_funiegan.sh
# name: test
# checkpoint_dir: pretrained
We provide a python program in command-line interface (CLI) to calculate PSNR
and SSIM
. You can see example below to use this simple tool.
# The input_dir may like 'results/raunet_test01/LSUI400/single/predicted/'
python calc_psnr_ssim.py \
--input_dir <YOUR_ENHANCED_IMAGES_DIRECTORY> \
--refer_dir <YOUR_REFERENCE_IMAGES_DIRECTORY> \
--output_dir <YOUR_OUTPUT_DIRECTORY> \
--resize \
--width 256 --height 256
We provide four programs for training with command-line interface. They are:
Please see them in details.
For example, you can train a model of RAUNE-Net like this:
python train_raunenet.py \
--cfg_file configs/train_lsui3879.yaml \
--name LSUI3879_PCONT1_SSIM1_SCONT1 \
--num_epochs 100 \
--batch_size 8 \
--lr 0.0001 \
--pcont_weight 1 \
--ssim_weight 1 \
--scont_weight 1 \
--seed 2023 \
--num_down 2 \
--num_blocks 3
If you have any question about this project, please contact fansuregrin
(or Wangzhen Peng
) through wangzhenpeng@stu.ouc.edu.cn or pwz113436@gmail.com! Btw, any pull request is welcome if you are interested in this project:)
[3] Fabbri, C., Islam, M.J., Sattar, J.: Enhancing underwater imagery using generative adversarial networks. In: 2018 IEEE International Conference on Robotics and Automation (ICRA). pp. 7159–7165. IEEE (2018)
[6] Islam, M.J., Xia, Y., Sattar, J.: Fast underwater image enhancement for improved visual perception. IEEE Robotics and Automation Letters 5(2), 3227–3234 (2020)
[7] Li, C., Anwar, S., Porikli, F.: Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recognition 98, 107038 (2020)
[8] Li, C., Guo, C., Ren, W., Cong, R., Hou, J., Kwong, S., Tao, D.: An underwater image enhancement benchmark dataset and beyond. IEEE Transactions on Image Processing 29, 4376–4389 (2019)
[9] Li, C., Guo, J., Guo, C.: Emerging from water: Underwater image color correction based on weakly supervised color transfer. IEEE Signal processing letters 25(3), 323–327 (2018)
[10] Li, H., Li, J., Wang, W.: A fusion adversarial underwater image enhancement network with a public test dataset. arXiv preprint arXiv:1906.06819 (2019)
[12] Liu, R., Fan, X., Zhu, M., Hou, M., Luo, Z.: Real-world underwater enhancement: Challenges, benchmarks, and solutions under natural light. IEEE Transactions on Circuits and Systems for Video Technology 30(12), 4861–4875 (2020)
[13] Peng, L., Zhu, C., Bian, L.: U-shape transformer for underwater image enhancement. IEEE Transactions on Image Processing (2023)
[14] Qi, Q., Li, K., Zheng, H., Gao, X., Hou, G., Sun, K.: Sguie-net: Semantic attention guided underwater image enhancement with multi-scale perception. IEEE Transactions on Image Processing 31, 6816–6830 (2022)
[17] Wen, J., Cui, J., Zhao, Z., Yan, R., Gao, Z., Dou, L., Chen, B.M.: Syreanet: A physically guided underwater image enhancement framework integrating synthetic and real images. arXiv preprint arXiv:2302.08269 (2023)
@InProceedings{10.1007/978-981-97-3623-2_2,
author="Peng, Wangzhen
and Zhou, Chenghao
and Hu, Runze
and Cao, Jingchao
and Liu, Yutao",
editor="Zhai, Guangtao
and Zhou, Jun
and Ye, Long
and Yang, Hua
and An, Ping
and Yang, Xiaokang",
title="RAUNE-Net: A Residual and Attention-Driven Underwater Image Enhancement Method",
booktitle="Digital Multimedia Communications",
year="2024",
publisher="Springer Nature Singapore",
address="Singapore",
pages="15--27",
abstract="Underwater image enhancement (UIE) poses challenges due to distinctive properties of the underwater environment, including low contrast, high turbidity, visual blurriness, and color distortion. In recent years, the application of deep learning has quietly revolutionized various areas of scientific research, including UIE. However, existing deep learning-based UIE methods generally suffer from issues of weak robustness and limited adaptability. In this paper, inspired by residual and attention mechanisms, we propose a more reliable and reasonable UIE network called RAUNE-Net by employing residual learning of high-level features at the network's bottle-neck and two aspects of attention manipulations in the down-sampling procedure. Furthermore, we collect and create two datasets specifically designed for evaluating UIE methods, which contains different types of underwater distortions and degradations. The experimental validation demonstrates that our method obtains promising objective performance and consistent visual results across various real-world underwater images compared to other eight UIE methods. Our example code and datasets are publicly available at https://github.com/fansuregrin/RAUNE-Net.",
isbn="978-981-97-3623-2"
}
Licensed under a Creative Commons Attribution-NonCommercial 4.0 International for Non-commercial use only. Any commercial use should get formal permission first.