A Survey of Deep Learning-Based Image Restoration ... - MDPI

44
Citation: Karavarsamis, S.; Gkika, I.; Gkitsas, V.; Konstantoudakis, K.; Zarpalas, D. A Survey of Deep Learning-Based Image Restoration Methods for Enhancing Situational Awareness at Disaster Sites: The Cases of Rain, Snow and Haze. Sensors 2022, 22, 4707. https:// doi.org/10.3390/s22134707 Academic Editor: Panagiotis Kasnesis, Dimitrios G. Kogias, Stelios A. Mitilineos and Charalampos Z. Patrikakis Received: 10 May 2022 Accepted: 16 June 2022 Published: 22 June 2022 Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affil- iations. Copyright: © 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https:// creativecommons.org/licenses/by/ 4.0/). sensors Article A Survey of Deep Learning-Based Image Restoration Methods for Enhancing Situational Awareness at Disaster Sites: The Cases of Rain, Snow and Haze Sotiris Karavarsamis , Ioanna Gkika , Vasileios Gkitsas , Konstantinos Konstantoudakis and Dimitrios Zarpalas * Visual Computing Lab, Information Technologies Institute, Centre for Research and Technology Hellas (CERTH), 57001 Thessaloniki, Greece; [email protected] (S.K.); [email protected] (I.G.); [email protected] (V.G.); [email protected] (K.K.) * Correspondence: [email protected]; Tel.: +30-2310-464-160 Abstract: This survey article is concerned with the emergence of vision augmentation AI tools for enhancing the situational awareness of first responders (FRs) in rescue operations. More specifically, the article surveys three families of image restoration methods serving the purpose of vision aug- mentation under adverse weather conditions. These image restoration methods are: (a) deraining; (b) desnowing; (c) dehazing ones. The contribution of this article is a survey of the recent literature on these three problem families, focusing on the utilization of deep learning (DL) models and meeting the requirements of their application in rescue operations. A faceted taxonomy is introduced in past and recent literature including various DL architectures, loss functions and datasets. Although there are multiple surveys on recovering images degraded by natural phenomena, the literature lacks a comprehensive survey focused explicitly on assisting FRs. This paper aims to fill this gap by presenting existing methods in the literature, assessing their suitability for FR applications, and providing insights for future research directions. Keywords: deraining; dehazing; desnowing; deep learning; deep neural networks 1. Introduction First responders (FRs) are trained professionals that provide aid in case of emergency and include medics, firefighters, law enforcement and civil protection officials. During op- erations, FRs often come across many stressful situations, experience extreme safety risks and may meet an additional inevitable obstacle—adverse weather conditions. Weather conditions that reduce visibility, such as rain, snow or haze, prevent FRs from exhibiting expected performance under normal conditions while they simultaneously increase the risk ofinjury. Moreover, adverse weather conditions may become an obstacle on the perfor- mance of any computer vision (CV) tool employed by FRs during operations. Typically, one would observe a degraded performance in these tools that handle raw visual data that is itself degraded by artifacts caused by the utilization of the tools under adverse weather conditions. In recent years, thanks to deep learning (DL), CV algorithms have been proposed for solving various vision tasks which, in the real world, are either difficult, time consuming or impossible for a human to carry out. In such tasks, DL-based CV algorithms have shown impressive results. Due to their great performance, recent works have concentrated on developing them further and making them more accurate and possibly faster. To make a DL model more accurate, many researchers have focused on designing sophisticated architectures. This led to the requirement of training large models that can achieve perfor- mance close to that of humans, but with such a computational complexity that would be impossible to be implemented in devices with constrained resources or in applications in Sensors 2022, 22, 4707. https://doi.org/10.3390/s22134707 https://www.mdpi.com/journal/sensors

Transcript of A Survey of Deep Learning-Based Image Restoration ... - MDPI

Citation: Karavarsamis, S.; Gkika, I.;

Gkitsas, V.; Konstantoudakis, K.;

Zarpalas, D. A Survey of Deep

Learning-Based Image Restoration

Methods for Enhancing Situational

Awareness at Disaster Sites: The

Cases of Rain, Snow and Haze.

Sensors 2022, 22, 4707. https://

doi.org/10.3390/s22134707

Academic Editor: Panagiotis

Kasnesis, Dimitrios G. Kogias, Stelios

A. Mitilineos and Charalampos Z.

Patrikakis

Received: 10 May 2022

Accepted: 16 June 2022

Published: 22 June 2022

Publisher’s Note: MDPI stays neutral

with regard to jurisdictional claims in

published maps and institutional affil-

iations.

Copyright: © 2022 by the authors.

Licensee MDPI, Basel, Switzerland.

This article is an open access article

distributed under the terms and

conditions of the Creative Commons

Attribution (CC BY) license (https://

creativecommons.org/licenses/by/

4.0/).

sensors

Article

A Survey of Deep Learning-Based Image RestorationMethods for Enhancing Situational Awareness atDisaster Sites: The Cases of Rain, Snow and HazeSotiris Karavarsamis , Ioanna Gkika , Vasileios Gkitsas , Konstantinos Konstantoudakisand Dimitrios Zarpalas *

Visual Computing Lab, Information Technologies Institute, Centre for Research and Technology Hellas (CERTH),57001 Thessaloniki, Greece; [email protected] (S.K.); [email protected] (I.G.); [email protected] (V.G.);[email protected] (K.K.)* Correspondence: [email protected]; Tel.: +30-2310-464-160

Abstract: This survey article is concerned with the emergence of vision augmentation AI tools forenhancing the situational awareness of first responders (FRs) in rescue operations. More specifically,the article surveys three families of image restoration methods serving the purpose of vision aug-mentation under adverse weather conditions. These image restoration methods are: (a) deraining;(b) desnowing; (c) dehazing ones. The contribution of this article is a survey of the recent literatureon these three problem families, focusing on the utilization of deep learning (DL) models and meetingthe requirements of their application in rescue operations. A faceted taxonomy is introduced inpast and recent literature including various DL architectures, loss functions and datasets. Althoughthere are multiple surveys on recovering images degraded by natural phenomena, the literaturelacks a comprehensive survey focused explicitly on assisting FRs. This paper aims to fill this gapby presenting existing methods in the literature, assessing their suitability for FR applications, andproviding insights for future research directions.

Keywords: deraining; dehazing; desnowing; deep learning; deep neural networks

1. Introduction

First responders (FRs) are trained professionals that provide aid in case of emergencyand include medics, firefighters, law enforcement and civil protection officials. During op-erations, FRs often come across many stressful situations, experience extreme safety risksand may meet an additional inevitable obstacle—adverse weather conditions. Weatherconditions that reduce visibility, such as rain, snow or haze, prevent FRs from exhibitingexpected performance under normal conditions while they simultaneously increase therisk of injury. Moreover, adverse weather conditions may become an obstacle on the perfor-mance of any computer vision (CV) tool employed by FRs during operations. Typically,one would observe a degraded performance in these tools that handle raw visual datathat is itself degraded by artifacts caused by the utilization of the tools under adverseweather conditions.

In recent years, thanks to deep learning (DL), CV algorithms have been proposed forsolving various vision tasks which, in the real world, are either difficult, time consuming orimpossible for a human to carry out. In such tasks, DL-based CV algorithms have shownimpressive results. Due to their great performance, recent works have concentrated ondeveloping them further and making them more accurate and possibly faster. To makea DL model more accurate, many researchers have focused on designing sophisticatedarchitectures. This led to the requirement of training large models that can achieve perfor-mance close to that of humans, but with such a computational complexity that would beimpossible to be implemented in devices with constrained resources or in applications in

Sensors 2022, 22, 4707. https://doi.org/10.3390/s22134707 https://www.mdpi.com/journal/sensors

Sensors 2022, 22, 4707 2 of 44

which CV algorithms need to perform in real time. This challenge led many scientists toresearch ways to achieve outstanding performance with models that contain a considerablylow number of parameters. At the same time, evolution in hardware has given us the abilityto use numerous sensors, processors and energy sources in small and lightweight gadgets.As a consequence, portable devices have paved the way for extending CV applications innumerous domains that require the existence of light equipment and quick response fromthe model. These devices can serve the enhancement of the FRs’ vision.

On the other hand, the development and deployment of CV systems for facilitatingthe FRs’ vision under adverse weather conditions (herein, we focus on rain, snow and haze)pose several challenges that need to be addressed. First, there is a lack of assortment inthe imagery provided by datasets which are focused on a specific task. Although there areseveral tasks directly related to the enhancement of visual recognition under adverse condi-tions (such as deraining, desnowing and dehazing), virtually none of them contains scenesrepresentative of the FRs’ working environment. Additionally, even in the design of theexisting datasets, it is impossible for the designer to simultaneously capture the same imagewith and without the visual artifacts caused by the adverse weather conditions. However,being able to do so is a requirement in standard supervised end-to-end approaches thatreceive pairs of clear and noisy images. In such models, the noisy image is fed at one endof the model and the clean image is output at the other end. Thus the models proposed inthe literature are either trained in an unsupervised manner having only the noisy images,or they use pairs of synthetic noisy images with their ground truth (clear) images. A directdrawback when training models on synthetic data, however, is that there is a domain gapamong the real and the synthetic datasets. That is, in such a case, it may be difficult forimage denoising models to generalize well on real-world data.

Another important issue in the application of CV systems for the facilitation of theFRs’ vision, is the required processing time. When employing a DL model to enhance thesituational awareness of FRs, the model should process visual data in near real-time andoften in infrastructure-less environments (e.g., environments with an absence of networkconnectivity and with power constraints). In the related literature, only a small numberof research works propose deep neural network models that are lightweight architecturesable to operate well in power-constrained infrastructures. Under these conditions, there isa requirement for the DL-based augmented vision tools to simultaneously be lightweightand adequately accurate.

Owing to the constraint on real-time processing and the requirement for lightweightmodels, there is an urgent need for solutions that can handle more than one task simultane-ously. These solutions are often called unified solutions, and we will be using this term inour article. So far, there has been a great amount of research works on image restorationtechniques focusing on only one specific adverse weather condition. In the real world,however, during a rescue operation (e.g., one taking more than one day to complete) theFRs could meet any weather condition. This means that the CV tool that could help theFRs’ vision should be able to adapt to the visibility changes of the scene, or it should beable to switch between modes of operation specific to one particular weather condition.Importantly, the unification of such models has already been addressed in the literature (atleast regarding the three problems that this article is concerned with). Hence, to the best ofour knowledge, there are a few models that suggest unified solutions.

In ideal weather conditions, CV tools can offer the means for FRs to keep themselvessafe and at the same time locate and save victims robustly and reliably. However, rain,snow and haze can degrade the performance of such CV systems. Consequently, thesesystems do not provide optimal assistance to FRs. In this survey article, we explore theuse of DL-based image denoising methods tailored towards the removal of atmosphericartifacts in captured images. These DL models could be applied for the augmentation ofthe FRs’ vision. We specifically require these inferences to be better possible under rain,snow and haze conditions.

Sensors 2022, 22, 4707 3 of 44

Clear vision can improve disaster response in multiple ways, including damage detection [1],traffic management [2,3], UAV-driven reconnaissance [4,5] and flood detection [6]. Objectdetection is a computer vision task which is usually impeded by adverse weather condi-tions; for instance, see the works of Rothmeier and Huber [7], Pfeuffer and Dietmayer [8],Hasirlioglu and Riener [9], Chaturvedi et al. [10] among many research papers on this topic.Morrison et al. [11] conducted a user study on the requirements of first responders in termsof communication systems. Although this study focuses on communication infrastructurerequirements, nevertheless it points out a need for robust technological solutions that mayimprove the capabilities of first responders in adverse weather conditions. However, firstresponders often need to operate in adverse weather conditions, which can degrade bothvision and the efficacy of vision-based algorithms. These can include rain, snow, haze,darkness, smoke, dust and others. From a visual computing point of view, these can bedivided in two broad categories:

• Conditions which degrade vision without fully obstructing it: such as rain, snowand haze. As objects or people are visible through such conditions both to the nakedeye and RGB camera sensors, visual computing algorithms can be used to restore suchimages and improve visibility.

• Conditions which fully obstruct some or all parts of the field of view: such are totaldarkness, heavy smoke or dense dust. Such conditions beyond the capabilities of RGBsensors and computer vision algorithms to restore, necessitating other modalities andapproaches, such as infrared sensors.

The present survey focuses specifically on the first category, that of conditions thatonly partially obscure vision, examining in particular the cases of rain, snow and haze.To that end, it presents and categorizes such image restoration methods that can improveclarity of vision in such adverse conditions, both for the responders themselves and for anycomputer algorithms that they may employ.

Due to the extensive research on the image restoration task, there are numerousresearch works devoted to the deraining, desnowing and dehazing tasks. While previoussurveys have focused on general-purpose image denoising, including deblurring and super-resolution (such as the recently published work by Su et al. [12]), an up-to-date survey ofimage restoration methods addressing specifically adverse weather conditions and theirapplicability for disaster response operations, has been missing. To this end, the presentwork focuses on image restoration methods that could serve the purpose of augmentingthe sight of FRs as the means to increase their situational awareness. The augmentationof vision scenarios that we consider regard the existence of adverse weather conditions.That implies that we have focused mostly on architectures that provide accurate resultsand could be implemented on portable devices that can provide real-time processing.However, non-real-time or near-real-time methods are also taken into consideration, sincethey may possess several properties or ideas which are desirable in the implementation ofnew, lightweight models in the future. In our article, we provide an up-to-date survey ofthe deraining, desnowing and dehazing methods, being employed as important tools foraugmenting the sight of FRs and for providing higher-quality input to CV algorithms thatare required to make critical inferences.

The key contributions of this survey are as follows:

• We survey the research literature on the deraining, desnowing and dehazing methodsthat employ DL-based architectures. To the best of our knowledge, this work is thefirst survey of image restoration methods in adverse conditions for assisting FRssituational awareness.

• We provide a faceted taxonomy of the abovementioned image denoising methods inadverse weather conditions in terms of their technical attributes.

• We compare the existing algorithms in terms of quantitative metrics and processingtime in order to decide the appropriateness of each method for the specific task offacilitating the FRs’ vision.

Sensors 2022, 22, 4707 4 of 44

Our work aims to conduct a detailed and comprehensive survey on single-imagerestoration methods in adverse weather conditions, namely rain, haze and snow. We expectthat this study can contribute toward understanding the current trends in the existingmethods, their applicability and limitations in an applications level for FRs. Additionally,the challenges that arise from these limitations provide valuable insights for future researchdirections. The article is structured as follows. In Section 1, we motivate the use of imagedenoising methods for the removal of rain, haze and snow under the more general contextof augmenting the vision capacity of FRs operating in rescue missions. In Section 2, wepresent the existing datasets for each task. In Sections 3–5 we survey the literature of DL-based deraining, desnowing and dehazing methods respectively. For each task, we presenta technical taxonomy of the existing literature based on the architecture of the models bothfor single and multi-images. In Section 6, we present the most common quantitative metricsand the results of the models proposed in the literature in terms of the quantitative metricsand processing time. Last, in Section 7 we summarize the paper.

2. Datasets

The availability of datasets for the deraining, desnowing and dehazing tasks is impor-tant for the development and cross-evaluation of methods built around these problems.Creating large-scale datasets for training DL models from scratch is not an easy task, there-fore, pre-existing publicly available datasets are very important, because they acceleratethe development of algorithms and researchers are not required to build their own datasetsthemselves. In this section we survey the available datasets for the deraining, desnow-ing and dehazing problems. Tables 1–3 summarize the proposed datasets for deraining,desnowing and dehazing, respectively.

2.1. Deraining Datasets

A plethora of datasets have become available for enabling the development and evalu-ation of deraining methods. When designing new deraining methods, a major problemthat is faced is the infeasibility of simultaneously capturing pairs of rainy and clean images,thereby imposing difficulties in training supervised models. A notable example refersto the deep end-to-end models, where a pipeline learns to map rainy images fed at theone end to clean images output at the other end. This difficulty has motivated the ideaof generating synthetic deraining image datasets. Based on this idea, the designer of thedataset injects noise in the images by means of a synthetic rain streak generation algorithm.Therefore, by knowing where each rain streak is placed, it is easy to obtain a rain streakmask image that can provide a supervisory signal to machine learning (ML) methods.A limitation that is often met when reusing synthetic datasets for training deraining models,is that the injected noise is often not realistic enough for the trained models to be able togeneralize well on real-world images with naturally generated rain streaks. Finally, apartfrom the datasets containing pairs of clean and synthetic rainy images, another class ofdatasets contain only real-world images for which ground-truth is unknown. Derainingdatasets that do not contain ground-truth are more suitable for unsupervised derainingmethods (for example, see [13–15]), or for semi-supervised deraining methods which canbe trained on a combination of a paired dataset and a dataset with no ground-truth (for anexample, see [16]).

The RainX datasets (where X is an integer value from a fixed set of integers) aresome of the most commonly used datasets in the evaluation of deraining algorithms.The Rain12600 dataset is a synthetic rain dataset that was originally contributed byFu et al. [17]. The dataset contains a total of 14,000 pairs of rainy and clean images ofthe same scene out of 900 original scenes. To create the dataset, originally 1000 images ofscenes were selected from the UCID [18], the BSD [19] dataset and Google image search.For each single image out of the available 1000 ones, 14 rainy images are syntheticallygenerated. Some authors reusing this dataset keep 90% of the available 14,000 14-tuples fortraining models (amounting to 12,600 images), and keep the rest of the images for testing

Sensors 2022, 22, 4707 5 of 44

(amounting to 1400 images). The Rain12000 dataset is another synthetic rain dataset thatcontains 12,000 images. The dataset was originally introduced by Zhang and Patel [20].Depending on the rain synthesis occurring in each image, an image is assigned one of threepossible labels, namely light-rain, medium-rain and heavy-rain. The Rain1400 dataset istraced back to the publication of Fu et al. [17]. This dataset is a subset of 100 14-tuplesfrom the Rain12600 dataset. The number of examples in this dataset amounts to 10% ofthe total example images provided by Rain12000. The dataset is commonly referred to asRain1400, and is a subset of the Rain12600 dataset. The Rain800 dataset was contributedby Zhang et al. [21]. The dataset is split into a training set of 700 synthetic images and atesting set of 100 images. The Rain12 dataset [22] contains 12 synthetic images which areinjected with one type of rain streak.

The Test100 and Test1200 datasets were contributed for the purpose of conductingmodel validation for deraining methods. More specifically, the Test100 dataset was in-troduced by Zhang et al. [21]. Test100 contains the last 100 examples of the Rain800dataset. The Test1200 dataset was contributed in the work of Zhang et al. [20]. It contains1200 images of rainy scenes.

The RainTrainH and RainTrainL datasets make a distinction among scenes with heavyrain and scenes with light rain. The RainTrainH and RainTrainL datasets were contributedin the work by Zhang and Patel [20].

The Rain100H, Rain100L and Rain200H are three of the most commonly used datasetson the deraining problem. Specifically, the Rain100H dataset was contributed in the workof Yang et al. [23]. It is a synthetic rain image dataset, originally composed of samples fromthe BSD200 dataset [19]. The Rain100L dataset, like the Rain100H dataset, was contributedby Yang et al. [23]. It is a synthetic rainy image dataset that comprises five differentorientations of rain streaks. This dataset can be specifically used to test the ability of a deepneural deraining model to learn the regularity of rain streaks. The Rain200H dataset wascontributed by Yang et al. [24]. The dataset contains 1800 paired training images and atesting set of 200 paired images.

The Rain in Driving (RID) and Rain in Surveillance (RIS) datasets resemble two usecases of image deraining algorithms, making both of them suitable for validating algorithmswhose target application partially or fully coincides with the two use cases. The RID dataset(contributed by Li et al. [25]) is a set of 2495 rainy images extracted from driving videos ofa high resolution. The dataset demonstrates the veiling effect of rain streaks on the cameralens. The dataset was captured under real multiple drives and traffic locations. Finally,the dataset contains ground-truth for four objects (namely, car, person, bus and bicycle).The RIS dataset (also contributed by Li et al. [25]) contains a total of 2048 real rainy imagesfrom lower-resolution surveillance video cameras. The dataset demonstrates the “rain andmist” scenario where the surveillance cameras introduce a fog-like effect and a mist-likeeffect caused by rain falling near the cameras.

The DAWN/Rainy dataset (contributed by Kenk and Hassaballah [26]) contains200 images of outdoor rainy scenes. The dataset can be used for the single image derainingtask. It can also be used for an object detection task where the detected targets can bematched directly with object ground-truth, which is available for each image. A limitationof the provided object ground-truth is that the provided ground-truth objects are selectiveof a small target set of objects.

The NTURain dataset was contributed by Chen et al. [27]. It is a synthetic rain datasetcomprising a total of eight video sequences. The frame-count of each video sequence rangesbetween 200 and 300 frames. A total of four of the videos contain short scenes capturedfrom a panning and unstable camera. The rest of the four videos were captured by a fastmoving camera.

The SPA-Data dataset by Wang et al. [28] contains 29,500 rainy or clean image pairs.The data are split into a training set of 28,500 examples and a testing set of 1000 examples.These rainy/clean image pairs are generated from a large dataset of 170 videos of real rain.The videos cover the topics of urban scenes, suburb scenes and outdoor fields.

Sensors 2022, 22, 4707 6 of 44

Table 1. Listing of datasets used for the deraining task.

Dataset Synthetic (S)/Real (R) Indoor (I)/Outdoor (O) Pairs YearRain12600 [17] S O 14,000 2017Rain12000 [20] S O 12,000 2018Rain1400 [17] S O 1400 2017Rain800 [21] R O 800 2020Rain12 [22] S O 12 2016Test100 [21] S O 100 2020

Test1200 [20] S O 1200 2018RainTrainH [20] S O 1800 2018RainTrainL [20] S O 200 2018Rain100H [23] S O 100 2020Rain100L [23] S O 100 2020Rain200H [24] S O 2000 2017

RID [25] R O 2495 2019RIS [25] R O 2048 2019

DAWN/Rainy [26] R O 200 2020NTURain [27] S O 8 (videos) 2018SPA-Data [28] R O 29,500 2019

2.2. Desnowing Datasets

A list of benchmark datasets has been compiled for use in the image desnowingproblem. The image desnowing problem has received a limited focus by researchers,in comparison to other problems such as the deraining and dehazing problems. In thissection, we describe four important datasets which have been used to assist the develop-ment and evaluation of desnowing methods.

The Snow-100K dataset contributed by Liu et al. [29] comprises both synthetic snowyimages and realistic images, all downloaded from the Flickr service. In particular, the datasetoffers 100,000 synthetic snow images paired with corresponding snow free ground-truthimages. The data are split in a training set and a testing set of 50,000 images each. A totalof 1329 realistic snowy images are provided along with the synthetic images. The datasetconsists of three subsets with small snow particles, medium-size and small-size snowparticles, and combined small, medium and large snow particles. Each subset containsaround 33,000 images.

The Snow Removal in Realistic Scenario (SRRS) dataset is a dataset that considers theveiling effect of snow, consisting of 15, 000 artificially synthesized images and 1000 real-lifesnow images downloaded from the internet. The popular RESIDE dataset is used topopulate the SRRS dataset with images, synthesizing the veiling effect in each one of theseimages. To provide labeled data of snow particle appearance, various types of snow aresynthesized by means of image processing software and labels are assigned to each particle.In terms of a training and validation set, 2500 images are chosen randomly for modeltraining and 1000 images are kept for testing models. The same partition of the images isconsidered to render the additional case where the veiling effect is absent.

The Comprehensive Snow Dataset (CSD), originally proposed by Chen et al. [30], con-tains 10,000 synthetic images, borrowed from the well-known RESIDE dataset. A snowflakeand snow streak synthesis algorithm is applied to generate synthetic images from the cleanimages. The variety in the data is controlled in terms of three factors, namely (a) the levelof transparency of snow streaks; (b) the size and the location of snow streaks. The goal is togenerate realistic looking image samples.

The SITD dataset was contributed in the work of Li et al. [31]. It contains 3000 snowyimages, snow-free images and snowflake images. The images were generated from50 videos that were collected from the web and 100 videos that were recorded by theauthors. White snow particles were generated synthetically based on a snow model andwere injected in the images. The dataset also promotes the variety of its samples by consid-ering: (a) the density of snowflakes; (b) the shapes of snow particles; (c) the transparency

Sensors 2022, 22, 4707 7 of 44

of snowflakes; (d) the time of the day at which samples were captured; (e) the scene thatwas captured.

Table 2. Listing of datasets used for the desnowing task.

Dataset Synthetic (S)/Real (R) Indoor (I)/Outdoor (O) Pairs YearSnow-100K [29] S & R O 100,000+ 2018

SRRS [32] S & R I & O 16,000 2020CSD [30] S I & O 10,000 2021SITD [31] S O 3000 2019

2.3. Dehazing Datasets

A great amount of datasets focusing on the dehazing task has been found in theliterature. Most of them are proposed for training supervised learning models and thereforecontain pairs of hazy and haze-free images. Since it is impossible to capture the exact sameimage with and without haze simultaneously, researchers have approached this problem inthree different ways. The first one is using datasets in which haze is synthetic and has beenadded to clear images based on mathematical models; the second one is using datasets inwhich haze is generated by a professional haze machine, and the third and most recent oneis using datasets where both hazy and clear images are real, but are not captured at the sametime. In this section, we describe the most popular datasets used for the dehazing task.

Tarel et al. [33] released the first synthetic dataset for single image dehazing, namedFoggy Road Image DAtabase (FRIDA). It consists of 72 pairs of synthetic images with andwithout fog, captured from virtual urban road scene using the SiVIC™ software. For eachone of the 18 clear images, four foggy ones were generated synthetically. In order togenerate the foggy images, four different types of synthetic fog were used. The FRIDA2dataset was introduced by Tarel et al. [34]. It consists of 264 pairs of synthetic images withand without fog, captured from virtual diverse road scenes, where for each clear imagefour synthetic foggy ones were generated. The synthetic images were generated in thesame way as in FRIDA.

Some years later, El et al. [35] released the CHIC dataset. It is a benchmark datasetcreated using two indoor scenes in a controlled environment that were captured both inhazy and clear conditions. In order to create haze, a haze machine was used and producednine different levels of haze density. The clear images and their corresponding nine hazyimages are also accompanied by some known parameters like the distance between eachobject and the camera, the local scene depth, etc.

One of the most widely used datasets for single image dehazing is the Realistic SingleImage Dehazing (RESIDE) dataset [36]. RESIDE is a benchmark dataset that consists of twoversions: the Standard and the extended (RESIDE-β) one. The standard version includes:(a) an Indoor Training Set (ITS) which contains over 10,000 pairs of hazy and clear indoorimages, where for each clear image 10 synthetic hazy ones were generated; (b) a SyntheticObjective Testing Set (SOTS), which contains 500 pairs of hazy and clear indoor imagesproduced in the same way as in ITS; (c) a Hybrid Subjective Testing Set (HSTS), whichcontains: 10 pairs of real, clean, outdoor images and the generated hazy ones and 10 real,outdoor hazy images. The RESIDE-β includes: (a) an Outdoor Training Set (OTS) whichcontains over 70,000 pairs of hazy and clear outdoor images, where for each clear image 35synthetic hazy ones were generated; (b) a Real-world Task-driven Testing Set (RTTS), whichcontains over 4000 real hazy images annotated with object categories and bounding boxes.

Other benchmark datasets include the D-HAZY [37], I-HAZE [38], O-HAZE [39],DENSE-HAZE [40] and NH-HAZE [41] datasets proposed by Ancuti et al. The D-HAZY [37]dataset contains over 1400 pairs of hazy and clear indoor images. It is built on the Mid-delbury [42] and NYU Depth [43] datasets, which are datasets that provide clear imagesand their corresponding depth information. The hazy images were synthesized by takinginto consideration depth information and by using the physics-based model presented inSection 6.3.5. The I-HAZE [38] dataset contains 35 pairs of hazy and clear indoor images.

Sensors 2022, 22, 4707 8 of 44

Here, both hazy and clear images are real and captured under the same illumination con-ditions. Haze is produced artificially by a professional haze machine and all the imageswere taken in a controlled environment. The O-HAZE [39] and the DENSE-HAZE [40]datasets contain pairs of hazy and clear outdoor images. DENSE-HAZE [40], which con-tains 33 pairs of images can be considered as the extension of the O-HAZE [39] dataset,which contains 45 pairs of images, in the way that it contains a much denser and challeng-ing haze than O-HAZE. The images in both datasets were captured in the same way as inI-HAZE. The NH-HAZE [41] dataset contains 55 pairs of hazy and clear outdoor images,captured in the same way as in I-HAZE, O-HAZE and DENSE-HAZE datasets, but herethe haze in the images is non-homogeneous.

The Haze Realistic Dataset (HAZERD) [44] is a benchmark dataset that contains14 clear, outdoor images accompanied by their corresponding depth maps estimated byfusing structure from motion and LIDAR. Using the depth map of a clear image, theysynthesized five hazed ones, where each one of them has different levels of haze density.

Zhao et al. [45] recently released the BEnchmark Dataset for Dehazing Evaluation(BeDDE). BeDDE is a real-world outdoor benchmark dataset where both hazy and clearimages are real and consists of two versions: the standard BeDDE and the EXtension ofthe BeDDE (exBeDDE). The standard BeDDE contains 208 pairs of hazy and clear images,collected from 23 provincial capital cities in China. One clear and more than one hazyimages were captured from the same place for each city. The images are accompaniedby manually labeled masks that provide the regions of interest and the hazy images arealso manually sorted into three categories on the basis of their haze density levels (“light”,“medium” and “heavy”). The exBeDDE contains 167 hazy images from 12 cities selectedfrom the BeDDE dataset accompanied by 1670 dehazed images. For each hazy image10 representative models were selected for producing the corresponding dehazed images,and for each dehazed result a subjective score from people is provided for assessing theperformance of the dehazing evaluation metrics.

Zheng et al. [46] introduced the 4KID dataset, a 4K resolution (3840× 2160) dataset,which consists of 10,000 pairs of hazy and clear images extracted from 100 videos. The hazewas synthesized using the physics-based model presented in Section 6.3.5 and a translationmodule that translates the hazy images from the synthetic to the real domain.

Recently, the first REal-world VIdeo DEhazing (REVIDE) dataset [47] was released. Ithas given the opportunity to further research the video dehazing problem. This benchmarkdataset contains 48 pairs of hazy and haze-free videos from scenes with 4 different styles.The haze is produced artificially by a professional haze machine and all the videos weretaken in a controlled environment.

Table 3. Listing of datasets used for the dehazing task.

Dataset Synthetic (S)/Real (R)/Generated (G) Indoor (I)/Outdoor (O) Pairs Year

FRIDA [33] S O 72 2010FRIDA2 [34] S O 264 2012

CHIC [35] G I 18 2016RESIDE [36] S & R I & O 10,000+ 2018

D-HAZY [37] S I 1400+ 2016I-HAZE [38] G I 35 2018O-HAZE [39] G O 45 2018

DENSE-HAZE [40] G O 35 2019NH-HAZE [41] G O 55 2020

HazeRD [44] S O 70 2017BeDDE [45] R O 200+ 20204KID [46] S O 10,000 2021

REVIDE [47] G I 40+ (videos) 2021

Although there exist a satisfactory number of publicly available datasets for thederaining, desnowing and dehazing problems, unfortunately none of them contain scenes

Sensors 2022, 22, 4707 9 of 44

representative of the FRs’ working environment. Additionally, in most of the paireddatasets the noise is either synthetic or generated artificially using a machine. This leadsto the fact that most of the models proposed in the literature suffer from the domain shiftproblem and they are not able to perform well in real-world scenarios.

3. A Review of the Deraining Literature

Rain is one of the most common bad weather conditions. When capturing an image ina rainy scene, the blurring effect and haziness are noticeable phenomena causing generallya degradation of the visual quality [48]. This degradation can affect FRs’ visual perceptionas well as the performance of any CV tool employed to help their operations in outdoorscenes. In the last decades, rain removal, which is referred to as deraining, is a task that hasattracted a lot of interest in the research community. In this Section, we survey the recentbibliography on the deraining problem. Most of the presented methods target the singleimage deraining problem, in which the goal is to restore a clean image from an input imagewhich is corrupted by rain streaks. Other methods target the deraining problem from thestandpoint of other modalities, such as image sequences (video) or stereo images.

3.1. A Taxonomy of the DL-Based Single Image Deraining Methods

In this Section, we provide a taxonomy of the surveyed deraining methods in terms ofthe methodological attributes that may characterize a particular deraining algorithm. Toprovide a more compact representation of the collected works on the deraining problem,Table 4 directly summarizes the methods they present in keywords, and further categorizesome of their common attributes (for instance, their general model idea or their use of anunderlying imaging equation).

Previous surveys on the subject [49–53] all provide a similar categorization of the sur-veyed methods. In this survey paper, we particularly highlight the attributes of more recentmethods; that is, methods that were published in the year 2020 onward. By inspecting thereviews compactly being presented in Table 4, we came up with a set of seven methodolog-ical attributes which characterize the surveyed methods. Each method presented in thisSection may be tied with at least one of the suggested attributes. For example, a methodmay simultaneously employ a generative adversarial network-like architecture and harnessthis architecture by means of an attention mechanism.

3.1.1. CNN-Based Deraining Methods

Deraining networks based on deep convolutional neural networks (CNNs) are net-works which are mainly composed by convolutions, pooling layers, dropout layers, orother layers which are common in vanilla deep CNNs. These plain models usually donot contain more sophisticated structures such as attention layers, multi-scale analysisconstructions, etc.

Fu et al. [17] propose the DetailNet method, which uses a CNN model leveraging priordomain knowledge from the high-frequency content of training images. Then, by doing sothe model can learn the structure of rain streaks from training data. The authors empiricallyobserve that the learnt model can also generalize well on real data when trained only ononly synthetic data. Fu et al. [48] propose the DerainNet model. The model employs adeep CNN and uses the high-pass frequency layer computed from input images. As in themodel presented by Fu et al. [17], this model can also generalize well on real-world rainimages with the underlying model being trained on synthetic data. Yang et al. [54] injecteda hierarchical wavelet transform into a recurrent process that considers the low-frequencycomponent computed from an input image. The learnt network can adapt to rain streaks ofa larger size while being trained on image examples with rain streaks of only one size.

Coarse-to-fine or local-to-global strategies are also common in the general deraining lit-erature. Fan et al. [55] propose the ResGuideNetwork model. This model can progressivelyproduce high quality deraining results, and it has a small set of parameters. It imple-ments a novel residual-guide feature fusion network. A cascaded network is proposed and

Sensors 2022, 22, 4707 10 of 44

residuals are adopted from shallower blocks to provide data to blocks that are at deeperpositions in the network. A coarse-to-fine estimation of negative residuals is also realizedin the overall model. Wang et al. [28] propose an image deraining method that is based onhuman supervision and temporal priors in order to derain images. It furthermore entailsa novel SPatial Attetive Network (SPANet) to remove rain streaks with a local-to-globalremoval strategy.

Li et al. [56] proposes the NLEDN model, an end-to-end encoder-decoder networkwith pooling layers, suitable for efficiently learning increasingly abstract feature repre-sentations. The encoder-decoder design implements non-locally enhanced dense blocks.These blocks are useful for learning features of a hierarchical structure, and for learninglong-range feature dependencies. Pan et al. [57] propose the dual CNN (DualCNN) forsuper-resolution, edge-preserving filtering, deraining and dehazing. The network consistsof two parallel branches, recovering the structures and details end-to-end. The Dual-CNN can be integrated with common CNN models. Huang and Zhang [58] propose aderaining network employing the Dynamic Multi-domain Translation (DMT) module. Theparameters of the DMT-Net are estimated end-to-end.

Zhang and Patel [20] propose the DID-MDN multi-stream densely connected CNNfor joint rain density estimation and deraining. The network can learn the density of rainfrom training images and can use features computed at different scales. The reader can alsobe pointed to the multi-scale methods reviewed in Section 3.1.5.

Cho et al. [59] proposes a deraining network based on a memory network thatexplicitly helps to capture long-term rain streak information. The memory network istied with an encoder-decoder architecture, similarly with the works by Li et al. [56];Pan et al. [57]; and, Li et al. [25]. The features that are extracted from the encoder are readand updated within the memory network.

Guo et al. [60] views single image deraining as a predictive filtering problem. Via theirmethod, they predict proper kernels via a deep neural network that filters out pixels inrainy images. The model addresses so-called residual rain traces, multi-scale and diverserain patterns by achieving deraining efficiency.

Zheng et al. [61] propose the Segmentation-aware Progressive Network (SAPNet)based on contrastive learning. First, a lightweight network comprising Progressive DilatedUnits (PDUs) is formed. The SAPNet model also implements an Unsupervised BackgroundSegmentation (UBS) network, which can preserve semantic information and improvethe generalization ability of the network. Learning the model occurs by optimizing theperceptual contrastive and perceptual image similarity loss functions.

Li et al. [25] proposes a novel multi-task learning, end-to-end architecture. Theapplication of multi-task learning in this model is similar to that applied by Yang et al. [23].The decomposition network splits rain images into two layers: the clean background imagelayer, and the rain-streak layer. During the training phase, the authors further employa composition structure to reproduce the input by the separated clean image and raininformation for improving the quality of the decomposition. Jiang et al. [62] propose theProgressive Coupled Network (PCNet) to separate rain streaks from the useful backgroundregions of rainy images. The integrated Coupled Representation Model (CRM) learns thejoint features and blending correlations. A low-memory requirement for the method ismade possible by the incorporation of depth-wise separable convolutions and a U-shapedneural network structure.

3.1.2. Different Learning Schemes for Deraining

In this Section, we mention a few works on the single image deraining problem thatare based on classic ML schemes that are common in the general literature of ML. Thetarget learning schemes met here are multi-task learning, self-supervised learning andsemi-supervised learning.

Yang et al. [23] developed the JORDER method. The method uses multi-task learningto jointly learn a binary rain map, and finally estimates the rain streak layers and the

Sensors 2022, 22, 4707 11 of 44

clean background. The authors introduced the contextual dilated network, able to exploitregional contextual information. A recurrent process progressively removing rain streakscan handle rain streak variability.

The Few Shot Self-Supervised Image Deraining (FLUID) method by Rai et al. [63]proposes a self-supervised technique for single image deraining. As the authors mention,self-supervised approaches rely on two assumptions: (a) the distribution of noise or rainis uniform, and (b) the value of a noisy or rainy pixel is independent of its neighbors.To overcome these problems, the authors of FLUID hypothesize a network trained withminimal supervision to compute the probabilities of pixel rain-class membership.

Wei et al. [16] propose a semi-supervised learning approach where singleton rainyimage examples are used, without their paired corresponding clean images. The networklearns to use the regularities of the data in the unlabeled examples by first bootstrapping amodel given the examples with labelled rain information.

3.1.3. Generative Models for Deraining

The adversarial learning framework developed by Goodfellow and colleagues [64]targets to learn a generative model of the underlying data distribution by being given afinite training dataset of labeled examples. The framework learns a generator function G(usually taking the form of a multilayer perceptron) from which we can sample examples,and a discriminating function D that computes the probability of an example being a realexample drawn from the training set. To learn both functions given the training data, theframework alternates the optimization of the parameters of the two functions (the generatorand the discriminator function), rendering a minimax 2-player optimization game. Thisgame is represented by a value function, which is in turn a function of the parameters ofthe generating and the discriminator functions. The framework attempts to sequentiallyalternate among minimizing the value function as a function of the generator’s parameters,and then maximize it in terms of the parameters of the discriminator function.

Four of the surveyed research papers and articles that we collected in this survey, arebased on the generative adversarial learning scheme by Goodfellow et al. [64]. Otherwise,an altered adversarial learning scheme is used. Here, we briefly discuss how they work.

Qian et al. [65] present an altered adversarial learning scheme for single imagederaining. The scheme employs an altered generative adversarial network formalism, inwhich the generator function uses a contextual autoencoder. The autoencoder requires aninput image, and the attentive-recurrent module computes an attention map for the inputimage. The multi-scale loss and the perceptual loss, are employed for the autoencoderto be able to extract contextual information off from different scales (see also multi-scalemethods in Section 3.1.5). The discriminator function enforces global and local imagecontent consistency. Global-to-local and coarse-to-fine strategies are also discussed inSection 3.1.1.

Guo et al. [15] propose the DerainAttentionGAN model for single image deraining.The generative adversarial model scheme used in this framework contains two generatorfunctions G and F and two corresponding discriminator functions Dclean and Drain. Thecycle consistency loss involves these functions and affects the parameters of the networkunder the condition that F(G(rain)) ≈ rain is obeyed approximately. The generatorfunction G is linked with three subnetworks: (a) the parameter sharing feature extractor GF;(b) the attention network GA; and, (c) the transformation network GT . Wei et al. [66] proposethe DerainCycleGAN model (for a graphical representation of this model, see Figure 1),providing an adversarial learning-based methodology for single image deraining. Themodel follows the CycleGAN formalism by Zhou et al. [67], and employs the unsupervisedrain attentive detector (URAD) to attend to rain information in both rain and rain-freeimage inputs.

Sensors 2022, 22, 4707 12 of 44

Figure 1. Schematic of the DerainCycleGAN method by Wei et al. [66]. The model follows theCycleGAN regime and employs URAD (Unsupervised Rain Attentive Detector) modules to attend torain information and guide intermediate data projections.

Yang et al. [68] propose a generative adversarial learning scheme, typically includinga generator and a discriminator function. The authors integrate capsule units in both thegenerator and the discriminator functions in order to further improve their discriminantabilities in generatively identifying the characteristic structure of rain streaks in targetimages. The method is optimized by iteratively minimizing the loss functor in terms ofthe generator and the discriminator functions. Moreover, a Rain Component Aware (RCA)loss is introduced to minimize the differences between the synthesized rain componentmap and the ground-truth by forwarding them through the pretrained network (termedthe RCA network).

Jin et al. [13] proposed the UD-GAN method, which relies on an altered adversariallearning formulation. This formulation includes a rain generation function that takes intoaccount the variation of style and color.

The RainGAN model [69] is trained to decompose a rainy image into a clean imagecomponent and a latent 2D matrix encoding raindrop structures in the rainy image. Itcan also reverse this mapping and map a clean image and a raindrop latent code into animage with raindrops. This mapping is defined by the implementation of domain-invariantresidual blocks. HeavyRainStorer [70] is a method specifically designed to handle imagescontaminated by heavy rain, which can cause the veiling effect. The proposed method is anetwork consisting of two processing stages. In the first stage, a physics-based model isapplied to remove rain streaks from an input image, the transmission and the atmosphericlight. In the second stage, a depth-guided Generative Adversarial Network (GAN) recoversthe background details and corrects artifacts caused by the first processing stage.

The ID-CGAN (Image Deraining Conditional Generative Adversarial Network) [21]method considers quantitative, visual and discriminative performance in a common objec-tive function mimicking the form of the conditional GAN loss functions.

3.1.4. Attention Mechanisms for Deraining

In DL, neural attentive mechanisms are mathematical constructs which are integratedin the DL model and are applied on training data, allowing for the network model to attendto some interesting portions of the data while adaptively neglecting other data that areprobably uninteresting to the model and most probably to a human interpreting the model.A model learns to adapt an attention construct to the data by means of model optimization.In this paragraph, we briefly mention the attentive mechanisms employed by the derainingmethods which are reviewed in our survey.

Jiang et al. [71] propose a neural network that decomposes images with rain streaksinto multiple rain-specific layers. Each consequent layer is individually analyzed in orderto identify rain streak structures. Improved non-local blocks are designed to exploit the self-similarity in rain information by learning spatial feature correlations, while simultaneouslyreducing computation time. A mixed attention mechanism is applied to guide the fusion ofrain layers.

Sensors 2022, 22, 4707 13 of 44

Hu et al. [72] present the DAF-Net model. The model is trained to learn depth-attentional features from the training image set. At first, the input image is forwardedthrough a deep neural network that calculates the depth map of the input image. Thenthe integrated attention mechanism is applied to the calculated depth map and attentionweights are, in turn, computed. The latter weights indicate the local and more distantstructure in depicted rain streaks.

The APAN method by Wang et al. [73] employs an attention module that oper-ates at an across-scale manner to capture long-range feature correspondences from com-puted multi-scale features. The dual attention residual group network (DARGNet) byZhang et al. [74] proposes a single image deraining model for increased performance onthe task. The framework comprises two attention mechanisms: a spatial attention mech-anism and a channel attention one. The spatial attention mechanism extracts multi-scalefeatures that can capture the variability of both shape and size in rain streaks. In turn, thechannel attention mechanism calculates signal dependencies that exist among differentchannels. We can draw a similarity with the method of Jiang et al. [71] in the computationof cross-channel signal dependencies. Wei et al. [75] proposed a hybrid neural networkwith an integrated robust attention mechanism, termed the Robust Attention DerainingNetwork (RadNet). The authors designed a lightweight network with a universal attentionmechanism for rain removal at a coarse level, and then proposed a deep neural networkdesign with multi-scale blocks for finer rain removal.

The recent method by Zhou et al. [76] proposes a task adaptive attention module to en-able the neural network to restore images with multiple degradation factors. Moreover, themodel also employs a task channel attention module and a task operation attention module.

Lin et al. [77] proposed an Efficient Channel Attention (ECA) module that can extractglobal information and detailed information adaptively, resulting in a network module thatis extremely lightweight.

Li et al. [78] proposed an encoder-decoder-based model employing the rain embeddingloss. This loss is used to supervise the encoder. The Rectified Local Contrast Normalization(RLCN) is used to extract candidate rain pixels. The authors also propose to use a LongShort-Term Memory (LSTM) network for recurrent deraining and fine-grained encoderfeature refinement at different scales. Figure 2 shows the schematic of the method.

Figure 2. Schematic of the SIR method by Li et al. [78]. The model employs an autoencoder forsingle image deraining, regularization loss functions and deterministically generated features thatencourage the autoencoder to attend to rain streak features.

Sensors 2022, 22, 4707 14 of 44

3.1.5. Multi-Scale Based Deraining Methods

Extracting information from multiple scales off from training data or their intermediaterepresentations may reveal complimentary structures which may be usable in subsequentneural processing. In this Section, we review the multi-scale data analysis strategiesfollowed by the single image deraining literature.

Wei et al. [75] introduces the RadNet model. The model comprises a Robust AttentionModule (RAM), and a universal attention mechanism for rain removal at a coarse level.Moreover, a Deep Refining Module (DRM) comprising multi-scale blocks allows for finerain removal.

The LFDN network by Zhang et al. [79] employs the Multi-scale Hierarchical FusionModule (MSHF). The input to this module is an initial feature map. Intermediate featurescomputed from the initial feature map are extracted using downblock modules. TheMSPFN method by Jiang et al. [80] employs a multi-scale collaborative representation forrain streaks. This framework is unified and can model rain streaks in terms of their scaleand their hierarchical deep features.

Fu et al. [81] exploits underlying complimentary information not only across mul-tiple scales but also between channels. The network is designed to transmit the inter-level and inter-scale features in the neural processing pipeline. The Dual AttentionResidual Group Network (DARGNet) by Zhang et al. [74] integrates a dual attentionmodel, which comprises spatial attention and channel attention that can integrate scale andchannel information.

The Dense Feature Pyramid Grids Network (DFPGN) by Wang et al. [82] computesmulti-scale features from an input image and then shares these features through severalpathways across layers.

Yamamichi and Han [83] introduces a novel, multi-scale, residual-aggregation derain-ing network called MRADN. A residual backbone extracts fine and detail context in aninitial scale. A multi-scale analysis module augments feature learning from a semanticcontext representation.

Zheng et al. [61] proposed the Segmentation-Aware Progressive Network (SAPNet)method. The model is lightweight and repetitively employs Progressive Dilated Units(PDUs). The PDU is reported to crucially expand the receptive field and allow for amulti-scale analysis of rain streaks.

Jasuga et al. [84] developed the SphinxNet deraining neural network (see Figure 3).The model can be described in terms of three computation phases: (a) data scaling; (b) theDerainBlock phase, which entails the feedforward pass of initial scaling data across hierar-chically placed local encoder and decoder networks; and, (c) final scaling. After the laststep, a derained output image is reported.

Figure 3. Schematic of the SphinxNet method by Jasuja et al. [84]. Encoder and decoder modules arearranged hierarchically to allow for multi-scale feature encoding and decoding.

Sensors 2022, 22, 4707 15 of 44

3.1.6. Recurrent Representations for Deraining

Recurrent representations in DL are deep neural networks which are implementedas a recursive function that is, in turn, a function of a set of parameters. The next term inthe recursive function depends on the previous term or on more than one of the previousterms, assuming a set of model parameters.

Zheng et al. [85] employ a model that is lightweight and contains recurrent layers tosequentially remove rain at each level of a pyramid. Figure 4 shows the schematic of themethod. Fu et al. [86] proposes the lightweight pyramid network (LPNet) for the task ofsingle image deraining. It leverages domain-specific knowledge and Gaussian-Laplacianpyramid decomposition in order to simplify learning globally and at each pyramid level,respectively. Recursive and residual network structures are employed to assemble theLPNet, maintaining a low count of network parameters and state-of-the-art performance.Li et al. [87] introduces the RESCAN model. Contextual dilated networks are employedby this model to iteratively (or progressively) remove rain streaks. Squeeze-and-excitationblocks are used to estimate the strength weight of various rain streak layers before they arelinearly combined to approximate the rain-free image. A recurrent neural network linksfeatures generated at the different stages of computation.

Figure 4. Schematic of the method by Zhang et al. [85]. The (a) shows the MSKSN model, and(b) shows the MKSB module.

The method presented by Su et al. [88] proposes a novel, unified and recurrentconvolutional residual network. The model employs the so-called R-blocks in convolutionalresidual networks. Non-local Channel Aggregation (NCA) simultaneously captures long-range spatial and channel correlations. In the popular PReNet model introduced byRen et al. [89], a recurrent layer exploits the correspondence of deep features computed atdifferent stages. The recursive evaluation of the ResNet model occurring at a single neuralprocessing stage can affect the count of network parameters at the expense of negligibleimage deraining quality.

Lin et al. [90] uses parallel recurrent subnetworks to distribute the load of identifyingrain streaks at certain ranges of streak size.

3.1.7. Data Fusion Strategies for Deraining

The strategies of fusing data in deep neural networks entail the extraction and/orcombination of data or information from different sources to different targets. In thisSection, we describe the fusion strategies used in the single image deraining bibliography.

The ResGuideNet method by Fan et al. [55] fuses feature activations computed in alayer-by-layer fashion. Similarly with the ResGuideNet model, the DFPGN model proposed

Sensors 2022, 22, 4707 16 of 44

by Wang et al. [82] fuses intermediate feature representations at neural processing layersby means of dense connection blocks.

The method by Zhang et al. [91] includes the SFNet and ViewNet neural submodules.SFNet computes neural activations from derained images and their scene segmentations,and concatenates the latent vector representations from both modalities before the concate-nated vector is fed to a deep CNN. The VFNet fuses the derained image representationand the scene-segmentation representation obtained from the initial left and right stereochannels. Then, an encoder-decoder network reconstructs the left and right derained im-ages. The Lightweight Fusion Distillation Network (LFDN) by Zhang et al. [79] employs amulti-scale, hierarchical fusion scheme (termed MSHF). The model can encode images withrain and blur artifacts. It also regulates the parameter count of the resulting neural model.

Wang et al. [92] proposes recurrent scale-guide networks for single image deraining.The method goes beyond the case of a monolithic single-stage deraining neural networkmodel; it introduces a recurrent network framework and employs a Long Short-term Mem-ory (LSTM) network to join link neural processing stages simultaneously. Cai et al. [93]propose the dual recursive network (DRN). The method is empirically shown to be fast atimage deraining. The increased running time speed is mainly attributed to the very lownumber of model parameters, and the variable recurrent count that is a model hyperpa-rameter. Besides the low model parameter count, the quality of its output is comparableor better than state-of-the-art approaches (PSNR and SSIM measures are considered bythe authors). The DRN method utilizes convolutional feature maps, and a recursive resid-ual network which consists of two residual blocks. The residual blocks are evaluatedrecurrently, allowing for the model to iteratively refine the deraining transformation. Themethod resembles other published techniques which progressively refine their output (forinstance, see the MSPFN method [80]).

3.2. Multi-Image Deraining

Zhang et al. [91] and Kim et al. [94] present two methods that receive as input stereoimages. The rest of the methods in this Section, including the work of Kim et al. [94], arevideo deraining methods.

3.2.1. Stereo-Based Methods for Deraining

Zhang et al. [91] propose the Paired Rain Removal Network (PRRNet). The modelcomprises a Semantic-Aware Deraining Module (SADM) which can simultaneously per-form semantic segmentation and deraining. The Semantic Fusion Network (SFNet) and theView-Fusion Network (VFNet) both fuse semantic information and information obtainedfrom multiple views. Additionally, the Enhanced Paired Rain Removal Network (EPRRNet)uses a prior that describes the semantic information in rain streaks in order to remove rainstreaks. A coarse deraining network first reduces the rain streaks on the input images, andthen a semantic segmentation network extracts semantic features from the coarse derainedimage. A better deraining result is obtained by another network that fuses the semanticand multi-view information.

Kim et al. [94] proposes a method for video deraining that makes use of temporalcorrelations in the sequence of video frames, and uses a low-rank matrix completion methodto remove rain streaks in a video sequence. Video deraining is achieved as a sequence ofsteps: at first, a rain pixel probability map is computed by frame subtraction; secondly,the rain probability map is analyzed in terms of sparse basis vectors; third, the obtainedbasis vectors are classified into vectors referring to rain streaks or vectors pertaining tooutlying information. A low-rank matrix completion technique is then applied to removerain streaks.

3.2.2. Video-Based Methods for Deraining

Xue et al. [95] propose a real-time method for video deraining. A fast attentivedeformable alignment module and a spatial-temporal reconstruction module are used.

Sensors 2022, 22, 4707 17 of 44

Moreover, deformable convolution based on the channel attention mechanism is employedto maintain frame continuity. Neural architecture search is also employed to discover aneffective architecture for temporal information aggregation. Zhang et al. [96] propose theEnhanced Spatio-Temporal Interaction Network (ESTINet). This model is optimized forbetter video deraining quality and for data processing speed. Deep residual networksand convolutional LSTM models are employed to capture spatial features and temporalcorrelations in successive frames. Yan et al. [97] employ a self-alignment network withtransmission-depth consistency. The method uses deformable convolution layers in anencoder function for feature-level frame alignment. The temporal relationships amongframes are considered in order to make a prediction for the current frame. The model canalso handle the accumulation of rain to resolve the ambiguity between the depth and water-droplet density. The network estimates the depth from consecutive rain-accumulation-removal outputs and calculates the transmission map using a physics-based model.

The ADN method by Yang and Lu [98] extracts multi-scale features from input shallowfeatures. The extracted hierarchical, multi-scale features are then concatenated togetherand fed into a rainy map generator that estimates the rain layer.

The method by Wang et al. [99] proposes the Recurrent Multi-level Residual GlobalAttention Network (RMRGN) to gradually utilize global attention information and imagedetails to remove rain streaks progressively.

Deng et al. [100] propose the RoDerain network. The model employs the so-calledrotation operator to remove the rain streaks in natural and stochastic scenes. The alternatingdirection method of multipliers is used to optimize the model.

Kulkarni et al. [101] proposes the Progressive Subtractive Recurrent LightweightNetwork (PSRLN) for video deraining. The Multi-kernel Feature Sharing Residual Block(MKSRB) learns to capture the structure in rain streaks of a varying size. This allows forthe removal of rain streaks through iterative subtraction. Features that are generated bythe MKSRB are merged with output generated at previous frames in a recurrent fashion tomaintain temporal consistency. The Multi-scale Multi-Receptive Difference Block (MMRDB)performs feature subtraction as the means to avoid detail loss and extract HF information.Yue et al. [102] present a semi-supervised video deraining method, in which a dynamicalrain generator is employed to fit the rain layer. The dynamical generator consists of oneemission model and one transition model. These both help encode the spatial appearanceand temporal dynamics of rain streaks. Different prior formats are designed for the labeledsynthetic and unlabeled real data so as to fully exploit their underlying common knowledge.An expectation-maximization algorithm is developed to learn the model.

Yang et al. [103] proposes a two-stage recurrent network with dual-level flow regu-larization. At a first stage, the architecture extracts motion information from the initiallyestimated rain-free frame, and motion modeling at a second stage. Furthermore, to main-tain motion coherence between frames, dual-level flow-based regularization is proposed.This regularization occurs at a coarse flow level and at a fine-pixel level. Wang et al. [99]and Kulkarni et al. [101] also introduce recurrent processing in the models they propose.

Sensors 2022, 22, 4707 18 of 44

Table 4. Listing of surveyed research papers.

Category Method Model Short Description YearDetailNet [17] ACM reduces mapping range; promotes HF details 2017

Residual-guide [55] ACM cascaded; residuals; coarse-to-fine 2018NLEDN [56] ACM end-to-end, non-locally-enhanced; spatial correlation 2018

DID-MDN [20] ACM density-aware multi-stream densely connected CNN 2018DualCNN [57] ACM estimation of structures and details 2018Scale-free [54] HRMLL wavelet analysis 2019DMTNet [58] ACM symmetry reduces complexity; multidomain translation 2021UC-PFilt [60] ACM predictive kernels; removes residual rain traces 2022

SAPNet [61] N/A PDUs; unsupervised background segmentation;perceptual loss 2022

DDC [25] SBM decomposition and composition network; rain mask 2019DerainNet [48] ACM non-linear rainy-to-clear mapping 2017

PCNet [62] MRSL learns joint features of rainy and clear content 2021Spatial Attention [28] ACM human supervision; global-to-local attention 2019

CNN-based

memory encoder–decoder [59] ACM encoder–decoder architecture with memory 2022APAN [73] ACM multi-scale pyramid representation; attention 2021IADN [71] ACM self-similarity of rain; mixed attention mechanism; fusion 2020

DECAN [77] ACM detail-guided channel attention module identifieslow-level features; background repair network

2021

DAF-Net [72] DRM end-to-end model; depth-attentional features learning 2019SIR [78] ACM encoder–decoder embedding; layered LSTM 2022

RadNet [75] ACM/Raindrop

restores raindrops and rainstreaks; handles single-type,superimposed-type or blended-type data 2021

DARGNet [74] HRM dual-attention (spatial and channel) 2021

Attention

task-adaptive attention [76] N/A task-adaptive, task-channel, task-operation attention 2022DerainAttentionGAN [15] ACM uses Cycle-GAN; attention 2022

DerainCycleGAN [66] ACM CycleGAN transfer learning; unsupervised attention 2021RCA-cGAN [68] ACM rain streak characteristics; integration with cGAN 2022RainGAN [69] Raindrop raindrop removal as many-to-one translation 2022UD-GAN [13] ACM GAN; self-supervised constraints from intrinsic statistics 2019

HeavyRainStorer [70] HRM 2-stage network; physics-based backbone; depth-guidedGAN 2019

ID-CGAN [21] ACM conditional GAN with additional constraint 2020

Generativemodels

AttGAN [65] Raindrop attentive GAN; learns rain structure 2018MSPFN [80] N/A streak correlations; multi-scale progressive fusion 2020

MRADN [83] ACM multi-scale residual aggregation; multi-scale contextaggregation; multi-resolution feature extraction 2021

LFDN [79] N/A encoder–decoder architecture; encoder with multi-scaleanalysis; decoder with feature distillation; module fusion 2021

SphinxNet [84] N/A AEs for maximum spatial awareness; convolutionallayers; skip concatenation connections

2021

DFPGN [82] ACM cross-scale information merge; cross-layer feature fusion 2021GAGN [81] ACM context-wise; multi-scale analysis 2022

Multi-scalebased

UMRL [104] ACM UMRL network learns rain content at different scales 2019

JORDER [23] HRM multi-task learning; priors on equation parameters 2020FLUID [63] N/A few-shot; self-supervised; inpainting 2022

Differentlearningschemes Semi-supervised CNN [16] ACM adapts to unpaired data by training on paired data 2019

PReNet [89] ACM repeated ResNet; recursive; multi-scale info extraction 2019recurrent residual

multi-scale [85]MRSL residual multi-scale pyramid; coarse-to-fine progressive

rain removal; attention map; multi-scale kernel selection2022

Scale-aware [90] HRM multiple subnetworks handle range of rain characteristics 2017RESCAN [87] Equation (A8) contextual dilated network; squeeze-and-excitation block 2018

Pyramid Derain [86] ACM Gaussian–Laplacian pyramid decomposition 2019DRN [93] ACM multi-stage residual network with two residual blocks 2019

NCANet [88] Equation (A10) rain streaks as residuals sum; recurrent 2022PRRNet [91] ACM stereo; semantic segmentation; multi-view fusion 2021

Recurrent

GTA-Net [105] ACM multi-stream coarse; single-stream fine 2021

Sensors 2022, 22, 4707 19 of 44

4. A Review of the Desnowing Literature

Sometimes, FRs need to detect victims in snowy environments, as for example moun-tain rescuers that search for victims who are lost in mountain areas. Snow is a commonweather condition that can reduce the visibility of objects and scenes and obstruct theFRs’ missions. Unlike other atmospheric particles, snow particles have more complexcharacteristics, opaqueness, different shapes and sizes, uneven densities and irregularfalling trajectories, which make the snow removal problem, also known as desnowing,more challenging [29].

4.1. Related Work on the Desnowing Problem

In this section, we review the published literature on the desnowing problem; thatis, image restoration methods which are designed to remove snow particles from images.Table 5 lists the related work that we survey. Unlike the deraining and dehazing problemswhich have attracted significant attention by scientists, the desnowing problem has onlyattracted few research works. Importantly, however, researchers working on this problemalso contributed datasets for use by the research community. Here we list the reviews on therelated literature that we have collected on DL-based methods for single image desnowing.

4.1.1. CNN-Based Desnowing Methods

Chen et al. [30] employ the Dual Tree Complex Wavelet Transform (DTCWT) todecompose an input image into a series of high-frequency (HF) and corresponding low-frequency (LF) components and propose the hierarchical DTCWT desnowing network(HDCW-Net). Each component is then recursively analyzed in terms of its HF and LFcontent, until i recurrence levels are reached. Then an HF reconstruction neural network isemployed to reconstruct the HF image content, hence alleviating the effect of snow streaksand the snow veiling effect.

4.1.2. Generative Models for Desnowing

Li et al. [106] uses a compositional Generative Adversarial Network (compositionalGAN) architecture to separate the clean background image regions and the regions con-taminated by snow. This becomes possible by mainly employing the compositional loss,and some other minor loss functions that control the ability of the network to generate thecorrect output, to guide training of the cGAN model. As a model employing a GAN-stylearchitecture, it contains a discriminator and a generator network. Figure 5 shows theschematic of the method.

Figure 5. Schematic of the method by Li et al. [106]. A GAN-based architecture learns to extract theclean background image component and the rain streak layer component from an input rainy imageby means of the composition loss function.

Sensors 2022, 22, 4707 20 of 44

Chen et al. [32] proposes the JSTASR model (see Figure 6). This model can simultane-ously handle transparent and non-transparent snow particles by applying the modifiedpartial convolution. The model is transparency-aware; hence, it can handle snow withdifferent transparencies. JSTASR can classify snow particles by considering their sizeand scale.

Figure 6. Schematic of the JSTASR method by Chen et al. [32]. The model comprises three modules.First, it contains a joint size and transparency-aware snow removal module. Secondly, it employs aveiling effect recovery module, and finally it comprises a size-aware clean image discriminator.

Jaw et al. [107] proposed a novel Deep Neural Network (DNN) architecture witha top-down pathway and lateral cross-resolution connections; it is depicted in Figure 7.The model (called DesnowGAN) exploits high-level semantic features and low-level spatialfeatures for improved efficiency and running time. A novel loss function helps to jointlylearn sharpness, structure and realism.

Figure 7. Schematic of the DesnowGAN method by Jaw et al. [107]. A GAN-based snow removalmodule, a refinement module and a discriminator module guide single image desnowing.

4.1.3. Multi-Scale Based Desnowing Methods

Liu et al. [29] proposes the DesnowNet model which deals with the removal of translu-cent and opaque snow particles instead of paying attention to the transparency of snowparticles (see Figure 8). The model corrects the image content by accurately estimatingand restoring details in the image that are lost due to opaque snow particles. It modelssnow by means of a snow mask, which considers only the translucency of snow at eachcoordinate, and also by means of a chromatic aberration map that captures fine colordistortions. The model interprets snow through context-aware features and loss functions.DesnowNet also implements multi-scale receptive fields.

Sensors 2022, 22, 4707 21 of 44

Figure 8. Schematic of the DesnowNet method by Liu et al. [29].

Li et al. [31] introduces a physics-based snow model and proposes a novel snowremoval method based on the snow model and deep neural networks. The model decom-poses a snowy image non-linearly as a combination of a snow-free image and dynamicsnowflakes. The authors designed the Multi-scale Stacked Densely Connected Convolu-tional Network (MS-SDN) to simultaneously detect and remove snowflakes in an image.The MS-SDN is composed of a multi-scale convolutional subnetwork for extracting featuremaps and two stacked modified DenseNets for snowflake detection and removal.

Zhang et al. [108] proposed the Deep Dense Multi-Scale Network (DDMSNet) forsingle image desnowing. The model can learn multi-scale representation off from pixel-level and feature-level input. Dense connections are applied to connect together multi-scalefeature-computing subnetworks. The network exploits semantic and geometric informationas prior knowledge. Semantic and geometric features are obtained in different stages tohelp recover snow and recover clean images.

Table 5. Listing of references on the desnowing problem.

Category Method Short Description YearCNN-based HDCW-Net [30] DTCWT analysis; recursively computes HF component; neural network

reconstructs the last HF component 2021

cGAN [106] separates the background from snowy regions; uses compositional loss 2019

JSTASR [32] handles transparent/non-transparent snow particles; modified partialconvolution; transparency aware; considers size and scale of snow particles 2020

Generativemodels

DesnowGAN [107]

DNN with top-down pathway and lateral cross-resolution connections;high-level and low-level spatial features; split-transform-merge topology

reduces model size and computational cost; atrous spatial pyramidpooling for multi-scale and global receptive field

2020

DesnowNet [29] accurately corrects image content by estimating and restoring details inthe image that are lost due to opaque snow particles 2018

MS-SDN [31] multi-scale convolutional subnetwork extracts feature maps; stackedmodified DenseNets for snowflakes detection and removal 2019Multi-scale

based

DDMSNet [108]multi-scale representation from pixel-level and feature-level input;

multi-scale subnetwork are desnely connected; semantic and geometricpriors; multistage analysis

2021

5. A Review of the Dehazing Literature

Haze can be defined as the atmospheric phenomenon where dust, smoke or other par-ticles absorb and scatter the light, resulting in reduction of transparency of the air and conse-quently reduction of visibility [109]. FRs often need to take action in scenes of crises wheretheir visual perception is affected by haze (fire, explosion, rescuing in the mountains, etc.),causing serious problems in the detection and navigation to possible victims.

Sensors 2022, 22, 4707 22 of 44

Images captured in hazy environments suffer from loss of contrast, color fidelity andedge information which can further cause reduction of the performance of CV algorithmsused for tasks like object detection, image segmentation, etc. Haze removal, which isreferred to as dehazing, is considered an important process. Since most CV models as-sume clear weather conditions [110]. However, it can also be considered as a challengingproblem, since the haze is dependent on the unknown depth information which varies atdifferent positions [111].

5.1. A Taxonomy of the DL-Based Single Image Dehazing Methods

In order to solve the dehazing problem numerous methods have been proposedin the bibliography. These methods can be divided in two main categories: The hand-crafted priors-based methods and the data-driven methods. The hand-crafted priors-basedmethods use handcraft priors from empirical observation in order to remove the haze.They include priors such as dark channel (DCP) [111], contrast maximization [112], colorattenuation [113] and non-local prior [114]. Unlike the hand-crafted priors-based methods,the data-driven methods use large-scale datasets to learn automatically the image prior.The main focus of this paper will be in data-driven methods. A synopsis of the methods ispresented in Table 6.

5.1.1. CNN-Based Dehazing Methods

Early data-driven methods include ML algorithms like random forest regressor [115],linear models [116] and the back propagation neural network [117]. In 2016, Cai et al. [118]adopted a trainable three-layer CNN to directly estimate the transmission map (for moreinformation, see Section 6.3.5) from a hazy image and also developed the activation func-tion called Bilateral Rectified Linear Unit (BReLU) to extract haze-relevant features fortransmission recovery. Their pioneering model is known as Dehazenet and achieved adramatically high efficiency compared to the state of the art of its times.

Li et al. [119] used CNNs in order to generate the clear image but their approach wasdifferent in the way that they wanted to avoid the estimation of the transmission map andthe atmospheric light intensity. They introduced a new K(x) parameter that integrates thesetwo parmeters in one as shown in Equation (2). Their model is known in the bibliographyas All-in-One Dehazing Network (AOD-Net).

J(x) = K(x)I(x)− K(x) + b (1)

where b is the constant bias with the default value 1 and

K(x) =1

t(x) (I(x)− A) + (A− b)

I(x)− 1(2)

Ullah et al. [120] exploited the transformed Atmospheric Scattering Model (ASM)proposed in [119]. They introduced a lightweight CNN architecture for estimating theK(x) parameter, that consists of eight convolution layers and three concatenation layersand an additional Color Visibility Restoration (CVR) module that helps recovering colorinformation and contrast of the image by averaging color intensity and equalizing perchannel histogram.

Qin et al. [121] considered two important facts in their model: (1) the haze distributionmay be uneven on different pixels of the image; (2) the different channel features havetotally different weighted information. Their proposed end-to-end Feature Fusion AttentionNetwork (FFA-Net) uses both pixel attention and channel attention modules and combinesthem to design a feature attention module. The combined model is added to stackedresidual blocks. In this way, they managed to focus on processing pixels with thick hazeand more important channel information.

Inspired by the FFA-Net, Wu et al. [122] proposed an AE-like dehazing network withfeature attention block as the basic block, the AECR-Net. They managed to make their

Sensors 2022, 22, 4707 23 of 44

network much more compact than the FFA-Net and also introduced a novel contrastiveregularization in order to force the network to learn that the output dehazed image shouldbe closer to the clear image and farther from the hazed one in the representation space.

5.1.2. Multi-Scale Based Dehazing Methods

Inspired by the feature fusion structure of [121], Hu [123] proposed a multi-scale gridnetwork (MSFFA-Net) in order to avoid the bottleneck issue happened in the conventionalmulti-scale approach.

Liu et al. [124] had a CNN-based approach to the problem, without taking the ASMinto consideration. They introduced the GridDehazeNet (GDNet), a network consistingof three modules: the pre-process module which is a convolutional layer and a residualdense block and generates multiple different versions of the input image, the backbonemodule which performs attention-based multi-scale estimation based on the generatedinputs and finally the post-processing module, which is a residual dense block followingby a convolutional layer and achieves reduction of the artifacts.

Ren et al. [125] employed a Multi-Scale CNN (MSCNN), where firstly a coarse-scalenetwork was used to learn the mapping between hazy inputs and their transmissions,and then a fine-scale network performs refined transmission estimation. The two networksare alternately merged and upsampled to maintain the original resolution. Extendingtheir work, in 2020, Ren et al. [126] introduced an improved version of the MSCNN withfewer pooling and up-sampling layers in the coarse-scale network and a novel holisticedge guided network (MSCNN-HE) that ensures that the objects with the same depth havethe same transmission values. Similarly, in Wang et al. [127] also followed the multi-scalecoarse-to-fine idea and proposed an Ensemble Multi-scale Residual Attention Network(EMRA-Net). Their coarse network is a Three-scale Residual Attention CNN (TRA-CNN),where the different scales of the input image are produced using the 2-D Discrete WaveletTransform (DWT). Their refined network is an Ensemble Attention CNN (EA-CNN) thatfuses the three outputs of the coarse network into a final haze-free image.

Dong et al. [128] introduced the Multi-Scale Boosted Dehazing Network with DenseFeature Fusion (MSBDN). Their network is based on an encoder–decoder architecture witha dense feature fusion module and incorporates the Strengthen-Operate-Subtract (SOS)boosting strategy in the decoder to boost features of different levels. It also incorporatesan error feedback principle in order to preserve the spatial information and exploit thefeatures from non-adjacent levels.

Zhang et al. [129] adopted the AOD-Net’s transformed ASM proposed in [119] andtried to estimate K(x) fast and accurately in order to produce the final haze-free imageaccording to Equation (2). Their proposed Fast and Accurate Multi-scale End-to-end De-hazing Network (FAMED-Net) has adopted the Gaussian/Laplacian pyramid architecturesfollowed by a fusion module. Specifically, the input images are down-sampled to twodifferent scales. The original and the two down-sampled images are fed to three encoders(one encoder for each scale) that do not share features. The outputs of the encoders areinterpolated to the original scale and fed to a fusion module that estimates the K(x).

Zhao et al. [130] exploited long-range dependencies in order to improve the haze-free results. They introduced a Pyramid Global Context (PGC) block plugged into aU-Net, which is further improved by a dilated residual bottleneck (DRB) block. Theirproposed network achieved to extract long-range dependencies among points and patchescomputationally more efficiently than by just stacking many convolutional layers.

In a recent work, Sheng et al. [131] considered the influence of the haze in the lu-minance of an image in the CIELAB colorspace. They proposed the multi-scale residualattention network (MSRA-Net), a network that consists of two subnetworks: one for theluminance and one for the chrominance. The network also embodies the Multi-Scale Resid-ual Attention block (MSRA-block) and the Feature Aggregation Building block (FAB-block)and manages to improve the quality of the colors in the haze-free results.

Sensors 2022, 22, 4707 24 of 44

Fan et al. [132] incorporated depth information in their multi-scale network (MSDFN).Specifically, they relied on the U-Net architecture. The depth image is first encoded andthen decoded so as to produce the final haze-free image, while both in the encoding anddecoding procedures the multi-level features of the hazy image are concatenated.

Das and Dutta [133] experimented with a deep multi-patch and a deep multi-scalenetwork and tried to build a network which could remove the haze fast and efficientlyeven if it is non-homogeneous. They deduced that their multi-patch hierarchical network(DMPHN) is faster and better than their multi-scale hierarchical network, because it aggre-gates local features generated from a finer to a coarser grid. The Trident Dehazing Network(TDN) proposed by Liu et al. [134], depicted in Figure 9, also tries to remove efficiently thedense and non-homogeneous haze by following a multi-scale approach. It consists of threesubnetworks: an encoder–decoder network that builds the first version of the haze-freeimage, a detail refinement network that focuses on the HF details and a haze density mapgeneration network that detects which regions of the image have thick and which thin haze.Non-homogeneous and dense haze was also considered by Jo et al. [135]. They ignored theASM and created a multi-scale architecture that employs the selective residual blocks (SRB)as its main functional module.

Figure 9. The architecture of Trident Dehazing Network [134]. ⊕ refers to tensor addition and ⊗ totensor multiplication.

5.1.3. Generative Models for Dehazing

Zhang et al. [136] proposed a dehazing GAN that takes the ASM into considera-tion, known in the bibliography as DCPDN. The generator consists of two subnetworks:an edge-preserving pyramid densely connected encoder–decoder network to estimatethe transmission map and an 8-block U-net to estimate the atmospheric light. By us-ing Equation (13), they obtained the dehazed image. They also introduced a new edge-preserving loss function to optimize the network that estimates the transmission map and ajoint-discriminator based GAN framework to learn the dependencies between the dehazedimage and the corresponding estimated transmission map. In [137,138], GANs are also

Sensors 2022, 22, 4707 25 of 44

used for image dehazing and depend on the ASM; however, the model in [138] is based onunpaired supervision.

Ren et al. [139] completely ignored the ASM. Their proposed Gated Fusion Network(GFN) is based on the fusion based encoder–decoder architecture, employed an originalhazy image and three derived images as inputs (white balance, contrast enhancement,and gamma correction methods) and learned to predict confidence maps. After that,the confidence maps were fused to give the dehazed image. Their model was trainedwith MSE and adversarial loss. Qu et al. [140] also tried to avoid ASM and exploited anEnhanced Pix2pix Dehazing Network (EPDN). They utilised the idea of a multi-resolutiongenerator, a multi-scale discriminator and an enhancer.

Li et al. [141] proposed an end-to-end trainable conditional GAN (cGAN) with en-coder–decoder architecture that takes pairs of haze and haze-free images and restoresdirectly the dehazed images without the need of ASM. Kan et al. [142] also exploited thepaired cGAN framework and the multi-loss function optimization idea. They proposed aU-connection Residual Network (UR-Net) as a generator and the spatial pyramid pooling(SPP) structure for the discriminator. The multi-loss function is composed of a combinationof adversary loss, L1 loss, the structural similarity index (SSIM) loss and a new peak-signal-to-noise ratio (PSNR) loss. Another innovation of this work is that it offers a flexibility inthe size of the input image by embedding the SPP structure.

Engin et al. [143] exploited the CycleGan architecture and enhanced it to recover thehaze-free image from a hazed one. The model is depicted in Figure 10. The advantagesof their so-called Cycle-Dehaze network are that it requires neither the estimation of thetransmission map nor pairs of hazed and their corresponding clear images. The differ-ences between CycleGan and Cycle-Dehaze are that Cycle-Dehaze uses an additional loss,the cyclic perceptual-consistency loss and the Laplacian pyramid as a post-processing stepto upscale the dehazed images. Cycle-Dehaze achieves good dehazing results, howeversometimes there exist some distortions.

Figure 10. The architecture of the Cycle-Dehaze Model [143]. G and F are the generators, and D_xand D_y the discriminators.

Other works inspired by CycleGan architecture are [144–147]. In [144] Dudhaneand Murala followed a similar training strategy as CycleGan, but their Cycle-consistentgenerative adversarial network (CDNet) differed in the architecture of the generator, wherethey introduced an encoder–decoder architecture with skip connections that estimatesthe transmission map. Liu et al. [145] proposed the Cycle-Defog2Refog network thatconsists of two transformation paths (hazy to haze-free, haze-free to hazy) and a two-stage mapping strategy in each transformation path so as to improve the dehazed results.In order to produce the synthetic hazy images they estimated the transmission map using aCNN approach. Jin et al. [146] was inspired by the CycleGAN architecture and added a

Sensors 2022, 22, 4707 26 of 44

Conditional Disentangle Network (UCDN), in order to manage different concentrations ofhaze. Mo et al. [147] wanted their network to have the ability to handle uneven and densehaze concentration and introduced the Dark Channel Attention optimized CycleGAN(DCA-CycleGAN). The DCA-CycleGAN consists of two subnetworks that compose thegenerator, two global discriminators and two local discriminators. The one subnetworkof the generator (Dark Channel Attention subnetwork) outputs attention maps that assistthe other subnetwork (AE) to be fine-tuned and produce the dehazed image. Additionally,the local discriminators proposed in this work help the network to manage differentconcentration of haze in the image.

Park et al. [148] tried to fuse heterogeneous GANs and specifically the CycleGANwith the cGAN architecture. The concept of their work is to combine the great colorbalance that CycleGAN offer with the preservation of spatial details provided by cGANs.The CycleGAN is trained on unpaired, real, outdoor hazy and haze-free images, and learnsto produce a haze-free result. The cGANs are trained on paired, synthesized indoor images.In their proposed method they exploited two different cGANs: the one learns to estimatethe transmission map while the other learns to estimate the atmospheric light. UsingEquation (13) they produce a second version of the haze-free result. After that, the twoestimations of the haze-free results are fused in a CNN, which consists of feature extractionlayers, a merge layer, and reconstruction layers.

Dong et al. [109] adopted the FD-GAN method, which is a GAN with a denselyconnected encoder–decoder as the generator and a fusion-discriminator that integrates HFand LF information as additional priors and constraints. Unlike Dong et al. [109], who usea fusion-discriminator, Fu et al. [149] suggested fusing the HF domain features into thegenerator. Their generative adversarial network (DW-GAN) uses a two-branch-designednetwork as a generator, where the one branch (DWT Branch) embeds the 2D DWT, so asto acquire clear texture details and the other (Knowledge Adaptation Branch) uses thepre-trained Res2Net as an encoder in order to avoid over-fitting.

In Wang et al. [150] introduced the TMS-GAN. Their architecture is based on twosubnetworks: the Haze-generation GAN (HgGAN) and the Haze-removal GAN (HrGAN).The HgGAN takes a synthetic hazy image as input and learns to output a synthetic butrealistic hazy image. The HrGAN learns to remove the haze both from synthetic andsynthetic but realistic hazy images. Its generator is trained by the paired synthetic andsynthetic but realistic data and achieves to produce a color-branch and detail-branchoutput. The element-wise addition of these two outputs and the clear image producedby the generator of HgGAN feed the discriminator of HrGAN so as to learn to producethe final haze-free images. Additionally, they proposed a plug-and-play Multi-attentionProgressive Fusion Module (MAPFM) for both HgGAN and HrGAN.

5.1.4. Deep Reinforcement Learning for Dehazing

Zhang and Dong [151] wanted to combine the simplicity of prior based methods withthe generalization ability of DL and followed an innovative methodology in the imagedehazing problem. They proposed a deep reinforcement learning (RL) based method(dehaze-RL), where a deep Q-learning network iteratively chooses actions in order toproduce the final haze-free image. The actions set of the agent includes 11 actions that areexisting dehazing algorithms and PSNR and SSIM metrics are used in the reward functionas the measurement. The main advantage of this method is that the dehazed result is muchmore interpretable since in every state the corresponding result can be acquired.

Guo and Monga [152] considered the depth information of the image and proposed aDepth-aware Dehazing using RL system (DDRL). DDRL works by following the assump-tion that haze is lighter closer to the camera and denser farther from it, and consists of twonetworks: a policy network that generates the depth slice and a dehazing network thatestimates the transmission map of each slice.

Sensors 2022, 22, 4707 27 of 44

5.1.5. Knowledge Distillation/Transferring for Dehazing

Hong et al. [153] presented a model that uses knowledge distillation to apply dehaz-ing (KDDN), where the teacher is an auto-encoder that receives clean images, learns toreconstruct them and transfers the knowledge from the intermediate representations tothe student network. The lightweight student network receives the hazy images and bymimicking features from the teacher learns to output the haze-free image. The studentnetwork’s supervision is based solely on dark channels loss and total variation loss.

Shao et al. [154] considered the domain shift problem. Usually, the models proposedfor dehazing do not perform well in dehazing real hazy images, because they are trained onsynthetic ones. Hence, they introduced a domain adaptation framework, which includes animage translation network and two dehazing networks (one for the synthetic domain andone for the real). In this way, they managed to reduce the domain discrepancy and achievea good performance in the dehazing task in both domains. Domain gap was a problemthat Chen et al. [155] also tried to address. They suggested a Principled Synthetic-to-realDehazing (PSD) framework, which consists of a pre-trained network, trained on syntheticdata as a backbone and they fine-tuned it in an unsupervised manner using real hazyimages and physical priors that guide the fine-tuning process.

Yu et al. [156] recently introduced a two-branch neural network that is able to removenon-homogeneous haze and perform well even when trained on a small-scale dataset.The first subnetwork is a transfer learning subnetwork that extracts global representationsfrom a pre-trained model, while the second one is trained from scratch on current trainingdata and complements the first one. The final output is produced by ensembling thetwo subnetworks.

5.1.6. Unsupervised/Semi-Supervised Learning for Dehazing

Overall, the CycleGAN-based networks described in Section 5.1.3 are trained in an un-paired manner. Yang et al. [138] was the first to adopt this unpaired training by introducingthe conditional GAN architecture. Motivated by this idea, Golts et al. [157] evolved it andproposed a fully convolutional network with six dilated residual blocks. Their networkaims to approximate the Dark Channel Prior (DCP) energy function and is trained usingonly hazy images. Additionally, an early stopping approach was followed, where a set of500 paired images was used as a validation set and the training stopped in the epoch inwhich the validation set had the best performance in terms of PSNR and SSIM metrics.

Li et al. [158] presented a semi-supervised learning algorithm to solve the problemof single image dehazing. Their CNN-based network consists of two branches sharingthe same weights and having the same encoder–decoder architecture: a supervised andan unsupervised one. The supervised branch is trained using synthetic, paired data andsupervised losses while the unsupervised branch is trained using only real hazy imagesand unsupervised losses based on image priors.

RefineDNet proposed by Zhao et al. [159] combines the advantages of both prior-basedand learning-based methods and requires unpaired hazy and haze-free images. Their two-stage network adopts the DCP in the first stage for visibility restoration and adversariallearning in the second stage for realness improvement. In order to improve the quality ofthe dehazed image even more, a perceptual fusion strategy is followed.

The You Only Look Yourself (YOLY) network proposed in [160] employs three jointdisentanglement subnetworks: the J-Net, the T-Net and the A-Net that are designed topredict the clear image, the transmission map and the atmospheric light, respectively, in aself-supervised manner, given the hazy image as input.

Sensors 2022, 22, 4707 28 of 44

Table 6. Listing of the surveyed research papers and articles on the dehazing problem.

Category Method Short Description YearDehazenet [118] 3-layer CNN, BReLU activation function 2016AOD-Net [119] lightweight, transformed ASM 2017

Light-DehazeNet [120] lightweight, transformed ASM, CVR module 2021FFA-Net [121] attention-based feature fusion structure 2020

CNN-based

AECR-Net [122] AE-like, contrastive learning, feature fusion 2021MSFFA-Net [123] multi-scale grid network, feature fusion 2021

GDNet [124] 3 sub-processes, multi-scale grid network 2019MSCNN [125] 2 nets: coarse- and fine-scale 2016

MSCNN-HE [126] 3 nets: coarse-, fine-scale and holistic edge guided 2020EMRA-Net [127] 2 nets: TRA-CNN and EA-CNN 2021

MSBDN [128] dense feature fusion module, boosted decoder 2020FAMED-Net [129] 3 encoders at different scales, fusion module 2019

PGC [130] PGC and DRB blocks 2020MSRA-Net [131] CIELAB, 2 subnets (luminance, chrominance) 2022

MSDFN [132] depth-aware dehazing 2021DMPHN [133] non-homogeneous haze, multi-patch architecture 2020

TDN [134] 3 subnets: coarse-, fine-scale and haze density 2020

Multi-scalebased

Jo et al. [135] selective residual blocks 2021DCPDN [136] generator with 2 subnets, edge-preserving loss function 2018

DehazeGAN [137] ASM-based GAN 2018DDN [138] ASM-based, unpaired supervision 2018GFN [139] fusion based, employs a hazy image and 3 derived inputs 2018

EPDN [140] multi-resolution generator, multi-scale discriminator, enhancer 2019cGAN [141] cGAN with encoder–decoder architecture 2018

Kan et al. [142] cGAN, UR-Net as a generator, flexibility in image size 2022Cycle-Dehaze [143] CycleGan based, unpaired supervision 2018

CDNet [144] CycleGan based, encoder–decoder architecture for the generator 2019Cycle-Defog2Refog [145] 2 transformation paths with 2-stage mapping strategy in each 2020

UCDN [146] CycleGan based with a conditional disentangle network 2020DCA-CycleGAN [147] generator with 2 subnets, 4 discriminators 2022

Park et al. [148] fusion of cGAN and CycleGAN 2020FD-GAN [109] integration of HF and LF information in the discriminator 2020DW-GAN [149] generator with a DWT and a Knowledge Adaptation Branch 2021

Generativemodels

TMS-GAN [150] 2 subnets: a haze-generation and a haze-removal GAN 2021

RL-basedDehaze-RL [151] actions: 11 dehazing algorithms, reward function: PSNR and SSIM 2020

DDRL [152] depth-aware dehazing 2020KDDN [153] teacher-student (dehazing) net 2020

Shao et al. [154] domain adaptation using a bidirectional translation net 2020

PSD [155] domain adaptation by unsupervised fine-tuning (real domain) apre-trained model (synthetic domain) 2021

Knowledgedistillation/transferring

Yu et al. [156] 2-branch net: transfer learning and current data fitting subnets 2021Golts et al. [157] unsupervised, DCP loss 2019

Li et al. [158] 2-branch: supervised and unsupervised subnets 2019RefineDNet [159] 2-stage network: DCP and adversarial learning stages 2021

Unsupervised/Semi-

supervised YOLY [160] self-supervised, 3 joint disentanglement subnetworks 2021

5.2. Multi-Image Dehazing

Other modalities in the DL-based dehazing literature (apart from the single image)include video dehazing and binocular image dehazing.

In the binocular, DL-based image dehazing few methods have been proposed recently.Song et al. [161] proposed an encoder–decoder architecture that jointly learns to improvethe quality of the disparity maps and the dehazing results. Pang et al. [162] tried to avoidthe time consuming estimation of the disparity map and proposed the BidNet, whichreceives binocular image pairs and try to dehaze them by exploring the correlation betweenthe images using the Stereo Transformation Module. The most recent Stereo RefinementDehazing Network (SRDNet) [163] follows a coarse-to-fine approach and tries to removethe haze of the binocular image pairs by avoiding exploiting the ASM. Their proposed

Sensors 2022, 22, 4707 29 of 44

framework consists of two networks: a weight-sharing coarse dehazing network and aguided separated refinement network, where the latter fuses the information from crossviews avoiding in this way the estimation of the disparity or correlation matrix.

Although, in the literature, the single image dehazing algorithms are often usedfor video dehazing, video dehazing can also be considered as a separate task, since thetemporal coherence of video frames offers a great amount of useful information that cannotbe found in single images. Li et al. [164] employed a CNN-based architecture for videodehazing. Their End-to-End Video Dehazing Network (EVD-Net) is based on the AOD-Net [119] architecture and follows three different types of temporal fusion structures inorder to consider the available information of the temporal coherence of video frames.Ren et al. [165] proposed an encoder–decoder architecture that collects information acrossframes in order to estimate the transmission map. Their network tries to restore the frameswith the same scenes in a way that the transmissions are smooth enough using two differentfusion strategies.

6. Results6.1. Quantitative Metrics

In order to evaluate the quality of the output image produced by a deraining, a desnow-ing or a dehazing model, there are two widely used metrics. These metrics compare thedifferences between an output and a reference image and for that reason they are alsocalled full-reference metrics. In this section, we mention the two most common quanti-tative metrics that are commonplace in every possible study on deraining, desnowingand dehazing.

6.1.1. Peak Signal-to-Noise Ratio

The peak-signal-to-noise ratio (PSNR) measures the discrepancy among two images I(a reference image) and K (a ground-truth image). The discrepancy is measured numericallyas the log (base 10) of the ratio of the square of the maximum intensity value in the image Ito the mean squared error (MSE) of the corresponding pixel intensity values in image I andimage K. Formally, the PSNR measure is given by the formula

PSNR = 10 log10(MAX2I /MSE) (3)

where MAXI is the maximum intensity value in the set of pixels in image I and MSE is thevalue of the mean squared error of the pixel intensities among the reference image I andthe ground-truth image K. The MSE value is given by the formula

MSE =1

mn

m−1

∑i=0

n−1

∑j=0

[I(i, j)− K(i, j)]2 (4)

6.1.2. Structural Similarity

The structural similarity index (SSIM) [166] is a metric that measures the perceptualdifference between two images x and y and is given by

SSIM(x, y) =(2µxµy + c1)(2σxy + c2)

(µ2x + µ2

y + c1)(σ2x + σ2

y + c2)(5)

where c1 = (k1L)2 and c2 = (k2L)2 and L is the dynamic range of the pixel values definedas L = 2bits per pixel − 1 and k1 = 0.01 and k2 = 0.03. µx and µy are the expected valuesof the variables x and y, and σx and σy are the variances of x and y.

6.2. Real-Time Performance Classification

Another critical aspect that must be considered when employing a DL algorithm forenhancing the situational awareness of FRs is whether they are able to process data quickly.

Sensors 2022, 22, 4707 30 of 44

However, reports on different methods often regard different image resolution, whichnaturally has a significant impact on execution time. Hence, in the comparisons of thissection, we have classified each method as real-, near-real- or non-real-time, taking intoconsideration both reported frames per second (FPS) and image resolution. The real-timecategory includes methods that report high FPS in at least medium resolutions, or mediumFPS in higher resolutions. Near-real-time methods may report medium FPS in mediumresolutions or lower, but still acceptable FPS, in high resolutions. With improving hardwareand computational advances, near-real-time methods may achieve higher FPS and bepromoted to real-time in the near future. Non-real-time methods report very low FPS. Itmust be stressed that this classification is still subjective and should not be interpreted as astrict classification scheme.

6.3. Comparison of Models

The deraining, desnowing and dehazing methods that could serve the purpose ofaugmenting the vision of FRs during rescue mission operations should have a low responsetime and produce a visually pleasing output. Therefore, in this section we discuss theresults of the models proposed in the literature for deraining, desnowing and dehazing interms of PSNR and SSIM values, as well as their processing time. The compared methodsare evaluated based on benchmarking datasets. In Tables 7–9 we report the results for thederaining, desnowing and dehazing methods respectively. The highest PSNR and SSIMvalues are highlighted using bold text and the algorithms have been sorted in a top tobottom order, where the top is the method with the highest processing time needed.

Additionally, we have classified the methods in three categories (real-time, near-real-time, non-real-time) based on the maximum images per second that the models can process(for more information please see Section 6.2), considering also the size of the images. Dueto the large number of algorithms found in the literature and the lack of availability ofcode for many of the methods, the reported results are only based on the literature and noexperiments have been made by the authors of this paper. Lastly, some methods have beenomitted due to lack of available information.

6.3.1. Mathematical Background of Deraining

Mathematically, a rainy image O can be modeled as a linear superimposition of theclean background image B, and the rain streak layer S which contains the rain streaks thatwe would like to remove, as shown in Equation (6).

O = B + S (6)

By estimating S we can, in turn, recover the clean image B through the difference asshown in Equation (7).

B = O− S (7)

The problem of decomposing the rainy image to a clear image and a rain streak layeris an ill-posed problem and to resolve this many different approaches have been proposed.In Appendix A we list the different imaging models that have been proposed for the imagederaining task.

6.3.2. Comparison of Deraining Models

In Table 7 we have collected the methods discussed in Section 3 and classified them interms of quality metrics and processing time. We observe that the best quantitative perfor-mance is obtained by the JORDER [23] and ResGuideNet3 [55] methods. Their performanceis justified by the reported PSNR and SSIM values. More specifically, the JORDER [23]method exhibits the highest PSNR value (namely, 32.95), while ResGuideNet3 [55] exhibitsthe highest SSIM value (estimated at 0.939). A drawback is that, while these two methodsscore high in terms of the PSNR or SSIM value, they both reach an average processingtime and can be classified as near-real-time performance. Two of the methods have an

Sensors 2022, 22, 4707 31 of 44

exceptionally short processing time, namely the PCNet-fast [167] and the LPNET [86].The MSPFN [80], IADN [71] and DDN [17] methods are observed to have an adequateperformance but at the same time they perform in near-real-time settings. We can concludethat for an augmented vision application around rescue mission operations, a model selec-tion strategy could be to select a method that has at least near-real-time processing time,and simultaneously a high value of the exhibited PSNR or SSIM. Using this principle wecan select the PCNet-fast, PCNet or ResGuideNet3 methods. By assuming a high PSNR orSSIM value, we can potentially improve the processing time capabilities of a method sothat it can be feasible to use it for the intended scenario. Such methods are, for instance,IADN, PReNet, MSPFN and RESCAN.

Table 7. Quantitative results of deraining methods along with their running time. ↑means higheris better. In bold we indicate the best result for each dataset.

Dataset Method PSNR ↑ SSIM ↑ FPS ↑ Image Resolution ClassificationRESCAN [87] 30.51 0.882 1.83 512× 512 non-real-timeMSPFN [80] 32.39 0.916 1.97 512× 512 non-real-timePReNet [89] 31.36 0.911 6.13 512× 512 near-real-timeIADN [71] 32.29 0.916 7.57 512× 512 near-real-timeDDC [25] 28.65 0.854 8.00 512× 512 near-real-time

DerainNet [48] 23.38 0.835 13.51 512× 512 near-real-timePCNet [167] 32.03 0.913 16.12 512× 512 near-real-timeUMRL [104] 21.15 0.770 20.00 512× 512 real-time

PCNet-fast [167] 31.45 0.906 35.71 512× 512 real-time

Test1200

LPNET [86] 25.00 0.782 37.03 512× 512 real-time

Rain100LJORDER [23] 32.95 0.921 5.55 481× 321 near-real-time

DDN [17] 31.12 0.926 6.25 481× 321 near-real-timeResGuideNet3 [55] 30.79 0.939 16.66 481× 321 near-real-time

6.3.3. Mathematical Background of Desnowing

Removing snow from a single image is an ill-posed problem. In the literature, it isoften assumed that a snowy image is composed of a clear image and an independent snowmask and the problem is formulated using Equation (8)

K(x) = J(x)(1− Z(x)) + C(x)Z(x) (8)

where K ∈ [0, 1]p×q×3 is a colored snowy image of size p× q, J ∈ [0, 1]p×q×3 is the cor-responding snow-free image, Z ∈ [0, 1]p×q×1 is the snow mask and C is the chromaticaberration map [29].

Recently, Chen et al. [32] reformulated this snow formation model, in order to addi-tionally take the veiling effect into consideration. Their Joint Size and Transparency-AwareSnow Removal (JSTASR) model is inspired by the Atmospheric Scattering Model used fordehazing (for more information the reader can refer to Section 6.3.5) and is formulatedusing Equation (9)

I(x) = K(x)T(x) + A(x)(1− T(x)) (9)

where A is the atmospheric light matrix, J is the scene radiance, T is the transmission mapand K is the snowy image without the veiling effect defined as:

K(x) = J(x)(1− Z(x)R(x)) + C(x)Z(x)R(x) (10)

where R is a binary mask which represents the snow location.

6.3.4. Comparison of Desnowing Models

Table 8 lists the reported quality metrics and processing time throughput of the listedmethods for desnowing discussed in Section 4. Again, in terms of execution time we

Sensors 2022, 22, 4707 32 of 44

use the same classification and the quantitative metrics are the PSNR and SSIM. Clearly,the DesnowGAN method by Jaw et al. [107] is the only method that exhibits real-time per-formance. This method proposes multiple model variations which are capable of generatinga variable number of FPS. In particular, it can generate between 14 and 33 FPS. We keepthe one with the highest number of FPS in the context of assisting FRs. The JSTASR [32],MS-SDN [31] and DesnowNet [29] methods belong to the class of non-real-time methods.It is also interesting to mention that the DesnowNet [29] and the MS-SDN [31] models scorehigher in PSNR and SSIM scores in contrast to the DesnowGAN [107] method. The lattermethod is very fast in terms of FPS, but it scores lower in PSNR and SSIM scores. In thiscase, the DesnowGAN [107] could be a good fit for enhancing the situational awarenessof FRs due to its high speed, but the MS-SDN model [31] could also be an option if weconsider improving its processing time.

Table 8. Quantitative results of desnowing methods along with their running time. ↑means higher isbetter. In bold we indicate the best result among the evaluated methods.

Dataset Method PSNR ↑ SSIM ↑ FPS ↑ Image Resolution ClassificationDesnowNet [29] 30.11 0.930 0.72 640× 480 non-real-time

MS-SDN [31] 29.25 0.936 2.38 640× 480 non-real-timeJSTASR [32] 28.61 0.864 2.77 640× 480 non-real-time

Snow-100K(overall)

DesnowGAN [107] 28.18 0.912 33.33 640× 480 real-time

6.3.5. Mathematical Background of Dehazing

In CV and computer graphics, a hazy image is modeled mathematically using theAtmospheric Scattering Model (ASM) (Equation 11) proposed by McCartney [168].

I(x) = J(x)t(x) + A(1− t(x)) (11)

where x represents the position of pixels, I(x) is the observed hazy image and J(x) thecorresponding haze-free image to be recovered. The first term on the right hand sideof Equation (11) J(x)t(x) is called direct attenuation, while the second term A(1− t(x))airlight. There are two critical parameters in the ASM: the parameter A that denotes theglobal atmospheric light, and the parameter t(x) that is the transmission map, defined asin Equation (12).

t(x) = e−βd(x) (12)

where β is the scattering coefficient of the atmosphere and d(x) is the distance between theobject and the camera.

Hence, solving Equation (11) for J(x), the model for the clean image can be written as

J(x) =1

t(x)I(x)− A

1t(x)

+ A (13)

In clear weather conditions β ≈ 0; therefore, I ≈ J. In hazy images the A and tparameters need to be estimated.

6.3.6. Comparison of Dehazing Models

In Table 9 we have collected the methods discussed in Section 5 and classified themin terms of quality metrics and processing time, based on the best performance found inthe literature. The quality metrics used for evaluation are the PSNR and SSIM metricsand the dataset used for evaluation is the RESIDE—SOTS. As can be seen, the MSFFA-Net [123] has the best PSNR value (36.69), the method proposed by Yu et al. [156] has thebest SSIM value (0.991), while the AOD-Net has the shortest processing time. There is atrend that the faster the model is the worse the metrics are, but there exist some exceptions.FAMED-Net [129] and FD-GAN [109] show a relatively good performance although they

Sensors 2022, 22, 4707 33 of 44

are real-time methods, while the MSCNN [125] is a near-real-time method but has poorerperformance. Since for the specific task we are interested both in execution time and thequality of the output images the method proposed by Yu et al. [156] and the DW-GAN [149]could be a good fit, while finding a way to reduce their processing time is also a great ideafor having a real-time algorithm with great quantitative metrics.

Table 9. Quantitative results of dehazing methods along with their running time. ↑ means higheris better. In bold we indicate the best result among the evaluated methods.

Dataset Method PSNR ↑ SSIM ↑ FPS ↑ Image Resolution ClassificationFFA-Net [121] 36.39 0.988 0.57 1600× 1200 non-real-timeLi et al. [158] 24.44 0.890 0.89 512× 512 non-real-time

MSCNN-HE [126] 21.56 0.860 1.20 427× 370 non-real-timeTDN [134] 34.59 0.975 1.58 1600× 1200 near-real-time

DW-GAN [149] 35.94 0.986 2.08 1600× 1200 near-real-timeLight-DehazeNet [120] 28.39 0.948 2.38 620× 460 non-real-time

PGC [130] 28.78 0.956 3.17 563× 752 near-real-timeMSFFA-Net [123] 36.69 0.990 3.23 620× 460 near-real-timeDehazeNet [118] 21.14 0.847 3.33 620× 460 near-real-time

EPDN [140] 25.06 0.923 3.41 563× 752 near-real-timeGolts et al. [157] 24.08 0.933 3.57 620× 460 near-real-time

GDNet [124] 32.16 0.983 3.60 620× 460 near-real-timeMSCNN [125] 17.57 0.810 3.85 620× 460 near-real-time

YOLY [160] 19.41 0.832 4.76 620× 460 near-real-timeYu et al. [156] 36.61 0.991 11.24 1600× 1200 real-timecGAN [141] 26.63 0.942 19.23 620× 460 real-timeGFN [139] 22.30 0.880 20.40 620× 460 real-time

DCPDN [136] 19.39 0.650 23.98 512× 512 real-timeFD-GAN [109] 23.15 0.920 65.00 1024× 1024 real-timeDMPHN [133] 16.94 0.617 68.96 1600× 1200 real-time

FAMED-Net [129] 25.00 0.917 86.20 620× 460 real-time

SOTS(RESIDE)

AOD-Net [119] 19.06 0.850 232.56 620× 460 real-time

7. Discussion and Conclusions

DL algorithms for enhancing vision in adverse weather conditions have raised a lotof interest in the past few years. This state-of-the-art report presents the vast increase ofresearch in this domain while focusing on the specific case of assisting FRs. We believethat such algorithms will have a profound impact on improving the situational awarenessof FRs in cases where their vision is bound by natural phenomena. We first introducecommonly used datasets, next we present a taxonomy and detailed review of the threefamilies of image restoration methods for adverse weather conditions based on the mostrecent research papers. Attention is also paid to unified image restoration methods that cansimultaneously handle rain, snow and haze. These models can blindly understand whichkind of weather-specific processing should be applied to an input image without requiringthe user to specify in what weather condition the image was captured. Finally, we discussthe suitability of each of the existing algorithms for rescue missions applications. In orderto decide the competence of each method, we present a comparison of them in termsof qualitative metrics and processing time. From our results, we can conclude that thereal-time performance for such image restoration tasks is of paramount importance tointegrating them into systems that can perform live in rescue missions.

Furthermore, there are very few real-world datasets and none of them are focused onthe specific task of rescue missions, thus it will be necessary to create future datasets focusedon rescue missions. We hope that this survey will introduce image restoration methods forFRs to a large research community, thereby facilitating the development of next-generationimage restoration applications focusing on both execution time and performance. Last but

Sensors 2022, 22, 4707 34 of 44

not least, research directions such as unified models that operate under multiple adverseconditions can open new pathways for deploying robust vision applications that will bevaluable both for researchers and experts.

Author Contributions: Conceptualization, S.K.; methodology, S.K., I.G. and V.G.; validation, S.K.,I.G. and V.G.; formal analysis, S.K., I.G. and V.G.; investigation, S.K. and I.G.; resources, D.Z.; datacuration, S.K. and I.G.; writing—original draft preparation, S.K., I.G. and V.G.; writing—review andediting, S.K., I.G., V.G., K.K. and D.Z.; visualization, S.K. and I.G.; supervision, K.K. and D.Z.; projectadministration, D.Z.; funding acquisition, D.Z. All authors have read and agreed to the publishedversion of the manuscript.

Funding: This research has been supported by the European Commission within the context of theproject RESCUER, funded under EU H2020 Grant Agreement 101021836.

Institutional Review Board Statement: Not applicable.

Informed Consent Statement: Not applicable.

Data Availability Statement: Not applicable.

Conflicts of Interest: The authors declare no conflict of interest.

AbbreviationsThe following abbreviations are used in this manuscript:

FR First ResponderCV Computer VisionDL Deep LearningML Machine LearningGPU Graphics Processing UnitCNN Convolutional Neural NetworkGAN Generative Adversarial NetworkAE AutoencoderHF High FrequencyLF Low FrequencyPDU Progressive Dilated UnitUBS Unsupervised Background SegmentationLSTM Long Short-Term MemoryDNN Deep Neural NetworkDWT Discrete Wavelet TransformRL Reinforcement LearningPSNR Peak-Signal-to-Noise RatioSSIM Structural Similarity Index MeasureFPS Frames Per Second

Appendix A

Appendix A.1. Underlying Equations for Image Deraining

This Appendix section supplements Section 6.3.1. Here we discuss several imagingmodels that have been proposed in methodologies on the deraining task. These imagingmodels, in general, make an attempt to separate the clean image data from the rain streakdata in various ways. Due to the complex data structure that exists in real image data,the imaging model equations that have been proposed may not model the underlying dataperfectly or perfectly separate the usable data from the undesired noise. Below, we list thesynthetic imaging models inline along with the references which introduced them.

1. Li et al. [90] model the observed rainy image O as a background layer B and asequence of s rain streak layers St. Each layer St can model rain streaks of differentcharacteristics (e.g., that of orientation and size). Their proposed model is known

Sensors 2022, 22, 4707 35 of 44

in the bibliography as Multiple Rain Streak Layers (MRSL) model and is shown inEquation (A1).

O = B +s

∑t=1

St (A1)

2. The Heavy Rain Model (HRM) proposed by Yang et al. [24] is a convex combinationof two quantities: (1) the underlying image modelled by a set of rain streak layers St(s layers are assumed) and the background image B; (2) the global atmospheric lightmatrix A. The s rain-streak layers St are able to capture the rain streaks with differentcharacteristics. a is the atmospheric transmission parameter. Equation (A2) shows theHRM model.

O = α (B +s

∑t=1

St) + (1− α) A (A2)

3. The model proposed by Xue et al. [54] assumes an image B + R that is gammacorrected by an exponent γ. This image is linearly combined with the atmosphericlight matrix A. This equation is similar to Equation (A2), except for the fact that thelatter requires multiple rain streak layers St. Their proposed model is known in thebibliography as Heavy Rain Model with Low Light (HRMLL) model and is shown inEquation (A3).

O = α(B + R)γ + (1− α)A (A3)

4. Luo et al. [169] introduced the Screen Blend Model (SBM). Their model is inspiredby the Equation (6), but a difference is that in the equation the last term B S issubtracted from the quantity B + S. The entries in the matrix B S weigh each pixelentry by the relative importance of the background pixel value of the associated entryand the corresponding value in matrix S. The entries of this matrix can be regardedas the linear correlation among the signals. Therefore, each pixel intensity associatedwith an entry of O is reduced by the product of the values from B and S. The model isshown in Equation (A4).

O = B + S− B S (A4)

5. The equation Rain Model With Occlusion (RMO) by Liu et al. [170] is similar to theHRM, except that they mutually differ in the last term (1− β) R. In this term βt(i,j)is a cancellation term that signifies whether the pixel at location (i, j) belongs to therain occluded region Ωs (Liu et al. [170] defines Ωs as “the region where the lighttransmittance of rain drop is low”) and matrix R is the rain reliance map.

O = β (B +s

∑t=1

St) + (1− β) R (A5)

6. Yang et al. [103] introduced the Comprehensive Rain Model (CRM). This model issimilar to HRM, however, here a matrix component U is added to the right factor ofthe first term of the equation in case 5. The CRM is shown in Equation (A6).

O = β [(B +s

∑t=1

St) + (1− α)A + U] + (1− β) R (A6)

7. The Depth-aware Rain Model (DRM) by Hu et al. [72] shares a similarity with theprevious equations in that terms that appear in the previous equations also appear inthis last equation but in a different order. The DRM is shown in Equation (A7)

O = β (1−s

∑t=1

St − (1− α)A) +s

∑t=1

St + αA (A7)

Sensors 2022, 22, 4707 36 of 44

8. Equation (A8) by Li et al. [87] combines linearly the clean background image B,the rain streak layers Ri and the atmospheric light matrix A. This linear combinationis especially a convex combination.

O = (1−n

∑i=0

ai)B + a0 A +n

∑i=1

αiRi (A8)

9. Equation (A9) by Yang et al. [23] is similar to Equation (6). The difference in theformer equation is that the entries of matrix R are linearly combined with cancellationentries in a matrix S. When an entry of S is zero, the respective entry in matrix R iscancelled. In turn, when an entry of S equals 1, then the respective intensity valuein the entry of R is promoted. Hence, the term O in O = B + S R is similar to theequation O = B + R, but a difference is that the entries in S R where the respectiveentry in S is zero are cancelled. When each entry in S equals 1, the respective entry inO is modelled as O = B + R.

O = B + S R (A9)

10. Equation (A10) by Su et al. [88] involves the iterative estimation of the backgroundimage B. The equation asserts that the rain streak layer R is determined by the paireddifferences of estimated background images Bi and Bi−1.

N

∑i=2

(Bi − Bi−1) = R (A10)

11. The Raindrop Equation (A11) proposed by Qian et al. [65] models the relationshipbetween the colored image I, a binary mask matrix M with zero-or-one entries,the clean background image B and the matrix R. Matrix R represents the raindropinformation, which is a combination of the background information and the lightreflected by the environment through the raindrops.

I = (1−M) B + R (A11)

12. The Rain Accumulation and Accumulation Flow (RAAF) model (Equation (A12))was introduced in the paper of Yang et al. [103]. The parameter At is the globalatmospheric light. βt is atmospheric transmission. Ut is the rain accumulation flowlayer, and St is the rain streak layer.

Ot = βtBt + (1− βt)At + Ut + St (A12)

Appendix A.2. Loss Functions Employed in the Deraining Problem

In this section, we review some notable loss functions which have been used by recentderaining methods.

The perceptual image similarity loss [171] measures the discrepancy of a derainedimage and the ground-truth of the image. To measure discrepancy, the loss functiontakes the difference between the feature activations of the derained image xD and thefeature activations of the ground-truth image xG. The components of the resulting featureactivation vector are reweighted by a vector θi.

The perceptual contrastive loss [61] is the weighted combination of n fraction termsreferring to the i-th layer Vi of an underlying CNN. Three quantities are considered in thisloss function: the derained image xD, the rainy image xR and the ground-truth image xG.In the fraction terms, the nominator is the `1 distance between the feature activations ofthe derained image and the ground-truth image. The denominator measures the distancebetween the feature activations of the i-th CNN network for the derained image xD andthe rainy image xR.

Sensors 2022, 22, 4707 37 of 44

The segmentation loss used in [61] is also commonly known as the focal loss func-tion [172]. The probability pij models the likelihood that the pixel at position (i, j) belongsto a particular class, and a and γ are standard parameters. The function computes theaverage bit count corresponding to the probability value pij, each weighted by the factor−a(1− pij)

γ. In this way, the supervision focuses on hard examples while reducing theloss contribution from easy examples.

The margin loss considers two extreme values m+ and m− and measures their relationto the squared `2 norm of a vector vk. If we consider two points m− and m+ on a straightline, then ||vk||2 can be placed at three possible point locations. Based on this location,the max operators of Tk and 1− Tk are either positive or are cancelled out, and the lossfunction varies among a minimum and a maximum value.

The color cycle-consistency loss used in [173] is a trivial modification of the cycle-consistency loss, originally introduced in [67]. In a Cycle-GAN, a network involvinga generator and a discriminator function is considered. This network employs a cycle-consistency constraint which ensures that a datum that is projected via projection functionF can be back-projected to the original vector space by another function G. The colorcycle-consistency loss function applies the cycle consistency constraint on the three colorchannels of an RGB image matrix of reference.

The identity preserving loss used in [174] considers the absolute error of the coor-dinates of two vectors. The first vector is a vector n and the mapping GN2R(n), and thesecond vector is a vector r and its projection vector GR2N(r). The target in this loss functionis to minimize the average value of the `1 vector norm of the difference of the two pairsof vectors.

The total variation loss is the sum of the absolute differences for neighboring valuesof an input image. It contributes to generating more smooth images while removing therough textures of the generated denoised image.

Adversarial loss [64] measures the distance between distribution of the generatedimages and the distribution of the ground truth images. In these settings, an adversarial net-work can have two loss functions, one for the generator who tries too fool the disciminator,and one for the discriminator who tries to classify a forged distribution from a real one.

The attention loss function used in [78] is one out of many loss functions designed forserving an attention mechanism in deep neural networks. This loss function considers the`2 matrix norm of the difference of an attention-mask matrix referring to an s-th scale and atarget attention matrix.

The Charbonnier loss function is a matrix function that involves the `2 matrix norm ofthe input variable and a lag term ε that ensures the numerical stability of the loss function.Since the `2 matrix norm of the input matrix variable is non-negative, the Charbonnier lossfunction adds a small parameter ε2 to the squared `2 matrix norm that ensures that thefinal loss function value equals at least ε.

Structural similarity is used in the deraining problem as a surrogate loss functionto optimize the weights of a deep neural network. Similarly, the peak-signal-to-noiseratio metric, which is also a no-reference metric for measuring the quality of an image,is also used as a surrogate loss function in some of the works on the deraining problem;for example, see [175]. Both functions are a natural choice of loss functions for a deepneural network capable of solving the single image deraining problem. Since in deepneural networks we seek to minimize the underlying loss function, we need to minimizethe negative of either function. In fact, in practice we want to maximize terms involvingthe structural similarity and peak signal-to-noise, usually taken to be the mean structuralsimilarity value and the mean peak signal-to-noise ratio.

Appendix A.3. Loss Functions Used by Desnowing Methods

In this part of the Appendix, we list the loss functions that are introduced by the DL-based single image desnowing methods available in Section 4.1. Except simple cases of lossfunctions that are common in the general literature of DL, Table A1 lists several important

Sensors 2022, 22, 4707 38 of 44

loss functions which do not fall within the common loss function cases. Importantly, thesenon-trivial loss functions are themselves alterations of basic loss functions. For instance,the complex wavelet loss function in the first row of Table A1 is a modified Charbonnierloss function. Importantly, the rest of the loss functions make use of the `1, `2 and Frobeniusnorms, respectively.

Table A1. Listing of loss functions used by desnowing methods.

Loss Function Ref. Description

∑i

√||(ηπ

i − ηπi ) + j(ηπ

i − ηπi )|2 + ε2 Chen et al. [30] Complex Wavelet Loss

||CC(J)− CC(JGT)||1 Chen et al. [30] Contradict Channel Loss1

Cj HjWj||φj(y)− φj(y)||22 Chen et al. [30] Perceptual Loss

∑τi=0 ||P2i (m)− P2i (m)||22 Liu et al. [29] Lightweight Pyramid Loss1N ∑N

i=1 ||xi − f (xi)||2 Liu et al. [29] Total Variation Loss||Gφ

j (y)− Gφj (y)||

2F Liu et al. [29] Style Loss

References1. Asami, K.; Shono Fujita, K.H.; Hatayama, M. Data Augmentation with Synthesized Damaged Roof Images Generated by GAN.

In Proceedings of the ISCRAM 2022 Conference Proceedings, 19th International Conference on Information Systems for CrisisResponse and Management, Tarbes, France, 7–9 November 2022.

2. Algiriyage, N.; Prasanna, R.; Doyle, E.E.; Stock, K.; Johnston, D. Traffic Flow Estimation based on Deep Learning for EmergencyTraffic Management using CCTV Images. In Proceedings of the 17th International Conference on Information Systems for CrisisResponse and Management, Blacksburg, VA, USA, 24–27 May 2020; pp. 100–109.

3. Algiriyage, N.; Prasanna, R.; Doyle, E.E.; Stock, K.; Johnston, D.; Punchihewa, M.; Jayawardhana, S. Towards Real-time TrafficFlow Estimation using YOLO and SORT from Surveillance Video Footage. In Proceedings of the 18th International Conferenceon Information Systems for Crisis Response and Management, Blacksburg, VA, USA, 23–26 May 2021.

4. Correia, H.R.; da Costa Rubim, I.; Dias, A.F.; França, J.B.; Borges, M.R. Drones to the Rescue: A Support Solution for EmergencyResponse. In Proceedings of the ISCRAM 2020 Conference Proceedings, 17th International Conference on Information Systemsfor Crisis Response and Management, Blacksburg, VA, USA, 24–27 May 2020.

5. Sainidis, D.; Tsiakmakis, D.; Konstantoudakis, K.; Albanis, G.; Dimou, A.; Daras, P. Single-handed gesture UAV control and videofeed AR visualization for first responders. In Proceedings of the International Conference on Information Systems for CrisisResponse and Management (ISCRAM), Blacksburg, VA, USA, 23–26 May 2021; pp. 23–26.

6. Zaffaroni, M.; Rossi, C. Water Segmentation with Deep Learning Models for Flood Detection and Monitoring. In Proceedings ofthe Conference on Information Systems for Crisis Response and Management (ISCRAM 2020), Blacksburg, VA, USA, 24–27 May2020; pp. 24–27.

7. Rothmeier, T.; Huber, W. Performance Evaluation of Object Detection Algorithms under Adverse Weather Conditions. InProceedings of the International Conference on Intelligent Transport Systems, Indianapolis, IN, USA, 20–23 September 2020;Springer: Berlin/Heidelberg, Germany, 2020; pp. 211–222.

8. Pfeuffer, A.; Dietmayer, K. Optimal Sensor Data Fusion Architecture for Object Detection in Adverse Weather Conditions. InProceedings of the 2018 21st International Conference on Information Fusion (FUSION), Cambridge, UK, 10–13 July 2018; pp. 1–8.[CrossRef]

9. Hasirlioglu, S.; Riener, A. Challenges in object detection under rainy weather conditions. In Proceedings of the First InternationalConference on Intelligent Transport Systems, Funchal, Portugal, 16–18 March 2018; Springer: Berlin/Heidelberg, Germany;pp. 53–65.

10. Chaturvedi, S.S.; Zhang, L.; Yuan, X. Pay “Attention” to Adverse Weather: Weather-aware Attention-based Object Detection.arXiv 2022, arXiv:2204.10803.

11. Morrison, K.; Choong, Y.Y.; Dawkins, S.; Prettyman, S.S. Communication Technology Problems and Needs of Rural FirstResponders. Public Law 2012, 112, 96.

12. Su, J.; Xu, B.; Yin, H. A survey of deep learning approaches to image restoration. Neurocomputing 2022, 487, 46–65. [CrossRef]13. Jin, X.; Chen, Z.; Lin, J.; Chen, Z.; Zhou, W. Unsupervised single image deraining with self-supervised constraints. In Proceedings

of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2761–2765.14. Yadav, S.; Mehra, A.; Rohmetra, H.; Ratnakumar, R.; Narang, P. DerainGAN: Single image deraining using wasserstein GAN.

Multimed. Tools Appl. 2021, 80, 36491–36507. [CrossRef]15. Guo, Z.; Hou, M.; Sima, M.; Feng, Z. DerainAttentionGAN: Unsupervised single-image deraining using attention-guided

generative adversarial networks. Signal Image Video Process. 2022, 16, 185–192. [CrossRef]16. Wei, W.; Meng, D.; Zhao, Q.; Xu, Z.; Wu, Y. Semi-Supervised Transfer Learning for Image Rain Removal. In Proceedings of the

IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019.

Sensors 2022, 22, 4707 39 of 44

17. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing Rain From Single Images via a Deep Detail Network. InProceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017.

18. Schaefer, G.; Stich, M. UCID: An uncompressed color image database. In Storage and Retrieval Methods and Applications forMultimedia 2004; International Society for Optics and Photonics: Bellingham, WA, USA, 2003; Volume 5307, pp. 472–480.

19. Martin, D.R.; Fowlkes, C.C.; Malik, J. Learning to detect natural image boundaries using local brightness, color, and texture cues.IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 530–549. [CrossRef]

20. Zhang, H.; Patel, V.M. Density-Aware Single Image De-Raining Using a Multi-Stream Dense Network. In Proceedings of theIEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018.

21. Zhang, H.; Sindagi, V.; Patel, V.M. Image De-Raining Using a Conditional Generative Adversarial Network. IEEE Trans. CircuitsSyst. Video Technol. 2020, 30, 3943–3956. [CrossRef]

22. Li, Y.; Tan, R.T.; Guo, X.; Lu, J.; Brown, M.S. Rain streak removal using layer priors. In Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2736–2744.

23. Yang, W.; Tan, R.T.; Feng, J.; Guo, Z.; Yan, S.; Liu, J. Joint Rain Detection and Removal from a Single Image with ContextualizedDeep Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 1377–1393. [CrossRef]

24. Yang, W.; Tan, R.T.; Feng, J.; Liu, J.; Guo, Z.; Yan, S. Deep Joint Rain Detection and Removal From a Single Image. In Proceedingsof the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017.

25. Li, S.; Ren, W.; Zhang, J.; Yu, J.; Guo, X. Single image rain removal via a deep decomposition–composition network. Comput. Vis.Image Underst. 2019, 186, 48–57. [CrossRef]

26. Kenk, M.A.; Hassaballah, M. DAWN: Vehicle detection in adverse weather nature dataset. arXiv 2020, arXiv:2008.05402.27. Chen, J.; Tan, C.H.; Hou, J.; Chau, L.P.; Li, H. Robust video content alignment and compensation for rain removal in a cnn

framework. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23June 2018; pp. 6286–6295.

28. Wang, T.; Yang, X.; Xu, K.; Chen, S.; Zhang, Q.; Lau, R.W. Spatial attentive single-image deraining with a high quality real raindataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17June 2019; pp. 12270–12279.

29. Liu, Y.F.; Jaw, D.W.; Huang, S.C.; Hwang, J.N. DesnowNet: Context-aware deep network for snow removal. IEEE Trans. ImageProcess. 2018, 27, 3064–3073. [CrossRef] [PubMed]

30. Chen, W.T.; Fang, H.Y.; Hsieh, C.L.; Tsai, C.C.; Chen, I.; Ding, J.J.; Kuo, S.Y. ALL Snow Removed: Single Image DesnowingAlgorithm Using Hierarchical Dual-Tree Complex Wavelet Representation and Contradict Channel Loss. In Proceedings of theIEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4196–4205.

31. Li, P.; Yun, M.; Tian, J.; Tang, Y.; Wang, G.; Wu, C. Stacked dense networks for single-image snow removal. Neurocomputing 2019,367, 152–163. [CrossRef]

32. Chen, W.T.; Fang, H.Y.; Ding, J.J.; Tsai, C.C.; Kuo, S.Y. JSTASR: Joint size and transparency-aware snow removal algorithm based onmodified partial convolution and veiling effect removal. In European Conference on Computer Vision; Springer: Berlin/Heidelberg,Germany, 2020; pp. 754–770.

33. Tarel, J.P.; Hautiere, N.; Cord, A.; Gruyer, D.; Halmaoui, H. Improved visibility of road scene images under heterogeneous fog. InProceedings of the 2010 IEEE Intelligent Vehicles Symposium, La Jolla, CA, USA, 21–24 June 2010; pp. 478–485.

34. Tarel, J.P.; Hautiere, N.; Caraffa, L.; Cord, A.; Halmaoui, H.; Gruyer, D. Vision enhancement in homogeneous and heterogeneousfog. IEEE Intell. Transp. Syst. Mag. 2012, 4, 6–20. [CrossRef]

35. El Khoury, J.; Thomas, J.B.; Mansouri, A. A color image database for haze model and dehazing methods evaluation. InInternational Conference on Image and Signal Processing; Springer: Berlin/Heidelberg, Germany, 2016; pp. 109–117.

36. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking single-image dehazing and beyond. IEEE Trans. ImageProcess. 2018, 28, 492–505. [CrossRef] [PubMed]

37. Ancuti, C.; Ancuti, C.O.; De Vleeschouwer, C. D-hazy: A dataset to evaluate quantitatively dehazing algorithms. In Proceedingsof the 2016 IEEE international conference on image processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2226–2230.

38. Ancuti, C.; Ancuti, C.O.; Timofte, R.; De Vleeschouwer, C. I-HAZE: A dehazing benchmark with real hazy and haze-free indoorimages. In International Conference on Advanced Concepts for Intelligent Vision Systems; Springer: Berlin/Heidelberg, Germany, 2018;pp. 620–631.

39. Ancuti, C.O.; Ancuti, C.; Timofte, R.; De Vleeschouwer, C. O-haze: A dehazing benchmark with real hazy and haze-free outdoorimages. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT,USA, 18–23 June 2018; pp. 754–762.

40. Ancuti, C.O.; Ancuti, C.; Sbert, M.; Timofte, R. Dense-haze: A benchmark for image dehazing with dense-haze and haze-freeimages. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September2019; pp. 1014–1018.

41. Ancuti, C.O.; Ancuti, C.; Timofte, R. NH-HAZE: An image dehazing benchmark with non-homogeneous hazy and haze-freeimages. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA,14–19 June 2020; pp. 444–445.

Sensors 2022, 22, 4707 40 of 44

42. Scharstein, D.; Hirschmüller, H.; Kitajima, Y.; Krathwohl, G.; Nešic, N.; Wang, X.; Westling, P. High-resolution stereo datasetswith subpixel-accurate ground truth. In German Conference on Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2014;pp. 31–42.

43. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In EuropeanConference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760.

44. Zhang, Y.; Ding, L.; Sharma, G. Hazerd: An outdoor scene dataset and benchmark for single image dehazing. In Proceedings ofthe 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3205–3209.

45. Zhao, S.; Zhang, L.; Huang, S.; Shen, Y.; Zhao, S. Dehazing evaluation: Real-world benchmark datasets, criteria, and baselines.IEEE Trans. Image Process. 2020, 29, 6947–6962. [CrossRef]

46. Zheng, Z.; Ren, W.; Cao, X.; Hu, X.; Wang, T.; Song, F.; Jia, X. Ultra-high-definition image dehazing via multi-guided bilaterallearning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN,USA, 20–25 June 2021; pp. 16180–16189.

47. Zhang, X.; Dong, H.; Pan, J.; Zhu, C.; Tai, Y.; Wang, C.; Li, J.; Huang, F.; Wang, F. Learning to restore hazy video: A new real-worlddataset and a new method. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville,TN, USA, 20–25 June 2021; pp. 9239–9248.

48. Fu, X.; Huang, J.; Ding, X.; Liao, Y.; Paisley, J. Clearing the skies: A deep network architecture for single-image rain removal.IEEE Trans. Image Process. 2017, 26, 2944–2956. [CrossRef]

49. Li, S.; Ren, W.; Wang, F.; Araujo, I.B.; Tokuda, E.K.; Junior, R.H.; Cesar, R.M., Jr.; Wang, Z.; Cao, X. A Comprehensive BenchmarkAnalysis of Single Image Deraining: Current Challenges and Future Perspectives. Int. J. Comput. Vis. 2021, 129, 1301–1322.[CrossRef]

50. Wang, H.; Wu, Y.; Li, M.; Zhao, Q.; Meng, D. Survey on rain removal from videos or a single image. Sci. China Inf. Sci. 2021,65, 111101. [CrossRef]

51. Du, S.; Liu, Y.; Zhao, M.; Shi, Z.; You, Z.; Li, J. A comprehensive survey: Image deraining and stereo-matching task-drivenperformance analysis. IET Image Process. 2022, 16, 11–28.

52. Hnewa, M.; Radha, H. Object Detection Under Rainy Conditions for Autonomous Vehicles: A Review of State-of-the-Art andEmerging Techniques. IEEE Signal Process. Mag. 2021, 38, 53–67. [CrossRef]

53. Yang, W.; Tan, R.T.; Wang, S.; Fang, Y.; Liu, J. Single Image Deraining: From Model-Based to Data-Driven and Beyond. IEEETrans. Pattern Anal. Mach. Intell. 2021, 43, 4059–4077. [CrossRef] [PubMed]

54. Yang, W.; Liu, J.; Yang, S.; Guo, Z. Scale-free single image deraining via visibility-enhanced recurrent wavelet learning. IEEETrans. Image Process. 2019, 28, 2948–2961. [CrossRef] [PubMed]

55. Fan, Z.; Wu, H.; Fu, X.; Huang, Y.; Ding, X. Residual-guide network for single image deraining. In Proceedings of the 26th ACMInternational Conference on Multimedia, Seoul, Korea, 22–26 October 2018; pp. 1751–1759.

56. Li, G.; He, X.; Zhang, W.; Chang, H.; Dong, L.; Lin, L. Non-locally enhanced encoder-decoder network for single image de-raining.In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Korea, 22–26 October 2018; pp. 1056–1064.

57. Pan, J.; Liu, S.; Sun, D.; Zhang, J.; Liu, Y.; Ren, J.; Li, Z.; Tang, J.; Lu, H.; Tai, Y.W.; et al. Learning dual convolutional neuralnetworks for low-level vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt LakeCity, UT, USA, 18–23 June 2018; pp. 3070–3079.

58. Huang, Z.; Zhang, J. Dynamic Multi-Domain Translation Network for Single Image Deraining. In Proceedings of the 2021 IEEEInternational Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 1754–1758. [CrossRef]

59. Cho, J.; Kim, S.; Sohn, K. Memory-guided Image De-raining Using Time-Lapse Data. arXiv 2022, arXiv:2201.01883.60. Guo, Q.; Sun, J.; Juefei-Xu, F.; Ma, L.; Lin, D.; Feng, W.; Wang, S. Uncertainty-Aware Cascaded Dilation Filtering for High-Efficiency

Deraining. arXiv 2022, arXiv:2201.02366.61. Zheng, S.; Lu, C.; Wu, Y.; Gupta, G. SAPNet: Segmentation-Aware Progressive Network for Perceptual Contrastive Deraining. In

Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022;pp. 52–62.

62. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Wang, Z.; Lin, C.W. PCNET: Progressive Coupled Network for Real-Time Image Deraining.In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September2021; pp. 1759–1763.

63. Rai, S.N.; Saluja, R.; Arora, C.; Balasubramanian, V.N.; Subramanian, A.; Jawahar, C. FLUID: Few-Shot Self-Supervised ImageDeraining. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8January 2022; pp. 3077–3086.

64. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarialnets. In Proceedings of the Advances in Neural Information Processing Systems 27, Montreal, QC, Canada, 8–13 December 2014.

65. Qian, R.; Tan, R.T.; Yang, W.; Su, J.; Liu, J. Attentive Generative Adversarial Network for Raindrop Removal From a SingleImage. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA,18–23 June 2018.

66. Wei, Y.; Zhang, Z.; Wang, Y.; Xu, M.; Yang, Y.; Yan, S.; Wang, M. DerainCycleGAN: Rain attentive CycleGAN for single imagederaining and rainmaking. IEEE Trans. Image Process. 2021, 30, 4788–4801. [CrossRef]

Sensors 2022, 22, 4707 41 of 44

67. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. InProceedings of the 2017 IEEE International Conference Computer Vision (ICCV), Venice, Italy, 22–29 October 2017.

68. Yang, F.; Ren, J.; Lu, Z.; Zhang, J.; Zhang, Q. Rain-component-aware capsule-GAN for single image de-raining. Pattern Recognit.2022, 123, 108377. [CrossRef]

69. Yan, X.; Loke, Y.R. RainGAN: Unsupervised Raindrop Removal via Decomposition and Composition. In Proceedings of theIEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops, Waikoloa, HI, USA, 4–8 January 2022;pp. 14–23.

70. Li, R.; Cheong, L.F.; Tan, R.T. Heavy rain image restoration: Integrating physics model and conditional adversarial learning. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019;pp. 1633–1642.

71. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Han, Z.; Lu, T.; Huang, B.; Jiang, J. Decomposition makes better rain removal: An improvedattention-guided deraining network. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 3981–3995. [CrossRef]

72. Hu, X.; Fu, C.W.; Zhu, L.; Heng, P.A. Depth-attentional features for single-image rain removal. In Proceedings of the IEEE/CVFConference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8022–8031.

73. Wang, Q.; Sun, G.; Fan, H.; Li, W.; Tang, Y. APAN: Across-scale Progressive Attention Network for Single Image Deraining. IEEESignal Process. Lett. 2021, 29, 159–163. [CrossRef]

74. Zhang, H.; Xie, Q.; Lu, B.; Gai, S. Dual attention residual group networks for single image deraining. Digit. Signal Process. 2021,116, 103106. [CrossRef]

75. Wei, Y.; Zhang, Z.; Xu, M.; Hong, R.; Fan, J.; Yan, S. Robust Attention Deraining Network for Synchronous Rain Streaks andRaindrops Removal. TechRxiv 2021. [CrossRef]

76. Zhou, J.; Leong, C.; Lin, M.; Liao, W.; Li, C. Task Adaptive Network for Image Restoration with Combined Degradation Factors.In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops, Waikoloa, HI,USA, 4–8 January 2022; pp. 1–8.

77. Lin, X.; Huang, Q.; Huang, W.; Tan, X.; Fang, M.; Ma, L. Single Image Deraining via detail-guided Efficient Channel AttentionNetwork. Comput. Graph. 2021, 97, 117–125. [CrossRef]

78. Li, Y.; Monno, Y.; Okutomi, M. Single Image Deraining Network with Rain Embedding Consistency and Layered LSTM. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022;pp. 4060–4069.

79. Zhang, Y.; Liu, Y.; Li, Q.; Wang, J.; Qi, M.; Sun, H.; Xu, H.; Kong, J. A Lightweight Fusion Distillation Network for ImageDeblurring and Deraining. Sensors 2021, 21, 5312. [CrossRef] [PubMed]

80. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Huang, B.; Luo, Y.; Ma, J.; Jiang, J. Multi-Scale Progressive Fusion Network for Single ImageDeraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA,13–19 June 2020.

81. Fu, B.; Jiang, Y.; Wang, H.; Wang, Q.; Gao, Q.; Tang, Y. Context-wise attention-guided network for single image deraining.Electron. Lett. 2022, 58, 148–150. [CrossRef]

82. Wang, Z.; Wang, C.; Su, Z.; Chen, J. Dense Feature Pyramid Grids Network for Single Image Deraining. In Proceedings of theICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada,6–11 June 2021; pp. 2025–2029. [CrossRef]

83. Yamamichi, K.; Han, X.H. Multi-scale Residual Aggregation Deraining Network with Spatial Context-aware Pooling andActivation. In Proceedings of the British Machine Vision Conference, Virtual, 23–25 November 2021.

84. Jasuja, C.; Gupta, H.; Gupta, D.; Parihar, A.S. SphinxNet-A Lightweight Network for Single Image Deraining. In Proceedings ofthe 2021 International Conference on Intelligent Technologies (CONIT), Hubli, India, 25–27 June 2021; pp. 1–6.

85. Zheng, Y.; Yu, X.; Liu, M.; Zhang, S. Single-image deraining via recurrent residual multiscale networks. IEEE Trans. Neural Netw.Learn. Syst. 2022, 33, 1310–1323. [CrossRef] [PubMed]

86. Fu, X.; Liang, B.; Huang, Y.; Ding, X.; Paisley, J. Lightweight pyramid networks for image deraining. IEEE Trans. Neural Netw.Learn. Syst. 2019, 31, 1794–1807. [CrossRef]

87. Li, X.; Wu, J.; Lin, Z.; Liu, H.; Zha, H. Rescan: Recurrent squeeze-and-excitation context aggregation net. In Proceedings of theEuropean Conference on Computer Vision, Munich, Germany, 8–14 September 2018.

88. Su, Z.; Zhang, Y.; Zhang, X.P.; Qi, F. Non-local channel aggregation network for single image rain removal. Neurocomputing 2022,469, 261–272.: 10.1016/j.neucom.2021.10.052. [CrossRef]

89. Ren, D.; Zuo, W.; Hu, Q.; Zhu, P.; Meng, D. Progressive image deraining networks: A better and simpler baseline. In Proceedingsof the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 3937–3946.

90. Li, R.; Cheong, L.F.; Tan, R.T. Single image deraining using scale-aware multi-stage recurrent network. arXiv 2017,arXiv:1712.06830.

91. Zhang, K.; Luo, W.; Yu, Y.; Ren, W.; Zhao, F.; Li, C.; Ma, L.; Liu, W.; Li, H. Beyond Monocular Deraining: Parallel Stereo DerainingNetwork Via Semantic Prior. arXiv 2021, arXiv:2105.03830.

92. Wang, C.; Zhu, H.; Fan, W.; Wu, X.M.; Chen, J. Single image rain removal using recurrent scale-guide networks. Neurocomputing2022, 467, 242–255. [CrossRef]

Sensors 2022, 22, 4707 42 of 44

93. Cai, L.; Li, S.Y.; Ren, D.; Wang, P. Dual recursive network for fast image deraining. In Proceedings of the 2019 IEEE internationalconference on image processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2756–2760.

94. Kim, J.H.; Sim, J.Y.; Kim, C.S. Video deraining and desnowing using temporal correlation and low-rank matrix completion. IEEETrans. Image Process. 2015, 24, 2658–2670. [CrossRef]

95. Xue, X.; Meng, X.; Ma, L.; Wang, Y.; Liu, R.; Fan, X. Searching Frame-Recurrent Attentive Deformable Network for Real-TimeVideo Deraining. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China,5–9 July 2021; pp. 1–6. [CrossRef]

96. Zhang, K.; Li, D.; Luo, W.; Lin, W.Y.; Zhao, F.; Ren, W.; Liu, W.; Li, H. Enhanced Spatio-Temporal Interaction Learning for VideoDeraining: A Faster and Better Framework. arXiv 2021, arXiv:2103.12318.

97. Yan, W.; Tan, R.T.; Yang, W.; Dai, D. Self-Aligned Video Deraining With Transmission-Depth Consistency. In Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11966–11976.

98. Yang, Y.; Lu, H. A Fast and Efficient Network for Single Image Deraining. In Proceedings of the ICASSP 2021–2021 IEEE Interna-tional Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2030–2034.[CrossRef]

99. Wang, M.; Li, C.; Ke, F. Recurrent multi-level residual and global attention network for single image deraining. Neural Comput.Appl. 2022, 1–12.. [CrossRef]

100. Deng, L.; Xu, G.; Zhu, H.; Bao, B.K. RoDeRain: Rotational Video Derain via Nonconvex and Nonsmooth Optimization. Mob.Networks Appl. 2021, 26, 57–66. [CrossRef]

101. Kulkarni, A.; Patil, P.W.; Murala, S. Progressive Subtractive Recurrent Lightweight Network for Video Deraining. IEEE SignalProcess. Lett. 2021, 29, 229–233. [CrossRef]

102. Yue, Z.; Xie, J.; Zhao, Q.; Meng, D. Semi-Supervised Video Deraining With Dynamical Rain Generator. In Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 642–652.

103. Yang, W.; Liu, J.; Feng, J. Frame-Consistent Recurrent Video Deraining With Dual-Level Flow. In Proceedings of the IEEE/CVFConference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019.

104. Yasarla, R.; Patel, V.M. Uncertainty guided multi-scale residual learning-using a cycle spinning cnn for single image de-raining.In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June2019; pp. 8405–8414.

105. Xue, X.; Meng, X.; Ma, L.; Liu, R.; Fan, X. Gta-net: Gradual temporal aggregation network for fast video deraining. In Proceedingsof the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON,Canada, 6–11 June 2021; pp. 2020–2024.

106. Li, Z.; Zhang, J.; Fang, Z.; Huang, B.; Jiang, X.; Gao, Y.; Hwang, J.N. Single image snow removal via composition generativeadversarial networks. IEEE Access 2019, 7, 25016–25025. [CrossRef]

107. Jaw, D.W.; Huang, S.C.; Kuo, S.Y. DesnowGAN: An efficient single image snow removal framework using cross-resolution lateralconnection and GANs. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1342–1350. [CrossRef]

108. Zhang, K.; Li, R.; Yu, Y.; Luo, W.; Li, C. Deep dense multi-scale network for snow removal using semantic and depth priors. IEEETrans. Image Process. 2021, 30, 7419–7431. [CrossRef]

109. Dong, Y.; Liu, Y.; Zhang, H.; Chen, S.; Qiao, Y. FD-GAN: Generative adversarial networks with fusion-discriminator for singleimage dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020;Volume 34, pp. 10729–10736.

110. Lee, S.; Yun, S.; Nam, J.H.; Won, C.S.; Jung, S.W. A review on dark channel prior based image dehazing algorithms. EURASIP J.Image Video Process. 2016, 2016, 1–23. [CrossRef]

111. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010,33, 2341–2353.

112. Tan, R.T. Visibility in bad weather from a single image. In Proceedings of the 2008 IEEE Conference on Computer Vision andPattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8.

113. Zhu, Q.; Mai, J.; Shao, L. Single Image Dehazing Using Color Attenuation Prior. In Proceedings of the British Machine VisionConference BMVC Citeseer, Nottingham, UK, 7–10 September 2014.

114. Berman, D.; Avidan, S.; Treibitz, T. Non-local image dehazing. In Proceedings of the IEEE Conference on Computer Vision andPattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1674–1682.

115. Tang, K.; Yang, J.; Wang, J. Investigating haze-relevant features in a learning framework for image dehazing. In Proceedings ofthe IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2995–3000.

116. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process.2015, 24, 3522–3533.

117. Mai, J.; Zhu, Q.; Wu, D.; Xie, Y.; Wang, L. Back propagation neural network dehazing. In Proceedings of the 2014 IEEEInternational Conference on Robotics and Biomimetics (ROBIO 2014), Bali, Indonesia, 5–10 December 2014; pp. 1433–1438.

118. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. ImageProcess. 2016, 25, 5187–5198. [CrossRef] [PubMed]

119. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE InternationalConference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778.

Sensors 2022, 22, 4707 43 of 44

120. Ullah, H.; Muhammad, K.; Irfan, M.; Anwar, S.; Sajjad, M.; Imran, A.S.; de Albuquerque, V.H.C. Light-DehazeNet: A NovelLightweight CNN Architecture for Single Image Dehazing. IEEE Trans. Image Process. 2021, 30, 8968–8982. [CrossRef] [PubMed]

121. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature fusion attention network for single image dehazing. In Proceedings ofthe AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11908–11915.

122. Wu, H.; Qu, Y.; Lin, S.; Zhou, J.; Qiao, R.; Zhang, Z.; Xie, Y.; Ma, L. Contrastive learning for compact single image dehazing. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021;pp. 10551–10560.

123. Hu, B. Multi-Scale Feature Fusion Network with Attention for Single Image Dehazing. Pattern Recognit. Image Anal. 2021,31, 608–615.

124. Liu, X.; Ma, Y.; Shi, Z.; Chen, J. Griddehazenet: Attention-based multi-scale network for image dehazing. In Proceedings of theIEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 7314–7323.

125. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks. InEuropean Conference on Computer Vision, Glasgow, UK, 23–28 August 2016; pp. 154–169.

126. Ren, W.; Pan, J.; Zhang, H.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks withholistic edges. Int. J. Comput. Vis. 2020, 128, 240–259. [CrossRef]

127. Wang, J.; Li, C.; Xu, S. An ensemble multi-scale residual attention network (EMRA-net) for image Dehazing. Multimed. Tools Appl.2021, 80, 29299–29319. [CrossRef]

128. Dong, H.; Pan, J.; Xiang, L.; Hu, Z.; Zhang, X.; Wang, F.; Yang, M.H. Multi-Scale boosted dehazing network with dense featurefusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June2020; pp. 2157–2167.

129. Zhang, J.; Tao, D. FAMED-Net: A fast and accurate multi-scale end-to-end dehazing network. IEEE Trans. Image Process. 2019,29, 72–84. [CrossRef]

130. Zhao, D.; Xu, L.; Ma, L.; Li, J.; Yan, Y. Pyramid global context network for image dehazing. IEEE Trans. Circuits Syst. Video Technol.2020, 31, 3037–3050. [CrossRef]

131. Sheng, J.; Lv, G.; Du, G.; Wang, Z.; Feng, Q. Multi-Scale residual attention network for single image dehazing. Digit. SignalProcess. 2022, 121, 103327. [CrossRef]

132. Fan, G.; Hua, Z.; Li, J. Multi-scale depth information fusion network for image dehazing. Appl. Intell. 2021, 51, 7262–7280.[CrossRef]

133. Das, S.D.; Dutta, S. Fast deep multi-patch hierarchical network for nonhomogeneous image dehazing. In Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 482–483.

134. Liu, J.; Wu, H.; Xie, Y.; Qu, Y.; Ma, L. Trident dehazing network. In Proceedings of the IEEE/CVF Conference on ComputerVision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 430–431.

135. Jo, E.; Sim, J.Y. Multi-Scale Selective Residual Learning for Non-Homogeneous Dehazing. In Proceedings of the IEEE/CVFConference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 507–515.

136. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3194–3203.

137. Zhu, H.; Peng, X.; Chandrasekhar, V.; Li, L.; Lim, J.H. DehazeGAN: When Image Dehazing Meets Differential Programming. InProceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 1234–1240.

138. Yang, X.; Xu, Z.; Luo, J. Towards perceptual image dehazing by physics-based disentanglement and adversarial training. InProceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32.

139. Ren, W.; Ma, L.; Zhang, J.; Pan, J.; Cao, X.; Liu, W.; Yang, M.H. Gated fusion network for single image dehazing. In Proceedingsof the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3253–3261.

140. Qu, Y.; Chen, Y.; Huang, J.; Xie, Y. Enhanced pix2pix dehazing network. In Proceedings of the IEEE/CVF Conference onComputer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 8160–8168.

141. Li, R.; Pan, J.; Li, Z.; Tang, J. Single image dehazing via conditional generative adversarial network. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8202–8211.

142. Kan, S.; Zhang, Y.; Zhang, F.; Cen, Y. A GAN-based input-size flexibility model for single image dehazing. Signal Process. ImageCommun. 2022, 102, 116599. [CrossRef]

143. Engin, D.; Genç, A.; Kemal Ekenel, H. Cycle-dehaze: Enhanced cyclegan for single image dehazing. In Proceedings of the IEEEConference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 825–833.

144. Dudhane, A.; Murala, S. Cdnet: Single image de-hazing using unpaired adversarial training. In Proceedings of the 2019 IEEEWinter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1147–1155.

145. Liu, W.; Hou, X.; Duan, J.; Qiu, G. End-to-end single image fog removal using enhanced cycle consistent adversarial networks.IEEE Trans. Image Process. 2020, 29, 7819–7833. [CrossRef]

146. Jin, Y.; Gao, G.; Liu, Q.; Wang, Y. Unsupervised conditional disentangle network for image dehazing. In Proceedings of the 2020IEEE International Conference on Image Processing (ICIP), Abu Dhabi, UAE, 25–28 October 2020; pp. 963–967.

147. Mo, Y.; Li, C.; Zheng, Y.; Wu, X. DCA-CycleGAN: Unsupervised single image dehazing using Dark Channel Attention optimizedCycleGAN. J. Vis. Commun. Image Represent. 2022, 82, 103431. [CrossRef]

Sensors 2022, 22, 4707 44 of 44

148. Park, J.; Han, D.K.; Ko, H. Fusion of heterogeneous adversarial networks for single image dehazing. IEEE Trans. Image Process.2020, 29, 4721–4732. [CrossRef] [PubMed]

149. Fu, M.; Liu, H.; Yu, Y.; Chen, J.; Wang, K. DW-GAN: A Discrete Wavelet Transform GAN for NonHomogeneous Dehazing. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021;pp. 203–212.

150. Wang, P.; Zhu, H.; Huang, H.; Zhang, H.; Wang, N. Tms-gan: A twofold multi-scale generative adversarial network for singleimage dehazing. IEEE Trans. Circuits Syst. Video Technol. 2021, 2, 2760–2772. [CrossRef]

151. Zhang, Y.; Dong, Y. Single Image Dehazing via Reinforcement Learning. In Proceedings of the 2020 IEEE International Conferenceon Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 6–8 November 2020; Volume 1,pp. 123–126.

152. Guo, T.; Monga, V. Reinforced depth-aware deep learning for single image dehazing. In Proceedings of the ICASSP 2020—2020 IEEEInternational Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 8891–8895.

153. Hong, M.; Xie, Y.; Li, C.; Qu, Y. Distilling image dehazing with heterogeneous task imitation. In Proceedings of the IEEE/CVFConference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3462–3471.

154. Shao, Y.; Li, L.; Ren, W.; Gao, C.; Sang, N. Domain adaptation for image dehazing. In Proceedings of the IEEE/CVF Conferenceon Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2808–2817.

155. Chen, Z.; Wang, Y.; Yang, Y.; Liu, D. PSD: Principled synthetic-to-real dehazing guided by physical priors. In Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7180–7189.

156. Yu, Y.; Liu, H.; Fu, M.; Chen, J.; Wang, X.; Wang, K. A two-branch neural network for non-homogeneous dehazing via ensemblelearning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25June 2021; pp. 193–202.

157. Golts, A.; Freedman, D.; Elad, M. Unsupervised single image dehazing using dark channel prior loss. IEEE Trans. Image Process.2019, 29, 2692–2701. [CrossRef]

158. Li, L.; Dong, Y.; Ren, W.; Pan, J.; Gao, C.; Sang, N.; Yang, M.H. Semi-supervised image dehazing. IEEE Trans. Image Process. 2019,29, 2766–2779. [CrossRef]

159. Zhao, S.; Zhang, L.; Shen, Y.; Zhou, Y. RefineDNet: A weakly supervised refinement framework for single image dehazing. IEEETrans. Image Process. 2021, 30, 3391–3404. [CrossRef]

160. Li, B.; Gou, Y.; Gu, S.; Liu, J.Z.; Zhou, J.T.; Peng, X. You only look yourself: Unsupervised and untrained single image dehazingneural network. Int. J. Comput. Vis. 2021, 129, 1754–1767. [CrossRef]

161. Song, T.; Kim, Y.; Oh, C.; Jang, H.; Ha, N.; Sohn, K. Simultaneous deep stereo matching and dehazing with feature attention. Int.J. Comput. Vis. 2020, 128, 799–817. [CrossRef]

162. Pang, Y.; Nie, J.; Xie, J.; Han, J.; Li, X. BidNet: Binocular image dehazing without explicit disparity estimation. In Proceedings ofthe IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5931–5940.

163. Nie, J.; Pang, Y.; Xie, J.; Pan, J.; Han, J. Stereo refinement dehazing network. IEEE Trans. Circuits Syst. Video Technol. 2021, 32,3334–3345. [CrossRef]

164. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. End-to-end united video dehazing and detection. In Proceedings of the AAAIConference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32.

165. Ren, W.; Zhang, J.; Xu, X.; Ma, L.; Cao, X.; Meng, G.; Liu, W. Deep video dehazing with semantic segmentation. IEEE Trans. ImageProcess. 2018, 28, 1895–1908. [CrossRef] [PubMed]

166. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEETrans. Image Process. 2004, 13, 600–612. [CrossRef] [PubMed]

167. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Wang, Z.; Wang, X.; Jiang, J.; Lin, C.W. Rain-free and residue hand-in-hand: A progressivecoupled network for real-time image deraining. IEEE Trans. Image Process. 2021, 30, 7404–7418. [CrossRef]

168. McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons, Inc.: New York, NY, USA, 1976.169. Luo, Y.; Xu, Y.; Ji, H. Removing rain from a single image via discriminative sparse coding. In Proceedings of the IEEE International

Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3397–3405.170. Liu, J.; Yang, W.; Yang, S.; Guo, Z. Erase or Fill? Deep Joint Recurrent Rain Removal and Reconstruction in Videos. In Proceedings

of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018.171. Yu, L.; Wang, B.; He, J.; Xia, G.S.; Yang, W. Single Image Deraining with Continuous Rain Density Estimation. IEEE Trans.

Multimed. 2021. [CrossRef]172. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International

Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988.173. Chen, X.; Pan, J.; Jiang, K.; Huang, Y.; Kong, C.; Dai, L.; Li, Y. Unpaired Adversarial Learning for Single Image Deraining with

Rain-Space Contrastive Constraints. arXiv 2021, arXiv:2109.02973.174. Ma, L.; Liu, R.; Zhang, X.; Zhong, W.; Fan, X. Video Deraining Via Temporal Aggregation-and-Guidance. In Proceedings of the

2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [CrossRef]175. Zhang, Y.; Zhang, J.; Huang, B.; Fang, Z. Single-image deraining via a recurrent memory unit network. Knowl.-Based Syst. 2021,

218, 106832. [CrossRef]