Next Article in Journal / Special Issue
Mode Recognition of Orbital Angular Momentum Based on Attention Pyramid Convolutional Neural Network
Previous Article in Journal
Total Variation Weighted Low-Rank Constraint for Infrared Dim Small Target Detection
Previous Article in Special Issue
Towards Robust Semantic Segmentation of Land Covers in Foggy Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Empirical Study on Retinex Methods for Low-Light Image Enhancement

1
College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China
2
Computer Science Department, Middlesex University, Hendon, London NW4 4BT, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2022, 14(18), 4608; https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184608
Submission received: 7 August 2022 / Revised: 9 September 2022 / Accepted: 11 September 2022 / Published: 15 September 2022

Abstract

:
A key part of interpreting, visualizing, and monitoring the surface conditions of remote-sensing images is enhancing the quality of low-light images. It aims to produce higher contrast, noise-suppressed, and better quality images from the low-light version. Recently, Retinex theory-based enhancement methods have gained a lot of attention because of their robustness. In this study, Retinex-based low-light enhancement methods are compared to other state-of-the-art low-light enhancement methods to determine their generalization ability and computational costs. Different commonly used test datasets covering different content and lighting conditions are used to compare the robustness of Retinex-based methods and other low-light enhancement techniques. Different evaluation metrics are used to compare the results, and an average ranking system is suggested to rank the enhancement methods.

Graphical Abstract

1. Introduction

Low-light enhancement methodologies try to recover buried details, remove the noise, restore the color details, and increase the dynamic range and contrast of the low-light images. Low light has inescapable effects on remote monitoring equipment and computer vision tasks. Low signal-to-noise ratio (SNR) causes severe noise in low-light imaging and makes it difficult to extract features for interpreting remote-sensing via computer vision tasks, whereas the performance of computer vision tasks entirely depends on accurate feature extraction [1]. Remote-sensing image enhancement has a wide range of applications in object detection [2,3], object tracking [4,5,6,7], video surveillance [8,9], military applications, daily life [10,11,12,13,14], atmospheric sciences [15], driver assistance systems [16], and agriculture. Earth is continuously being monitored by analyzing the images taken by satellites. Analyzing remotely taken images to help in fire detection, flood prediction, and understanding other environmental issues. Low-light enhancement of these images is playing a vital role in understanding these images in a better way. Even the accuracy of other remote sensing algorithms, such as classification and object detection, depends entirely on the image’s quality. In the literature, different methodologies exist for enhancing such degraded low-light images. Retinex theory-based enhancement methods are widely accepted among these enhancement methodologies due to their robustness. The main purpose of this study is to compare the Retinex-based methods with other non-Retinex-based enhancement methods experimentally. For comparison, we have categorized all the enhancement methods into two major groups (i.e., Retinex-based and non-Retinex-based methods). The Retinex group includes classical and deep learning-based Retinex enhancement methods. Meanwhile, the non Retinex group includes histogram equalization, gamma correction, fusion, and deep learning-based enhancement methods.
According to Retinex theory [17], an image can be decomposed into reflectance and illumination component. The reflectance component is considered an intrinsic component of the image and remains consistent under any lighting condition, whereas the illumination component represents the different lighting conditions. Later on, different Retinex theory based methods, such as single-scale retinex (SSR), [18] multiscale retinex with color restoration (MSRCR) [19], simultaneous reflectance and illumination estimation (SRIE) [20], and low-light illumination map estimation (LIME) [21] were developed for low-light enhancement. These methods produce promising results but may require fine-tuning of parameters and may fail to decompose the image correctly into reflectance and illumination parts. Wei et al. is the first one to introduce a real low/normal-light LOw-Light (LOL) dataset and Retinex theory-based deep network (Retinex-Net) in [22]. Retinex-Net comprises Decom-Net for decomposing the image into reflectance and illumination parts and an Enhance-Net for illumination adjustment. Later on, different Retinex theory-based deep learning methods were developed for low-light image enhancement algorithm [22,23,24,25].
Non-Retinex method such as histogram equalization is one of the simplest methods for enhancing low-light images. It flattens the distribution of pixel values throughout the image to improve contrast. In addition, using entire histogram information may over brighten some regions of the image, deteriorate its visual quality and introduce some artifacts in it. Different histogram-based methods such as local histogram equalization [26] and dynamic histogram equalization [27] were introduced to address these issues. However, these methods require higher computation power, the quality of the output depends on the fine-tuning of parameters, and in case of severe noise, it may produce artifacts. On the other hand, gamma correction based methods [28,29,30] apply the pixel-wise nonlinear operation to enhance the image. The main drawback of these methods is that each pixel is considered an individual entity, and their relationship with neighbor pixels is entirely ignored. Due to this, the output may be inconsistent with real scenes. Lore et al. [31] is the first to propose a learning-based enhancement network named LLNet using a synthetic dataset. Later on, different low-light training datasets (e.g., LOL [22], SID [32], SICE [33], VV (https://sites.google.com/site/vonikakis/datasets (accessed on 7 July 2021)), TM-DIED (https://sites.google.com/site/vonikakis/datasets (accessed on 7 July 2021)), and LLVIP [34]) were developed in order to assist the development of learning-based architectures [35,36,37,38].
Wang et al. [39] present a technical evaluation of different methods for low-light imaging. Most of the methods reviewed are classical, and comparing evaluations on five images is quite unfair. Later on, Qi et al., in [40], provide an overview of low-light enhancement techniques, whereas the quantitative analysis of a few methods only on the synthetic dataset (without noise) is provided. Noise is the most critical part of low-light enhancement and a single synthetic low-light dataset cannot compare performance. In [41], Li et al. propose a low-light image and video dataset to examine the generalization of existing deep learning-based image and video enhancement methods. In sum, low-light enhancement has a wide range of applications and is one of the most important image processing fields. To the best of our knowledge, no such study paper is present in the literature mentioned above that extensively provides the technical evaluation of low-light enhancement methods.
The main purpose of this research is to fairly compare the performance of Retinex-based enhancement methods with non-Retinex enhancement methods on a wide range of test datasets covering different contents and lighting conditions. For a fair comparison, the experimental evaluation criteria are defined first, and then all the methods are compared based on the criteria. In addition, an average ranking system is suggested to rank the enhancement methods based on their robustness. Computational complexity analysis of methods is also carried out on four different image sizes for real-time application. This experimental comparison and suggested ranking system of enhancement methods help the research community to understand their shortcomings and to design more robust models in the future.
The main contribution of this research can be summarized as follows:
  • A comprehensive literature review is presented for Retinex-based and non-Retinex methods.
  • A detailed experimental analysis is provided for a variety of Retinex-based and non-Retinex methods on a variety of publicly available test datasets using well-known image quality assessment metrics. Experimental results provide a holistic view of this field and provide readers with an understanding of the advantages and disadvantages of existing methodologies. In addition, the inconsistency of commonly used evaluation metrics is pointed out.
  • An analysis of the computational effectiveness of enhancement methods is also conducted on images of different sizes. As a result of this computation cost, we can determine which enhancement methods are more suitable for real-time applications.
  • Publicly available low-light test datasets were ranked based on experimental analysis. In developing more robust enhancement methods, the reader will benefit from this ranking of benchmarking test datasets.
The rest of the paper is organized as follows. Section 2 presents the relevant background knowledge of non-Retinex-based, and Retinex-based classical and advanced low-light enhancement methodologies. Section 3 presents the objectives of overall paper. In Section 4, experimental setup is defined, a detailed discussion of the qualitative, quantitative, and computational analysis of the classical and advanced low-light enhancement methodologies are provided. Section 5, reports the challenges and the future trends. Finally, the conclusion is drawn in Section 6.

2. Fundamentals

A thorough review of the literature related to Retinex-based and non-Retinex-based classical and advanced learning-based low-light enhancement methods is presented in this section. The following subsections contain literature on each of the categories mentioned above.

2.1. Retinex-Based Methods

Classical Retinex-based methods: The Retinex theory was developed by Land after he studied the human retina-and-cortex system in detail [17]. According to the presented theory, an image can be decomposed into two parts: reflectance and illumination. Reflectance is considered an intrinsic property and remains the same regardless of the lighting condition. Illumination is determined by the intensity of light. The following representation can be used to explain it:
S ( x , y ) = R ( x , y ) I ( x , y ) ,
where S, R and I represent the source image, reflectance and illumination, respectively and the operator ∘ denotes the element-wise multiplication between R and I. As time progressed, different implementations of Retinex theory were proposed in the literature. Path-based implementation of the Reinex [42,43,44,45,46,47] uses different geometry to calculate the relative brightness of adjacent pixels to obtain the reflection component. Marini and Rizzi proposed a biologically inspired implementation of Retinex for dynamic adjustment and color constancy in their article [45]. In [44], the authors examine the different path-wise approaches in detail and propose a mathematical formula to analyze these approaches. It is worth noting that the number of paths has a significant impact on the accuracy of the results. As a result, these path-wise implementations of Retinex theory suffer from a high degree of dependency on the path and sampling noise, as well as a high cost of computation when fine-tuning parameters.
The new method, random spray Retinex (RSR), was developed by Provenzi after replacing the paths with 2-D pixels sprays in [48]. When paths are replaced with 2-D random points distributed across the image, it is possible to determine the locality of color perception. Even though this approach is faster, the spray radius, radial density function, number of sprays, and pixels per spray must be adjusted. Jobson et al., in [18], used a single-scale Retinex (SSR) to implement Retinex for color constancy, and lightness and color rendition of grayscale images. It is not possible for the SSR to provide both dynamic range compression (small scale) and tonal rendition (large scale) simultaneously. However, it can only perform one of these tasks. Later, the authors of SSR extended their idea to multiscale retinex with color restoration (MSRCR) [19]. As a result of MSRCR, dynamic range compression, color consistency, and tonal rendition can be provided. SSR and MSRCR both improve lighting and scene restoration for digital images, but halo artifacts are visible near edges [49]. The majority of Retinex-based algorithms ignore the illumination component and only extract the reflection component as an enhanced result, but this results in unnaturalness. Enhancing an image is not just about enhancing details but also about maintaining its natural appearance. To solve the unnatural appearance, Wang et al. [50] make three contributions: (1) lightness-order-error metrics are proposed to measure objective quality, (2) bright-pass filters decompose images into reflectance and illumination, and (3) bi-log transformations to map illumination while maintaining the balance between details and naturalness.
Zosso et al. reviewed Retinex-based methods and classified them into five broad categories in [51]. Additionally, a two-step non-local unifying framework is proposed to enhance the results and address the Retinex problem. In the first step, a quasi gradient filter is obtained which satisfies gradient-sparsity and gradient-fidelity prior constraints. As a second step, additional constraints are applied to the calculated quasi-gradient filter in order to make it fit the reflectance data. Guo et al. devised a method named low-light illumination map estimation (LIME) [21] to estimate the illumination of each pixel first; then, apply a structure to that illumination map and use it as the final illumination map. A variational based framework (VF) was introduced for Retinex for the first time by Kimmel et al. [52]. In accordance with previous methods, the objection function is based on the assumption that the illumination field is smooth. On the other hand, this model lacks information regarding reflectance. Later on, different variational approaches to Retinex theory are presented [53,54,55]. In [56], a total variational model (TVM) for Retinex is proposed, assuming spatial smoothness of illumination and piecewise continuity of reflection. In order to minimize TVM, a split Bregman iteration is used. VF and TVM differ primarily in that TVM also takes into account reflection.
Fu et al. proposed a linear domain probabilistic method for simultaneous illumination and reflectance estimation (PM-SIRE) [49]. By using an alternating direction multiplier method, maximum a posteriori (MAP) is employed to estimate illumination and reflectance effectively. Later, Fu et al. presented a weighted variational model for simultaneous illumination and reflectance estimation (WV-SIRE) [20]. A WV-SIRE model is capable of preserving more details about the estimated reflectance as well as suppressing noise more effectively than a log-transformed model. The PM-SIRE and WV-SIRE both assume that illumination changes smoothly over time, which may lead to incorrect illumination estimation. Based on the luminous source, different surfaces are illuminated in different directions.
A fusion-based method for enhancing weakly illuminated images is proposed in [57]. This fusion method decomposes a weakly illuminated image into a reflectance map and an illumination map. By using sigmoid and adaptive histogram equalization functions, the illumination map is further decomposed into luminance-improved and contrast-enhanced versions, and two weights are designed for each. Finally, an enhanced image is obtained by combining the luminance-improved and contrast-improved versions with their corresponding weights in a multi-scale manner. For the purpose of preserving intrinsic and extrinsic priors, Cai et al. proposed a joint intrinsic-extrinsic prior (JieP) model [58]. In JieP, shape prior is used to preserve structure information, texture prior is used to estimate illumination with fine details, and illumination prior is used to capture luminous information. Ying et al. [59] simulate the camera response model (CRM) by investigating the relationship between two different exposure images and use the illumination estimation to estimate the exposure ratio map. Later, the CRM and exposure ratio map are used to produce the enhanced image. According to the CRM algorithm, some dark parts of the body, such as the hair, are misinterpreted as dark backgrounds, and they are over-enhanced as well.
Advanced Retinex-based methods: The robustness of Retinex theory makes it applicable to deep learning methods as well. Wei et al. were the first to combine the idea of Retinex theory with deep learning by proposing the Retinex-Net network. Retinex-Net consists of a Decom-Net for decomposing the image into reflectance and illumination parts and an Enhance-Net for adjusting illumination. Furthermore, they introduce a real low/normal-light Low-Light (LOL) dataset [22]. As a further development of the Retinex theory, Zhang et al. proposed the kindling the darkness (KinD) network in [36]. There are three components of KinD: layer decomposition, reflectance restoration, and illumination adjustment. As a result of layer decomposition, the input image is divided into reflectance and illumination elements, the reflectance part is improved by reflecting restoration and the illumination part is smoothed piece-by-piece by illumination adjustment. By combining the outputs of the reflectance and illumination modules, the final result is achieved. Artifacts, overexposure, and uneven lighting are common problems with KinD outputs. For mitigating these effects, Zhange et al. proposed an improved version of KinD in [60]. This improved version of KinD implements a multi-scale illumination attention module, known as KinD++. KinD++ has improved the quality of output images, but it has a lower computational efficiency than KinD. In [61], a Retinex-based real-low to real-normal network (R2RNet) was proposed. R2RNet consists of a decomposition network, a denoise network, and a relight network, each of which is trained separately using decomposition loss, denoise loss, and relight loss, respectively. As a result of decomposition, illumination and reflectance maps are produced. The denoise-net uses the illumination map as a constraint to reduce the noise in the reflectance map, and the relight-net utilizes the denoised illumination map and reflectance map in order to produce an enhanced output. It is noteworthy that three separately trained networks are utilized to solve the low-light enhancement problem, which is not an optimal strategy. Decomposing an image into illumination and reflectance is a computationally inefficient process. Retinex-based transfer functions were introduced by Lu and Zhange in [23] to solve this decomposition problem. As opposed to decomposing the image, the network learn the transfer function to obtain the enhanced image. Liu et al. [62] introduces reference free Retinex-inspired unrolling with architecture search (RUAS) to reduce computational burden and construct lightweight yet effective enhancement. First, RAUS exploits the intrinsic underexposed structure of low-light images; then, it unrolls the optimization process to establish a holistic propagation model. Wang et al. [63] presents paired seeing dynamic scene in the dark (SDSD) datasets. A self-supervised end-to-end framework based on Retinex is also proposed in order to simultaneously reduce noise and enhance illumination. This framework consists of modules for progressive alignment, self-supervised noise estimation, and illumination map prediction. With progressive alignment, temporal information is utilized to produce blur-free frames, self-supervised noise estimation estimates noise from aligned feature maps of the progressive module, and illumination estimation estimates illumination maps consistent with frame content.
Retinex theory is also used in semi-supervised and zero-shot learning-based techniques for enhancing low light visibility. In Zhang et al. [24], a self-supervised maximum entropy Retinex (ME-Retinex) model is presented. In the ME-Retinex model, a network for enhancing image contrast is coupled with a network for re-enhancing and denoising. Zhao et al. [64] proposed a zero-reference framework named RetinexDIP that draws inspiration from the concept of a deep image prior (DIP). The Retinex decomposition is carried out in a generative manner in RetinexDIP. From random noises as input, RetinexDIP generates both reflectance and illumination maps simultaneously, and enhances the illumination map resulting from this process. The proposed model generalizes well to various scenes, but producing an illumination map requires hundreds of iterations. This iterative learning approach consumes a lot of time to produce optimized results. The robust retinex decomposition network (RRDNet) is a three-branch zero-shot network that is proposed in RRDNet [25] to decompose low-light input images into illumination, reflectance, and noise. RRDNet weights are updated by a zero-shot scheme using a novel non-reference loss function. In the proposed loss function, there are three components: the first part reconstructs the image, the second part enhances the texture of the dark region, and the third part suppresses noise in the dark regions. Qu et al., in [65], segmented an image into sub-images, applied deep reinforcement learning to learn the local exposure for each sub-image and finally adversarial learning is applied to approximate the global aesthetic function. It is also proposed to learn discriminators asynchronously and reuse them as value functions.

2.2. Non-Retinex Methods

Histogram equalization (HE) [66] is one of the earlier methods used for enhancing the dynamic range of low-light images. It is a well-known method due to its simplicity. When the entire image histogram is balanced, the visual quality of the image is deteriorated, false contours are introduced, and annoying artifacts are introduced into the image [67]. As a result, some uniform regions become saturated with very bright and very dark intensities [68]. Gamma correction [69] is a non-linear classical technique that is used for image enhancement. It increases the dark portion of the image while suppressing the bright portion. During gamma correction, each pixel is treated as an individual. It is possible that some regions of the image will be under- or over-enhanced due to a single transformation function used for each pixel.
In later years, deep learning has been applied to my field of study. Lore et al. [31] were the first one to use a stacked sparse based autoencoder approach called LLNet for joint enhancement and noise reduction. There is evidence that deeper networks perform better than non-deeper networks; however, deeper networks suffer from gradient vanishing problems. To use a deeper network and solve the gradient vanishing problem, Tao et al. in LLCNN [70] proposed a special module to utilize multiscale feature maps for low-light enhancement. A multi-branch low-light enhancement network (MBLLEN) is designed by Lv et al. in [71] to extract features of different levels, enhance these multi-level features, and fuse them in order to produce an enhanced image. Additionally, Lv et al. also propose a novel loss function that takes into account the structure information, context information, and regional differences of the image. Wang et al., in [72], propose the global illumination-aware and detail-preserving network (GLADNet). In the first step, GLADNet uses an encoder-decoder network to estimate the global illumination and then reconstructs the details lost during the rescaling process. The major disadvantage of LLNet, LLCN, MBBLEN and GLADNet is that they were trained on synthetically darkened and noise-added datasets. Chen et al. [32] used a Unet based pipeline for enhancing and denoising extremely low-light images using the RAW training see-in-the-dark (SID) dataset. This Unet-based pipeline is designed specifically for images in RAW format. Practically, the most common image format is sRGB. The majority of previous methods have used pixel-wise reconstruction losses and failed to provide effective regularization of the local structure of the image, which in turn undermines the network’s performance. The pixel-to-pixel deterministic mapping results in improperly exposed regions, introduces artifacts, and fails to describe the visual distance between the reference and the enhanced image. A flow-based low-light enhancement method (LLFlow) has been proposed by Wang et al. [38] to address this pixel-to-pixel mapping issue. It is possible to map multi-modal image manifolds into latent distributions using the normalizing flow. Effectively enhanced manifolds can be constructed using the latent distribution.
Getting low-light and normal-light images paired can be difficult, expensive, or impractical. An unpaired low-light enhancement method called EnlightenGAN is proposed by Jiang et al. [73] to eliminate the need for paired training datasets. A global-local discriminator structure and an easy-to-use attention U-net generator are proposed in EnlightenGAN. By designing the attention U-net only to enhance the dark regions more, the image is neither overexposed nor underexposed. A dual global-local discriminator strategy contributes to the balance between local and global enhancement of low-light images. Xiong et al. [74] considered low-light enhancements as two subtasks: illumination enhancement and noise reduction. A two-stage framework referred to as decoupled networks is proposed for handling each task. In decoupled networks, there are two encoder-decoder architectures, the first architecture enhances illumination, and the second architecture suppresses noise by taking the original input along with the enhanced output from stage one. To facilitate unsupervised learning, an adaptive content loss and pseudo triples are proposed. Xia et al. [75] used two images of the scene taken in quick succession (with and without a flash) to generate a noise-free and accurate display of ambient colors. Using a neural network, an image taken without flash is analyzed for color and mood, while an image taken with a flash is analyzed for surface texture and details. One of the major disadvantages of this method is that paired images with and without flash are not generally available.
The camera sensors on mobile phones perform poorly in low-light conditions. An improved face verification method using a semisupervised decomposition and reconstruction network is proposed in [76] to improve accuracy for low-light images of faces. Yang et al. [77] proposes a deep semi-supervised recursive band network (DRBN) to address the decreased visibility, intensive noise, and biased color of low-light images. DRBN learns in two stages, the first stage involves learning the linear band representation by comparing low- and normal-light images, and the second stage involves recomposing the linear band representation from the first stage to fit the visual properties of high-quality images through adversarial learning. Further improvement of the DRBN is impeded by the separation of supervised and unsupervised modules. Qiao et al. [78] further improved DRBN performance by introducing a joint training based semi-supervised algorithm. Wu et al. [79] proposed the lightweight two stream method to overcome the limitations of the training data due to sample bias and the hurdle of the large number of parameters in real-time deployment. Additionally, a self-supervised loss function is proposed to overcome the sample bias of the training data.
Guo et al. [80] proposes zero-reference deep curve estimation (Zero-DCE) rather than performing image-to-image mapping. In order to preserve the contrast of the neighboring pixels, Zero-DCE creates high-order curves from low-light images and then adjusts low-light images pixel-by-pixel using these high-order curves. It is superior to existing GAN-based methods since it does not require paired or unpaired data for its training. Enhanced images are produced with four non-reference loss functions: spatial consistency loss, exposure control loss, color constancy loss, and illumination smoothness loss. The re-design and reformulation of the network structure were subsequently carried out by Li et al., who introduced Zero-DCE++, which is an accelerated and lighter version of Zero-DCE.

3. Objectives of Experimental Study

This research study aims to address the following questions:
1.
It has been noted that although there have been a large number of algorithms developed for low-light enhancement, Retinex theory-based models are gaining more attention due to their robustness. Retinex theory is even used in deep learning-based models. Specifically, this paper attempts to compare the performance of Retinex theory-based classical and deep learning low-light enhancement models with other state-of-the-art models.
2.
Several low-light enhancement methods perform well on some test datasets but fail in real-world scenarios. An extensive range of real-world images should be used to test the robustness of the low-light enhancement models. As a means of assessing the robustness of enhancement methods in real-world scenarios, various test datasets spanning a wide range of lighting conditions and contents need to be selected, and the performance of Retinue-based models needs to be compared with that of other enhancement techniques on these test datasets.
3.
The trend of real-time cellphone night photography is increasing day by day. Therefore, analyzing the computational costs associated with low-light enhancement methods is necessary. Comparison of not only the parameters of these methods but also the processing time for the images of four different sizes (i.e., 400 × 600 × 3 , 640 × 960 × 3 , 2304 × 1728 × 3 and 2848 × 4256 × 3 ) is required. A computational analysis of different sizes of images will enable the researchers to determine whether the computational cost increases linearly or exponentially as the image size increases.
4.
The quality of low-light enhancement methods needs to be evaluated using a variety of image quality assessment (IQA) methods. Every metric aims to identify the particular quality of the predicted image. The LOE measures the naturalness of the image, whereas the information entropy measures the information contained in the image. What is the most effective method of comparing the robustness of low-light enhancement methods when comparing results based on these evaluation metrics?

4. Quantitative and Qualitative Analysis

This subsections of this section present the experimental setup for farily comparing the methods, qualitative, quantitative comparison, and computational cost analysis of enhancement methods. In addition, it also discusses the evaluation metrics and test datasets.

4.1. Experimental Criteria for Enhancement Methods Comparison

To conduct a fair comparison to analyze the enhancement methods generalization, we have selected the nine different publicly available test datasets widely used in the literature for comparing the performance of enhancement methods [64,73,81]. The selected datasets include LIME [21], LOL [22], DICM [82], VV (https://sites.google.com/site/vonikakis/datasets (accessed on 7 July 2021)), MEF [83], NPE [50], LSRW [61], SLL [84] and ExDark [85]. The main purpose of selecting these different nine test datasets is to cover diversified scenes, camera devices, lighting conditions (i.e., weak lighting, under exposure, twilight, dark), and contents. In summary, each test dataset covers a different aspect of low-lighting, scene or content. Therefore, these test datasets are useful to compare the performance of enhancement methods from different aspects.
The four most commonly used no-reference metrics for the quantitative evaluation of low-light enhancement methods are used. These metrics include entropy [86], BRISQUE [87], NIQE [88], and LOE [50]. The entropy measures the information content of an image. a higher value of entropy indicates richer details and a higher contrast level of an image. Blind/referenceless image spatial quality evaluator (BRISQUE) is another commonly used model to quantify the quality of low-light enhancement methods. It does not compute the distortion specific feature, but instead it uses the scene statistics to quantify the loss of naturalness in an image due to the presence of distortion. BRISQIE uses a space vector machine (SVM) regressor to predict the quality of the image. Natural image quality evaluator (NIQE) quantifies the quality of the distorted image by measuring the distance of natural scene statistic (NSS) feature model and the multivariate Gaussian (MVG) feature model of distorted image. Lightness order error (LOE) is designed to measure the order of lightness. The order of lightness represents the direction of the light source and helps to quantify the naturalness preservation. LOE can be defined as follows:
L O E = 1 m n x = 1 m y = 1 n U ( Q ( i , j ) , Q ( x , y ) ) U Q r ( i , j ) , Q r ( x , y ) ,
where U ( x , y ) is a unit step function. It returns 1 if x > y and returns 0 otherwise. m, n represents height and width of the image, respectively. Moreover, Q ( i , j ) and Q r ( i , j ) are maximum values among the three color channels at location ( i , j ) for the original image and enhanced image, respectively.
In this study, the performance of 17 Retinex-based methods and 17 non-Retinex will be compared. We consider the publicly available codes and recommended settings of these methods to have a fair comparison. The higher value of entropy indicates better quality and for the other three methods (i.e., LOE, NIQE, and BRISQUE) lower values of entropy indicate the better image quality. To show a better understanding the comparison, average ranking has been suggested to enhancement methods based on these IQA methods. For example, the enhancement methods that got the highest average score of entropy on all test datasets are given rank 1 and vice versa. Similarly, the enhancement methods show the lowest average score according to LOE or NIQE or BRISQUE are assigned rank 1 and the highest average score is assigned the highest rank. Rank 1 indicates the best performance and the rank with higher value indicates the worst performance.
In addition, we compare the computational complexity of classical methods on images of four different sizes. The classical codes computational complexity is computed on CPU, whereas those of deep learning-based methods on NVIDIA Titan Xp GPU.

4.2. Qualitative Evaluation of Enhancement Methods

In this section, we provide a detailed description of the qualitative evaluation of enhancement methods. The comparative visual results of the top ten classical and advanced methods on six publicly available test datasets are shown in Figure 1 and Figure 2, respectively. These figures’ first to sixth columns indicate the enhancement results of different methods on LIME, LARW, DICM, ExDark, LOL, and SLL datasets, respectively. For simplicity, deep learning and classical methods are discussed one by one. It is encouraged to zoom in to compare the results.
Zero-shot learning-based methods (i.e., ZeroDCE and RetinexDIP) produce darker and noisy images compared to other methods. The results of GLADNet, TBEFN, and LLFlow are more realistic, sharper, less noisy, and have accurate color rendition. The output images of MBLLEN are over-smoothed and darker but less dark than ZeroDCE. GLADNet, TBEFN, LLFlow, MBLLEN, and KinD are trained on paired data. The supervised learning-based models achieved the appropriate restoration of color and textures, noise suppression, and better generalization. However, no method has produced good results on all the datasets. For example, GLADNet results on DICM are too noisy and produce artifacts on the ExDark image. Similarly, strange artifacts on DICM images are produced by TBEFN. LLFLow produces greenish color around the edges of LSRM image. As it can be seen, StableLLVE has a lighter washed-out effect and smoothed edges on all the results. KinD results look realistic, but some parts of the image look too dark, such as the background chairs in the LOL image. SS-Net produces a good result on the VV test image but produces poor results on DICM and ExDark. Moreover, the strange pattern, missing color information, and other details can be observed easily on the ExDark image. The results of Retinex-based methods (i.e., TBEFN, KinD, SS-Net, RetinexDIP) look more natural and real.
The classical methods (i.e., CVC, DHE, BIMEF, IAGC, and AGCWD) shown in the Figure 2 belong to the non-Retinex category, and PM-SIRE, WV-SRIE, JieP, EFF, and NPE belongs to Retinex theory. If we closely observe their visual results, one thing that is common among majority of these methods is noise. Except for BIMEF and EFF, most results can easily observe noise. The average brightness of BIMEF is too low and does not enhance the overall image. On the other hand, EFF produces higher brightness results, but the image’s details are not too sharp. CVC and IAGC do not accurately render the color information, making their results look black and white. Although some classical methods’ results quality is good, their results are still darker than deep learning-based methods. The results produced by BIMEF, IAGE, and CVC are darker as compared to other classical methods. Over-enhancement, severe noise and loss of color information can be seen in the results of DHE. The results of CVC are not only darker but also lost color information. AGCWD produces low contrast and less bright images, and some parts of the image are too dark (for example background buildings in the LIME test image and the background wall in the LOL test image). Gamma correction-based methods (i.e., AGCWD and IAGC) enhance some parts of the image while darker parts become darker. Strange artifacts around the fire can be easily seen in the IAGC result on the ExDark image.
The results of Retinex-based methods (NPE and WV-SIRE) enhance the image’s brightness, contrast, and sharpness, but fail to suppress the noise. The major issue with the majority of traditional methods is noise suppression. Histogram-based methods work to balance the histogram of the image to increase the brightness and contrast, but there is no such mechanics to remove the noise. Meanwhile, gamma correction-based methods treat each pixel individually and fail to exploit their relationship with neighbor pixels, which results in different artifacts and noise. In contrast, Retinex theory-based methods create different algorithms for successfully decomposing low-light images into reflectance and illumination components. In the case of severe noise, decomposing the image becomes difficult. The noise is not considered a major factor in any of these approaches. Therefore, noise dominates the visual results of these methods. When Figure 1 and Figure 2 are compared, it is evident that deep learning-based methods produce brighter, sharper, cleaner, and higher contrast results. There is still some noise in some results, but compared to traditional methods, it is very low. Contrary to this, traditional visual results have many shortcomings. For example, some results have a lower average brightness, a lesser contrast level, a lesser degree of sharpness, failure to remove noise, and serious color shifts. Some of them enhance the image and the noise associated with it.

4.3. Quantitative Comparison of Enhancement Methods

Four non-reference evaluation metrics were used for the quantitative comparison. There are two reasons for using no-reference-based IQA metrics: (1) the majority of widely used test datasets are no-reference, and (2) unsupervised methods are emerging. Metrics adopted for evaluation include NIQE [88], BRISQUE [87], LOE [50], and Entropy. Low NIQE, BRISQIE, and LOE values indicate better image quality. In contrast, higher values of entropy indicate richer information. Table 1, Table 2, Table 3 and Table 4 provide quantitative results for these metrics. Red indicates the best scores obtained on each dataset, while blue and green indicate the second and third best scores. The LOE indicates that non-Retinex methods perform better, whereas the other three metrics show that performance is uniform across both categories (i.e., Retinex and non-Retinex). Each method is evaluated by four metrics. There is no winner on all four metrics. To determine which method generalizes well, the enhancement methods score on all test data is averaged. The last column of the aforementioned tables represents the average score of enhancement methods on all test datasets. Based on averaged score, ranking number is assigned to each method and we summarize these rankings in Figure 3. Ranking 1 goes to the method with the best average score, and ranking 31 to the method with the worst average score. Different metrics rank enhancement methods differently. For instance, AGCWD ranked first according to LOE metric, whereas the same method is ranked as fifth, eighteenth, and twenty-ninth according to BRISQIE, NIQE, and entropy, respectively. Instead of analyzing the enhancement methods based on different metrics, we have taken the average of the ranking assigned based on the mentioned metrics and discussed the results of this average ranking.
The red line in Figure 3 represents the average ranking achieved by enhancement methods on all test datasets. The average ranking puts GLADNet, TBEFN, and LLFlow in first, second, and third, respectively. GLADNet generalizes well despite being trained on 5000 synthetic images using L 1 loss. Retinex-based methods TBEFN, WV-SIRE, JieP, and KinD also generalized well and received the 2nd, 4th, 5th, and 6th rankings, respectively. TBEFN [23] is trained on a mixture of 14,531 patches collected from SICE [33] and LOL [22] datasets using SSIM, total variation, and VGG loss. KinD is based on Retinex theory and trained on LOL. A self-supervised network and a zero-shot-based Retinex method (i.e., self-supervised network and RetinexDIP) ranked 12th and 21st, respectively. MBLLEN is also a supervised learning-based networks and ranked 10th. Meanwhile, MBLLEM is a multi-branch fusion network trained on the PASCAL VOC dataset [101]. Zero-shot learning-based methods such as Zero-DCE got 19th. Among all deep learning-based methods, the top six methods are supervised learning-based methods. It is also worth noticing that among the top ten methods, five are Retinex-based methods and 5 are non-Retinex methods.

4.4. Computational Complexity Analysis of Enhancement Methods

The computational complexity analysis of classical methods and deep learning-based methods is presented in Table 5 and Table 6, respectively. The analysis is conducted on four different test datasets (i.e., LOL, LSRW, VV, and SID). Table 5 and Table 6 report the average time taken and the resolution of a single image for each dataset. For each of these tables, red, blue, and green colors are used to indicate the best, second best, and third best performance, respectively. Results shown in Table 5 have been obtained using a CPU, while results shown in Table 6 have been obtained using an NVIDIA Titan Xp GPU. HE has the shortest runtime of all classical methods. HE just takes around 20.3ms to process an image of resolution 2848 × 4256 × 3 . The majority of HE-based methods, such as BPDHE, WAHE, LDR, CVC, and BiHE, are time-efficient, except for DHE. DHE continuously divides an image into several sub-histogram units in order to avoid leaving a dominant portion in newly created sub-histograms. Due to the continual dividing process, this method is the slowest of all the HE-based methods mentioned. Gamma correction-based methods also have good computational efficiency. IAGC takes relatively longer than other methods because it truncates an image’s cumulative distribution function (CDF) and adaptively corrects each truncated CDF.
Furthermore, Retinex-based methods are more computationally expensive than HE and gamma correction-based methods. NPE, PM-SIRE, and WV-SRIE are among the Retinex-based methods that experience significant increases in computation costs with increasing image size. These methods are computationally inefficient due to their iterative approach to finding the optimal solution and use of Gaussian filtering operations. The efficiency of deep learning-based methods depends on the number of parameters that are used. Zero-DCE is the fastest deep learning-based method due to its simplest network architecture and fewer parameters. The majority of deep learning-based methods’ average runtime is between 1.7 ms and 2.57 s. RRDNet iteratively minimizes the error to produce the final enhanced output. The number of iterations varies for different inputs. The iteratively solving the problem makes it the slowest among all the networks. RetinexDIP is another zero-shot learning-based method and performs 300 iterations on each input to produce the final output. The iteratively solving problem makes RetinexDIP and RRDNet the slowest methods. A scatter plot of methods’ performance versus time taken on CPU and GPU is shown in Figure 4 and Figure 5, respectively. We consider CPU methods with less than 1s processing time and GPU methods with less than 0.5s. Methods closer to the origin have a lower computational cost and better performance.

4.5. Difficulty Analysis of Test Datasets

Results of enhancement methods have also been used to rank the nine test datasets. The last row of Table 1, Table 2, Table 3 and Table 4 shows the average score of different enhancement methods on the test datasets based on LOE, NIQE, Entropy, and BRISQUE, respectively. Figure 6 shows the difficulty rank for each dataset across IQA methods. A red line shows the average of all rankings in Figure 6. As determined by the average ranking score, VV is the easiest test dataset, while SLL is the most challenging. SLL is the synthetic dataset with severe noise added. There is too much noise to remove and produce better results. Meanwhile, VV has a lower noise level, making it the easiest dataset. LOL and ExDark are the second and third most difficult datasets. A test dataset’s difficulty level is determined by its noise level. The higher the noise level, the harder it is to recover color details and other information.

4.6. Evaluation IQA methods

To analyze the objective quality of different enhancement methods, we have used LOE, NIQE, Entropy, and BRISQUE as described in Table 1, Table 2, Table 3 and Table 4, respectively. We can easily identify the differences in the evaluations of these metrics if we compare their values among themselves. The best result was produced by the BPDHE enhancement method, according to LOE and NIQE, whereas BRISQUE evaluated MSRCR and Entropy evaluated GLADnet as best methods. Each metric measures a different aspect of the predicted image, which makes their results different. To easily depict and analyze the overall performance of enhancement methods, we have suggested the average rank from 1 to 31 (depending on how many methods are compared) to compare their performance. The best performance is ranked 1, and the worst performance is ranked 31. These rankings can be seen in Figure 3, where the x-axis represents the enhancement methods, and the y-axis represents the ranking. Green, dotted black, blue, and yellow lines in Figure 3 show the rankings of enhancement methods based on LOE, NIQE, BRISQUE, and Entropy metrics, respectively. Moreover, the red line in Figure 3 represents the average of all the rankings mentioned earlier (given based on different metrics). The best method can be chosen based on this average ranking system.

5. Discussion

In this section, we summarize the results obtained and the findings of the overall paper.
i
The enhancement methods are evaluated using four evaluation metrics. No method has emerged as the clear winner on all four metrics (LOE, entropy, NIQE, BRISQUE). This is due to the fact that each evaluation method measures a different aspect of enhancement methods (e.g., LOE measures naturalness, entropy measures information content, and NIQE measures distortion). A suggested average ranking system is found to be the most reliable method of comparing the overall performance of enactment methods.
ii
In the average ranking system, it has been observed that the three most successful enhancement methods (GLADNet, TBEFN, LLFlow) are based on supervised learning. Among the top ten methods, five are based on Retinex. In comparison to classical, advanced self-supervised, and zero-short learning methods, supervised learning is more effective. Denosing is the most challenging part in enhancement. Noise can be observed in the visual results of outperforming methods.
iii
There is no Retinex-based method among the top three fastest methods. As a result of the image decomposition, these methods are more time consuming. As the size of the image increases, the computational time of classical Retinex-based methods increases dramatically. Zero-DCE is the fastest learning-based method, taking approximately 0.017 s to process an image of size 2848 × 4256 × 3 . However, it ranks 20th in terms of performance. GLADNet, on the other hand, is ranked first, but it takes approximately 2.772 s to process an image of the same size.
iv
The average ranking of all enhancement methods is observed in a broader sense. The results indicate that five methods in the top ten are based on Retinex theory (i.e., TBEFN, WV-SRIE, JieP, KinD, and PM-SIRE). The remaining five fall into different categories (i.e., HE, gamma correction, deep learning). When it comes to real-world scenarios, Retinex theory algorithms are more robust. In contrast, decomposing the image into illumination and reflectance makes them more computationally intensive and, therefore, slower. Computational complexity is the bottleneck for their development in real-world scenarios.

6. Conclusions

In this study, we present an experimental comparison of Retinex-based methods with other non-Retinex methods on nine diversified datasets. According to this study, five out of the top 10 methods are based on Retinex. Researchers are aiming to develop methods that can be generalized and produce enhanced, denoised, color rendered results in real time. Based on the comparisons, ZeroDCE is considered to be the fastest method for processing high-resolution images within 17 milliseconds. However, ZeroDCE ranked 19th and its results were darker and noisy. In contrast, Retinex-based methods have a greater degree of robustness and generalization. The decomposition of the image is a time-consuming process and is a bottleneck in the processing time of Retinex-based methods. Based on the overall ranking, supervised learning methods (e.g., GLADNet, TBEFN, LLFLow) perform better than all other methods. Training images for GLADNet and patches for TBEFN are 5000 images and 14,531 patches, respectively. Both GLADNet and TBEFN are able to generalize well due to their large training data, as well as their Unet architecture which makes them more efficient as compared to other heavy network designs. Moreover, this research evaluated the results of enhancement methods on four different metrics and suggested a method for ranking enhancement methods according to their performance. This research study may help the research community develop more robust and lightweight models for real-time photography and video shooting.

Author Contributions

Conceptualization, M.T.R.; methodology, M.T.R.; software, M.T.R.; validation, M.T.R. and G.G.; formal analysis, M.T.R.; investigation, G.G.; resources, D.S.; data curation, H.K.; writing—original draft preparation, G.G.; writing—review and editing, H.K.; visualization, H.K.; supervision, D.S.; project administration, X.C.; funding acquisition, D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Ministry of Science and Technology China (MOST) Major Program on New Generation of Artificial Intelligence 2030 No. 2018AAA0102200. It is also supported by Natural Science Foundation China (NSFC) Major Project No. 61827814 and Shenzhen Science and Technology Innovation Commission (SZSTI) project No. JCYJ20190808153619413.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, J.; Wang, W.; Wang, R.; Gao, W. CSPS: An adaptive pooling method for image classification. IEEE Trans. Multimed. 2016, 18, 1000–1010. [Google Scholar] [CrossRef]
  2. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Chen, Y.; Cai, L.; Ling, H. M2det: A single-shot object detector based on multi-level feature pyramid network. In Proceedings of the AAAI Conference on Artificial Intelligence, Atlanta, GA, USA, 8–12 October 2019; Volume 33, pp. 9259–9266. [Google Scholar]
  3. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  4. Bertinetto, L.; Valmadre, J.; Henriques, J.F.; Vedaldi, A.; Torr, P.H. Fully-convolutional siamese networks for object tracking. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 850–865. [Google Scholar]
  5. He, A.; Luo, C.; Tian, X.; Zeng, W. A twofold siamese network for real-time object tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4834–4843. [Google Scholar]
  6. Luo, W.; Sun, P.; Zhong, F.; Liu, W.; Zhang, T.; Wang, Y. End-to-end active object tracking via reinforcement learning. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 3286–3295. [Google Scholar]
  7. Ristani, E.; Tomasi, C. Features for multi-target multi-camera tracking and re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6036–6046. [Google Scholar]
  8. Saini, M.; Wang, X.; Atrey, P.K.; Kankanhalli, M. Adaptive workload equalization in multi-camera surveillance systems. IEEE Trans. Multimed. 2012, 14, 555–562. [Google Scholar] [CrossRef]
  9. Feng, W.; Ji, D.; Wang, Y.; Chang, S.; Ren, H.; Gan, W. Challenges on large scale surveillance video analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 69–76. [Google Scholar]
  10. Ko, S.; Yu, S.; Kang, W.; Park, C.; Lee, S.; Paik, J. Artifact-free low-light video enhancement using temporal similarity and guide map. IEEE Trans. Ind. Electron. 2017, 64, 6392–6401. [Google Scholar] [CrossRef]
  11. Rasheed, M.T.; Shi, D. LSR: Lightening super-resolution deep network for low-light image enhancement. Neurocomputing 2022, 505, 263–275. [Google Scholar] [CrossRef]
  12. Khan, H.; Wang, X.; Liu, H. Handling missing data through deep convolutional neural network. Inf. Sci. 2022, 595, 278–293. [Google Scholar] [CrossRef]
  13. Khan, H.; Wang, X.; Liu, H. Missing value imputation through shorter interval selection driven by Fuzzy C-Means clustering. Comput. Electr. Eng. 2021, 93, 107230. [Google Scholar] [CrossRef]
  14. Khan, H.; Liu, H.; Liu, C. Missing label imputation through inception-based semi-supervised ensemble learning. Adv. Comput. Intell. 2022, 2, 1–11. [Google Scholar] [CrossRef]
  15. Ellrod, G.P. Advances in the detection and analysis of fog at night using GOES multispectral infrared imagery. Weather. Forecast. 1995, 10, 606–619. [Google Scholar] [CrossRef]
  16. Negru, M.; Nedevschi, S.; Peter, R.I. Exponential contrast restoration in fog conditions for driving assistance. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2257–2268. [Google Scholar] [CrossRef]
  17. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  18. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef] [PubMed]
  19. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed]
  20. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2782–2790. [Google Scholar]
  21. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  22. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  23. Lu, K.; Zhang, L. TBEFN: A two-branch exposure-fusion network for low-light image enhancement. IEEE Trans. Multimed. 2020, 23, 4093–4105. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Di, X.; Zhang, B.; Li, Q.; Yan, S.; Wang, C. Self-supervised Low Light Image Enhancement and Denoising. arXiv 2021, arXiv:2103.00832. [Google Scholar]
  25. Zhu, A.; Zhang, L.; Shen, Y.; Ma, Y.; Zhao, S.; Zhou, Y. Zero-shot restoration of underexposed images via robust retinex decomposition. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
  26. Gonzalez, R.C. Digital Image Processing, 2nd ed.; Addison-Wesley: Boston, MA, USA, 1992. [Google Scholar]
  27. Abdullah-Al-Wadud, M.; Kabir, M.H.; Dewan, M.A.A.; Chae, O. A dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  28. Rahman, S.; Rahman, M.M.; Abdullah-Al-Wadud, M.; Al-Quaderi, G.D.; Shoyaib, M. An adaptive gamma correction for image enhancement. EURASIP J. Image Video Process 2016, 35, 2016. [Google Scholar] [CrossRef]
  29. Huang, S.C.; Cheng, F.C.; Chiu, Y.S. Efficient contrast enhancement using adaptive gamma correction with weighting distribution. IEEE Trans. Image Process. 2012, 22, 1032–1041. [Google Scholar] [CrossRef]
  30. Wang, Z.G.; Liang, Z.H.; Liu, C.L. A real-time image processor with combining dynamic contrast ratio enhancement and inverse gamma correction for PDP. Displays 2009, 30, 133–139. [Google Scholar] [CrossRef]
  31. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  32. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to see in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  33. Cai, J.; Gu, S.; Zhang, L. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef] [PubMed]
  34. Jia, X.; Zhu, C.; Li, M.; Tang, W.; Zhou, W. LLVIP: A Visible-infrared Paired Dataset for Low-light Vision. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 3496–3504. [Google Scholar]
  35. Park, J.; Lee, J.Y.; Yoo, D.; Kweon, I.S. Distort-and-recover: Color enhancement using deep reinforcement learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5928–5936. [Google Scholar]
  36. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  37. Zheng, C.; Shi, D.; Shi, W. Adaptive Unfolding Total Variation Network for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 4439–4448. [Google Scholar]
  38. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.P.; Kot, A.C. Low-Light Image Enhancement with Normalizing Flow. arXiv 2021, arXiv:2109.05923. [Google Scholar] [CrossRef]
  39. Wang, W.; Wu, X.; Yuan, X.; Gao, Z. An experiment-based review of low-light image enhancement methods. IEEE Access 2020, 8, 87884–87917. [Google Scholar] [CrossRef]
  40. Qi, Y.; Yang, Z.; Sun, W.; Lou, M.; Lian, J.; Zhao, W.; Deng, X.; Ma, Y. A Comprehensive Overview of Image Enhancement Techniques. Arch. Comput. Methods Eng. 2021, 29, 583–607. [Google Scholar] [CrossRef]
  41. Li, C.; Guo, C.; Han, L.H.; Jiang, J.; Cheng, M.M.; Gu, J.; Loy, C.C. Low-light image and video enhancement using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 1. [Google Scholar] [CrossRef]
  42. Land, E.H.; McCann, J.J. Lightness and retinex theory. Josa 1971, 61, 1–11. [Google Scholar] [CrossRef]
  43. Land, E.H. Recent advances in retinex theory and some implications for cortical computations: Color vision and the natural image. Proc. Natl. Acad. Sci. USA 1983, 80, 5163. [Google Scholar] [CrossRef]
  44. Provenzi, E.; De Carli, L.; Rizzi, A.; Marini, D. Mathematical definition and analysis of the Retinex algorithm. JOSA A 2005, 22, 2613–2621. [Google Scholar] [CrossRef]
  45. Marini, D.; Rizzi, A. A computational approach to color adaptation effects. Image Vis. Comput. 2000, 18, 1005–1014. [Google Scholar] [CrossRef]
  46. Land, E.H. An alternative technique for the computation of the designator in the retinex theory of color vision. Proc. Natl. Acad. Sci. USA 1986, 83, 3078–3080. [Google Scholar] [CrossRef] [PubMed]
  47. Cooper, T.J.; Baqai, F.A. Analysis and extensions of the Frankle-McCann Retinex algorithm. J. Electron. Imaging 2004, 13, 85–92. [Google Scholar] [CrossRef]
  48. Provenzi, E.; Fierro, M.; Rizzi, A.; De Carli, L.; Gadia, D.; Marini, D. Random spray Retinex: A new Retinex implementation to investigate the local properties of the model. IEEE Trans. Image Process. 2006, 16, 162–171. [Google Scholar] [CrossRef] [PubMed]
  49. Fu, X.; Liao, Y.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A probabilistic method for image enhancement with simultaneous illumination and reflectance estimation. IEEE Trans. Image Process. 2015, 24, 4965–4977. [Google Scholar] [CrossRef] [PubMed]
  50. Wang, S.; Zheng, J.; Hu, H.M.; Li, B. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef] [PubMed]
  51. Zosso, D.; Tran, G.; Osher, S.J. Non-Local Retinex—A Unifying Framework and Beyond. SIAM J. Imaging Sci. 2015, 8, 787–826. [Google Scholar] [CrossRef]
  52. Kimmel, R.; Elad, M.; Shaked, D.; Keshet, R.; Sobel, I. A variational framework for retinex. Int. J. Comput. Vis. 2003, 52, 7–23. [Google Scholar] [CrossRef]
  53. Ma, W.; Osher, S. A TV Bregman iterative model of Retinex theory. Inverse Probl. Imaging 2012, 6, 697. [Google Scholar] [CrossRef]
  54. Ma, W.; Morel, J.M.; Osher, S.; Chien, A. An L 1-based variational model for Retinex theory and its application to medical images. In Proceedings of the CVPR, Colorado Springs, CO, USA, 20–25 June 2011; pp. 153–160. [Google Scholar]
  55. Fu, X.; Zeng, D.; Huang, Y.; Ding, X.; Zhang, X.P. A variational framework for single low light image enhancement using bright channel prior. In Proceedings of the 2013 IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; pp. 1085–1088. [Google Scholar]
  56. Ng, M.K.; Wang, W. A total variation model for Retinex. SIAM J. Imaging Sci. 2011, 4, 345–365. [Google Scholar] [CrossRef]
  57. Fu, X.; Zeng, D.; Huang, Y.; Liao, Y.; Ding, X.; Paisley, J. A fusion-based enhancing method for weakly illuminated images. Signal Process. 2016, 129, 82–96. [Google Scholar] [CrossRef]
  58. Cai, B.; Xu, X.; Guo, K.; Jia, K.; Hu, B.; Tao, D. A joint intrinsic-extrinsic prior model for retinex. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4000–4009. [Google Scholar]
  59. Ying, Z.; Li, G.; Ren, Y.; Wang, R.; Wang, W. A new low-light image enhancement algorithm using camera response model. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 3015–3022. [Google Scholar]
  60. Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond brightening low-light images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
  61. Hai, J.; Xuan, Z.; Yang, R.; Hao, Y.; Zou, F.; Lin, F.; Han, S. R2RNet: Low-light Image Enhancement via Real-low to Real-normal Network. arXiv 2021, arXiv:2106.14501. [Google Scholar]
  62. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  63. Wang, R.; Xu, X.; Fu, C.W.; Lu, J.; Yu, B.; Jia, J. Seeing Dynamic Scene in the Dark: A High-Quality Video Dataset With Mechatronic Alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9700–9709. [Google Scholar]
  64. Zhao, Z.; Xiong, B.; Wang, L.; Ou, Q.; Yu, L.; Kuang, F. RetinexDIP: A Unified Deep Framework for Low-light Image Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1076–1088. [Google Scholar] [CrossRef]
  65. Yu, R.; Liu, W.; Zhang, Y.; Qu, Z.; Zhao, D.; Zhang, B. Deepexposure: Learning to expose photos with asynchronously reinforced adversarial learning. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 2153–2163. [Google Scholar]
  66. Cheng, H.; Shi, X. A simple and effective histogram equalization approach to image enhancement. Digit. Signal Process. 2004, 14, 158–170. [Google Scholar] [CrossRef]
  67. Kim, Y.T. Contrast enhancement using brightness preserving bi-histogram equalization. IEEE Trans. Consum. Electron. 1997, 43, 1–8. [Google Scholar]
  68. Ibrahim, H.; Kong, N.S.P. Brightness preserving dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  69. Guan, X.; Jian, S.; Hongda, P.; Zhiguo, Z.; Haibin, G. An image enhancement method based on gamma correction. In Proceedings of the 2009 Second International Symposium on Computational Intelligence and Design, Changsha, China, 12–14 December 2009; Volume 1, pp. 60–63. [Google Scholar]
  70. Tao, L.; Zhu, C.; Xiang, G.; Li, Y.; Jia, H.; Xie, X. LLCNN: A convolutional neural network for low-light image enhancement. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  71. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using CNNs. In Proceedings of the BMVC, Newcastle, UK, 3–6 September 2018; p. 220. [Google Scholar]
  72. Wang, W.; Wei, C.; Yang, W.; Liu, J. GLADNet: Low-light enhancement network with global awareness. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Jodhpur, India, 15–18 December 2018; pp. 751–755. [Google Scholar]
  73. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  74. Xiong, W.; Liu, D.; Shen, X.; Fang, C.; Luo, J. Unsupervised real-world low-light image enhancement with decoupled networks. arXiv 2020, arXiv:2005.02818. [Google Scholar]
  75. Xia, Z.; Gharbi, M.; Perazzi, F.; Sunkavalli, K.; Chakrabarti, A. Deep Denoising of Flash and No-Flash Pairs for Photography in Low-Light Environments. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2063–2072. [Google Scholar]
  76. Le, H.A.; Kakadiaris, I.A. SeLENet: A semi-supervised low light face enhancement method for mobile face unlock. In Proceedings of the 2019 International Conference on Biometrics (ICB), Crete, Greece, 4–7 June 2019; pp. 1–8. [Google Scholar]
  77. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From fidelity to perceptual quality: A semi-supervised approach for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3063–3072. [Google Scholar]
  78. Qiao, Z.; Xu, W.; Sun, L.; Qiu, S.; Guo, H. Deep Semi-Supervised Learning for Low-Light Image Enhancement. In Proceedings of the 2021 14th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Online. 23–25 October 2021; pp. 1–6. [Google Scholar]
  79. Wu, W.; Wang, W.; Jiang, K.; Xu, X.; Hu, R. Self-Supervised Learning on A Lightweight Low-Light Image Enhancement Model with Curve Refinement. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 1890–1894. [Google Scholar]
  80. Guo, C.G.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  81. Wang, L.W.; Liu, Z.S.; Siu, W.C.; Lun, D.P. Lightening network for low-light image enhancement. IEEE Trans. Image Process. 2020, 29, 7984–7996. [Google Scholar] [CrossRef]
  82. Lee, C.; Lee, C.; Kim, C.S. Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef]
  83. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef] [PubMed]
  84. Lv, F.; Li, Y.; Lu, F. Attention guided low-light image enhancement with a large scale low-light simulation dataset. arXiv 2019, arXiv:1908.00682. [Google Scholar] [CrossRef]
  85. Loh, Y.P.; Chan, C.S. Getting to know low-light images with the exclusively dark dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
  86. Gonzalez, R.C. Digital Image Processing; Pearson Education India: Noida, India, 2009. [Google Scholar]
  87. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
  88. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  89. Papasaika-Hanusch, H. Digital image PROCESSING Using Matlab; Institute of Geodesy and Photogrammetry, ETH Zurich: Zurich, Switzerland, 1967; Volume 63. [Google Scholar]
  90. Celik, T.; Tjahjadi, T. Contextual and variational contrast enhancement. IEEE Trans. Image Process. 2011, 20, 3431–3441. [Google Scholar] [CrossRef]
  91. Pizer, S.M. Contrast-limited adaptive histogram equalization: Speed and effectiveness stephen m. pizer, r. eugene johnston, james p. ericksen, bonnie c. yankaskas, keith e. muller medical image display research group. In Proceedings of the First Conference on Visualization in Biomedical Computing, Atlanta, GA, USA, 22–25 May 1990; Volume 337. [Google Scholar]
  92. Cao, G.; Huang, L.; Tian, H.; Huang, X.; Wang, Y.; Zhi, R. Contrast enhancement of brightness-distorted images by improved adaptive gamma correction. Comput. Electr. Eng. 2018, 66, 569–582. [Google Scholar] [CrossRef]
  93. Ying, Z.; Li, G.; Gao, W. A bio-inspired multi-exposure fusion framework for low-light image enhancement. arXiv 2017, arXiv:1711.00591. [Google Scholar]
  94. Afifi, M.; Derpanis, K.G.; Ommer, B.; Brown, M.S. Learning Multi-Scale Photo Exposure Correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9157–9167. [Google Scholar]
  95. Zhang, F.; Li, Y.; You, S.; Fu, Y. Learning Temporal Consistency for Low Light Video Enhancement From Single Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4967–4976. [Google Scholar]
  96. Li, C.; Guo, J.; Porikli, F.; Pang, Y. LightenNet: A convolutional neural network for weakly illuminated image enhancement. Pattern Recognit. Lett. 2018, 104, 15–22. [Google Scholar] [CrossRef]
  97. Hu, Y.; He, H.; Xu, C.; Wang, B.; Lin, S. Exposure: A white-box photo post-processing framework. ACM Trans. Graph. (TOG) 2018, 37, 1–17. [Google Scholar] [CrossRef]
  98. Ying, Z.; Li, G.; Ren, Y.; Wang, R.; Wang, W. A new image contrast enhancement algorithm using exposure fusion framework. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Ystad, Sweden, 22–24 August 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 36–46. [Google Scholar]
  99. Pu, T.; Wang, S. Perceptually motivated enhancement method for non-uniformly illuminated images. IET Comput. Vis. 2018, 12, 424–433. [Google Scholar] [CrossRef]
  100. Al-Ameen, Z. Nighttime image enhancement using a new illumination boost algorithm. IET Image Process. 2019, 13, 1314–1320. [Google Scholar] [CrossRef]
  101. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
Figure 1. A visual representation of results from top ten deep learning methods on six datasets. The rows are showing the results produced by different algorithms, whereas the columns are showing datasets.
Figure 1. A visual representation of results from top ten deep learning methods on six datasets. The rows are showing the results produced by different algorithms, whereas the columns are showing datasets.
Remotesensing 14 04608 g001
Figure 2. A visual representation of results from top ten classical methods on six datasets. The rows are showing the results produced by different algorithms, whereas the columns are showing datasets.
Figure 2. A visual representation of results from top ten classical methods on six datasets. The rows are showing the results produced by different algorithms, whereas the columns are showing datasets.
Remotesensing 14 04608 g002
Figure 3. Different IQA metrics are used to rank the enhancement methods. Rank values range from 1 to 31. A rank value of 1 indicates the highest performance based on a particular IAQ method, and a rank value of 1 indicates the worst performance. The average rank is shown in red.
Figure 3. Different IQA metrics are used to rank the enhancement methods. Rank values range from 1 to 31. A rank value of 1 indicates the highest performance based on a particular IAQ method, and a rank value of 1 indicates the worst performance. The average rank is shown in red.
Remotesensing 14 04608 g003
Figure 4. Avg. ranking versus Time is shown for each enhancement method. Only the methods take less than 1 s on CPU (Intel(R) Core(TM) i7-6700 CPU @ 3.40 GHz 3.41 GHz) with 16 GB RAM to process the image of size 400 × 600 × 3 is shown in the figure. Red dots represent non-Retinex methods, while blue dots represent Retinex methods.
Figure 4. Avg. ranking versus Time is shown for each enhancement method. Only the methods take less than 1 s on CPU (Intel(R) Core(TM) i7-6700 CPU @ 3.40 GHz 3.41 GHz) with 16 GB RAM to process the image of size 400 × 600 × 3 is shown in the figure. Red dots represent non-Retinex methods, while blue dots represent Retinex methods.
Remotesensing 14 04608 g004
Figure 5. Avg. ranking versus Time is shown for each enhancement method. Only the methods take less than 0.5 s on GPU (NVIDIA Titan Xp GPU) to process the image of size 400 × 600 × 3 is shown in the figure. Red dots represent non-Retinex methods, while blue dots represent Retinex methods.
Figure 5. Avg. ranking versus Time is shown for each enhancement method. Only the methods take less than 0.5 s on GPU (NVIDIA Titan Xp GPU) to process the image of size 400 × 600 × 3 is shown in the figure. Red dots represent non-Retinex methods, while blue dots represent Retinex methods.
Remotesensing 14 04608 g005
Figure 6. Each test dataset has been ranked based on its difficulty. Rank values range from 1 to 9. A lower rank indicates less difficulty, a higher rank indicates more difficulty.
Figure 6. Each test dataset has been ranked based on its difficulty. Rank values range from 1 to 9. A lower rank indicates less difficulty, a higher rank indicates more difficulty.
Remotesensing 14 04608 g006
Table 1. Quantitative comparison of enhancement algorithms on nine test datasets using LOE metric. A lower value of the LOE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
Table 1. Quantitative comparison of enhancement algorithms on nine test datasets using LOE metric. A lower value of the LOE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
DatasetsLIMELOLDICMVVMEFNPELSRWSLLExDarkAverage
Methods
Non-Retinex MethodsHE [89]290.280423.910283.980280.750406.930184.590122.84753.990408.76358.222
DHE [27]7.66322.22775.60821.0137.85223.97413.93010.177138.04935.610
BPDHE [68]6.960125.04614.9364.1105.4807.6435.985382.146134.77476.342
CVC [90]99.386286.840135.32491.21797.464131.478124.946324.260189.896164.534
CLAHE [91]183.094397.432386.183209.867224.280379.588242.572504.013252.236308.807
AGCWD [29]10.0750.132557.48214.7776.04631.4321.4636.132137.99031.932
IAGC [92]63.028170.19053.50255.94366.71041.48877.123278.054165.790113.600
BIMEF [93]136.898141.159239.271102.891155.616225.588117.777480.848237.563212.589
MBLLEN [71]122.188302.577176.58079.013131.243123.871168.128484.809190.384207.076
GLADNet [72]123.603349.720285.239145.034199.632203.488204.887518.189262.524254.702
DLN [81]132.594264.065404.673325.572189.831-176.527528.411212.723-
Zero-DCE [80]135.032209.426340.803145.435164.262312.392219.127539.673315.084280.775
Exposure Correction [94]242.461438.420362.552220.876275.476314.833288.659588.132307.881349.604
StableLLVE [95]134.130267.686476.374192.262198.069394.811179.101344.573248.400287.660
LightenNet [96]681.834387.204772.380328.510896.201714.390930.978924.638636.000698.788
White-box [97]90.876125.682195.516124.11596.704120.68784.279370.972135.606156.695
LLFlow [38]365.530367.153563.765300.058430.534538.078685.344764.261445.274511.808
Retinex-based MethodsLIME [21]559.618404.114818.660460.440618.480870.215434.4851103.98575.987649.553
NPE [50]300.505317.399264.604352.294344.953257.010435.676293.158358.018327.889
JieP [58]249.137314.798287.305137.026292.798305.435216.597690.829345.754323.818
PM-SIRE [49]113.63173.558152.779113.031166.640104.945143.945189.09193.194142.148
WV-SRIE [20]106.30883.806162.22469.480210.261155.683131.724236.846220.823158.856
MSRCR [19]842.0291450.951185.111280.68973.8931252.07893.2161211.11676.4151115.43
CRM [59]271.65221.818450.102174.751285.250534.275119.712619.537352.672314.419
EFF [98]136.898141.159239.271102.891155.616255.588117.777480.848237.563207.512
pmea [99]491.663725.647477.792318.569679.002610.183418.0461005.66529.189595.511
RetinexNet [22]472.189770.105636.160391.745708.250838.310591.278950.895548.905679.456
KinD [36]214.893434.595261.771134.844275.474241.221379.899479.139308.869303.412
RetinexDIP [64]767.0421084.35852.782396.417926.9481099.39572.4291283.77633.489856.197
RRDNet [25]72.91721.438261.429168.601100.735-136.011380.7471.100-
KinD++ [60]573.877720.025493.882258.744629.841-727.695555.363484.989-
IBA [100]14.6570.1616445.574169.71412.823364.810137.72721.758284.333179.613
Self-supervised Network [24]241.639322.628737.847282.273311.342581.691261.280467.892333.842412.349
TBEFN [23]289.754464.947617.100271.871419.666527.675386.583859.878389.558492.160
Average178.196342.070387.311227.201313.656378.930286.698548.053320.401-
Table 2. Quantitative comparison of enhancement algorithms on nine test datasets using NIQE metric. A lower value of the NIQE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
Table 2. Quantitative comparison of enhancement algorithms on nine test datasets using NIQE metric. A lower value of the NIQE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
DatasetsLIMELOLDICMVVMEFNPELSRWSLLExDarkAverage
Methods
Non-Retinex MethodsInput4.3576.7484.2743.5244.2633.7175.3915.3585.1284.800
HE [89]3.8848.4133.8502.6623.8703.5353.9636.4384.7524.685
DHE [27]3.9148.9873.7802.6483.5183.5103.6266.2924.5184.610
BPDHE [68]3.827NaN3.7862.8573.9023.5313.935NaN4.727-
CVC [90]4.0298.0143.8232.6923.6363.4984.1275.8284.6624.535
CLAHE [91]3.9077.2683.7922.7843.6063.4614.5815.7564.7344.490
AGCWD [29]4.0327.5283.8682.9703.6293.5443.7335.6604.5824.434
IAGC [92]3.9517.4184.0153.0123.6523.5983.9635.7404.5574.494
BIMEF [93]3.8597.5153.8452.8073.3293.5403.8795.7474.5144.397
MBLLEN [71]4.5134.3574.2304.1794.7393.9484.7223.9794.4784.329
GLADNet [72]4.1286.4753.6812.7903.3603.5223.3975.0663.7674.009
DLN [81]4.3414.8833.7893.2284.022-4.4194.3764.415-
Zero-DCE [80]3.7697.7673.5673.2163.2833.5823.7205.9983.9174.381
Exposure Correction [94]4.2157.8863.5883.0784.4563.4143.8204.9424.3574.443
StableLLVE [95]4.2344.3724.0613.4203.9243.4864.3674.1854.0533.984
LightenNet [96]3.7317.3233.5392.9953.3503.4073.5835.4534.0254.209
White-box [97]4.5987.8194.6303.5584.6224.0044.3147.1385.5345.202
LLFlow [38]3.9565.4453.7653.0263.4413.4983.5644.7224.0943.944
Retinex-based MethodsLIME [21]4.1098.1293.8602.4943.5763.6583.6556.3724.5884.542
NPE [50]3.5788.1583.7362.4713.3373.4263.5765.7714.2204.337
JieP [58]3.7196.8723.6782.7653.3903.5224.0155.6224.2154.260
PM-SIRE [49]4.0507.5063.9783.0103.4503.5313.9845.4354.3834.410
WV-SRIE [20]3.7867.2863.8982.8493.4743.4503.8265.4534.2414.310
MSRCR [19]3.9398.0063.9482.8143.6883.7803.8725.5744.9044.573
CRM [59]3.8547.6863.8012.6173.2643.5623.7216.0084.5254.391
EFF [98]3.8597.5153.8452.8073.3293.5403.8795.7474.5144.390
pmea [99]3.8438.2813.8362.5733.4313.5983.6946.2374.2964.493
RetinexNet [22]4.5978.8794.4152.6954.4104.4644.1507.5734.5515.142
KinD [36]4.7634.7094.1503.0263.8763.5573.5434.4504.3403.956
RetinexDIP [64]3.7357.0963.7052.4963.2453.6384.0815.88284.2344.297
RRDNet [25]3.9367.4363.6372.8143.508-4.1265.5244.010-
KinD++ [60]4.3854.6163.8042.6603.738-3.3545.0904.343-
IBA [100]4.0627.8843.7233.3103.5363.6303.7285.8374.2734.490
Self-supervised Network [24]4.8193.7534.7173.5484.3514.6024.0615.4004.0484.310
TBEFN [23]3.9543.4363.5032.8843.2273.2923.4784.6483.6213.511
Average3.9356.7283.8892.9563.6983.6263.9335.4094.403-
Table 3. Quantitative comparison of enhancement algorithms on nine test datasets using entropy [86] metric. A higher value of the entropy metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
Table 3. Quantitative comparison of enhancement algorithms on nine test datasets using entropy [86] metric. A higher value of the entropy metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
DatasetsLIMELOLDICMVVMEFNPELSRWSLLExDarkAverage
Methods
Non-Retinex MethodsInput6.1484.9156.6866.7156.0757.0175.4155.6165.7446.023
HE [89]7.3427.1847.2217.3837.1187.7566.8746.6626.7087.113
DHE [27]7.0976.7497.1417.2256.9137.5126.5316.7416.6136.930
BPDHE [68]6.6105.9326.9686.9776.4207.3486.2605.1916.1886.413
CVC [90]6.8756.4097.0557.2166.7557.4026.3186.5496.4656.772
CLAHE [91]6.7645.6797.0887.0566.5837.4086.0336.5916.3026.595
AGCWD [29]6.7926.4156.9257.0216.6487.3986.3946.2786.2486.666
IAGC [92]6.9916.2477.0157.1936.8787.3516.3186.6986.5546.782
BIMEF [93]7.0066.1457.0297.2436.8987.3116.5166.4526.4646.760
MBLLEN [71]7.1647.3037.2557.3337.0817.3867.2367.1977.1327.240
GLADNet [72]7.5027.3567.4047.4477.4087.4527.3937.5817.2507.412
DLN [81]7.1217.2777.2507.5357.255-7.2027.5767.129-
Zero-DCE [80]7.1666.5317.2247.5727.0937.4027.0356.5456.9327.042
Exposure Correction [94]7.1127.2447.2566.9626.9557.5317.0397.2476.9077.142
StableLLVE [95]7.2276.6257.0107.3857.2417.0426.8467.4397.1297.090
LightenNet [96]7.2346.1197.2637.4117.3087.3987.5996.1306.6886.990
White-box [97]5.9845.9256.0515.4755.3917.3806.3525.4605.2755.914
LLFlow [38]7.4687.4627.4257.5657.3667.5647.3437.3047.1257.394
Retinex-based MethodsLIME [21]7.3157.1296.9467.3957.1397.3327.2796.4186.5827.031
NPE [50]7.3686.9717.2087.5507.4057.4467.3186.4186.7727.139
JieP [58]7.0876.4437.2187.4577.1047.4276.7946.4736.6316.943
PM-SIRE [49]7.0066.3227.0847.3096.8947.4046.6966.3256.4416.812
WV-SRIE [20]6.9996.3487.0887.4016.9427.3866.6636.1906.4636.812
MSRCR [19]6.5636.8416.6776.9576.4556.7626.8955.9366.3196.605
CRM [59]6.4874.9716.6406.5596.2037.0265.4946.0685.9216.115
EFF [98]7.0066.1457.0297.2436.8987.3116.5166.4526.4646.760
pmea [99]7.2846.8247.2207.4797.2737.4497.0746.6386.7257.088
RetinexNet [22]7.4897.2337.4137.5757.4487.4637.2437.3857.2737.379
KinD [36]7.3887.0177.2117.4987.3287.4357.2097.4086.9057.251
RetinexDIP [64]6.9745.3757.2147.5576.6617.3816.3526.2136.6686.678
RRDNet [25]6.6465.4577.1427.2756.453-6.7756.0776.426-
KinD++ [60]7.4867.0657.3327.6277.463-7.3167.4527.034-
IBA [100]5.9054.9136.8267.2555.7497.0357.1465.4656.9716.420
Self-supervised Network [24]7.4977.4046.6757.2987.4696.9977.3977.4847.2967.253
TBEFN [23]7.4366.8757.3287.5077.3837.3667.0477.5197.3137.292
Average7.0006.4817.0727.2476.9047.3406.7986.6056.659-
Table 4. Quantitative comparison of enhancement algorithms on nine test datasets using BRISQUE metric. A lower value of the BRISQUE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
Table 4. Quantitative comparison of enhancement algorithms on nine test datasets using BRISQUE metric. A lower value of the BRISQUE metric indicates better performance. The first, second, and third best scores are highlighted with red, blue, and green colors, respectively.
DatasetsLIMELOLDICMVVMEFNPELSRWSLLExDarkAverage
Methods
Non-Retinex MethodsInput25.14221.92928.11529.38029.06626.67332.72625.30434.01528.401
HE [89]21.41139.55925.35918.93725.31325.44428.21940.01529.03428.985
DHE [27]22.33637.86625.99324.38021.46627.00826.47738.24828.95128.719
BPDHE [68]21.728NaN25.097225.18322.34526.42525.129NaN27.417-
CVC [90]22.58927.10124.62021.76619.28525.69326.80829.00726.97925.126
CLAHE [91]23.27429.46324.24823.48022.70125.36829.57031.57928.54326.825
AGCWD [29]21.96428.42124.72523.96119.42026.411723.36729.74026.16125.276
IAGC [92]24.31424.05827.02626.61721.84326.04423.85432.81327.42926.211
BIMEF [93]23.13527.65126.81122.54220.22025.50424.07734.98227.91026.174
MBLLEN [71]30.38623.07831.60335.07632.38929.42330.32822.10329.01229.127
GLADNet [72]22.28626.07326.25324.06822.90824.96922.80233.75424.76525.657
DLN [81]27.71528.98526.91429.78228.378-33.59726.79831.187-
Zero-DCE [80]23.33430.30530.65330.78625.48430.15925.82736.57226.76129.568
Exposure Correction [94]27.48328.35729.84731.69429.59726.76826.39128.63232.52029.204
StableLLVE [95]28.88532.19428.15028.29528.47525.66230.56325.85027.74928.367
LightenNet [96]19.52328.06228.79123.50221.46927.66725.14428.05525.92426.077
White-box [97]28.80731.72133.21235.73333.59926.67125.08139.45037.42932.862
LLFlow [38]22.85629.70925.07223.15725.67325.39222.01128.04126.13325.649
Retinex-based MethodsLIME [21]23.57233.97327.13725.39425.15828.57627.65835.82928.70428.986
NPE [50]22.50633.85825.49324.65422.32024.98627.19533.86128.45227.539
JieP [58]22.19327.08723.63322.94121.21425.49823.42130.20725.30924.914
PM-SIRE [49]24.65927.69427.59724.28724.32127.34225.34530.01426.67626.635
WV-SRIE [20]24.18127.61127.69824.43422.08825.76024.70028.28126.75025.894
MSRCR [19]19.38430.34525.79919.28219.09124.18925.78930.30025.41524.957
CRM [59]23.47729.59926.60122.36820.71625.72624.39637.72328.73326.939
EFF [98]23.13527.65126.81122.54220.22025.50424.07734.98227.91026.174
pmea [99]21.39032.91325.83224.97221.75626.35825.35838.13228.32127.874
RetinexNet [22]26.10139.58626.65622.45926.03629.08629.02141.50630.17030.565
KinD [36]26.77326.64530.69628.88730.43827.75326.76330.53929.25628.872
RetinexDIP [64]21.72319.67925.19925.33823.60526.67125.08132.61832.17526.296
RRDNet [25]24.49926.83429.62123.39617.750-27.10029.20527.606-
KinD++ [60]20.02525.08627.85228.16430.024-26.97334.97831.775-
IBA [100]24.33631.11732.10334.64623.74829.93325.82632.53726.63929.569
Self-supervised Network [24]30.19219.76829.52930.18328.35529.15926.20532.01627.99027.901
TBEFN [23]25.72017.34623.60623.65124.43524.035522.92930.67625.06423.968
Average23.00927.75227.26725.84124.31226.62126.28031.26728.425-
Table 5. Computational time of classical methods in terms of seconds on CPU is reported. The red, blue, and green colors are used to indicate the best, second best, and third best performance, respectively.
Table 5. Computational time of classical methods in terms of seconds on CPU is reported. The red, blue, and green colors are used to indicate the best, second best, and third best performance, respectively.
Image Size400 × 600 × 3640 × 960 × 32304 × 1728 × 32848 × 4256 × 3Avg.
Methods
Non-Retinex MethodsHE [89]0.000790.00140.00710.02030.00742
DHE [27]23.59059.625409.6281253.897436.685
BPDHE [68]0.0780.3381.6303.3181.341
CVC [90]0.0860.2301.1503.5331.250
CLAHE [91]0.000330.000990.00580.02260.00743
AGCWD [29]0.0310.0530.3441.0790.377
IAGC [92]0.0380.1551.0252.2530.867
BIMEF [93]0.1230.3591.8115.1011.848
Exposure Correction [94]0.7210.7780.90318.5015.226
LightenNet [96]3.0917.12645.990137.83548.510
LLFlow [38]24.74060.022363.2811403.92462.991
Retinex-based MethodsLIME [21]0.0900.2961.5064.6501.635
NPE [50]13.06131.025213.168648.832226.522
JieP [58]0.6460.8742.3076.5972.606
PM-SIRE [49]0.4021.34028.94828.42314.778
WV-SRIE [20]0.9153.13640.701182.26756.755
MSRCR [19]0.3220.7042.7878.5313.086
CRM [59]0.1660.4362.6268.1342.840
EFF [98]0.1360.4071.9735.4221.984
pmea [99]0.6460.8742.3076.5972.606
IBA [100]0.0320.08290.5121.3850.503
Table 6. The computation time (seconds) and number of parameters (millions) for deep learning-based methods on GPUs (NVIDIA TITAN Xp) are reported. The red, blue, and green colors are used to indicate the best, second best, and third best performance, respectively.
Table 6. The computation time (seconds) and number of parameters (millions) for deep learning-based methods on GPUs (NVIDIA TITAN Xp) are reported. The red, blue, and green colors are used to indicate the best, second best, and third best performance, respectively.
Image Size400 × 600 × 3640 × 960 × 32304 × 1728 × 32848 × 4256 × 3Avg.Parameters.
Methods
Non-RetinexStableLLVE [95]0.00470.0050.00760.0970.0284.310 M
MBLLEN [71]0.2400.3271.6018.1332.5750.450 M
GLADNet [72]0.1470.1610.6762.7720.9390.930 M
White-box [97]6.0406.4839.83315.2009.3898.560 M
DLN [81]0.0090.0150.0580.1970.0700.700 M
Zero-DCE [80]0.00250.00260.0210.0430.0170.079 M
RetinexRetinexNet [22]0.1550.1620.5911.2890.5490.440 M
KinD [36]0.3340.6043.5395.2132.4230.255 M
RetinexDIP [64]33.92437.01563.443112.54561.7320.707 M
RRDNet [25]59.479128.217893.03003.51021.10.128 M
KinD++ [60]0.3370.8575.40819.7466.5878.275 M
Self-supervised Net [24]0.0220.0540.3661.2120.4140.485 M
TBEFN [23]0.1710.1660.5500.8870.4440.490 M
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rasheed, M.T.; Guo, G.; Shi, D.; Khan, H.; Cheng, X. An Empirical Study on Retinex Methods for Low-Light Image Enhancement. Remote Sens. 2022, 14, 4608. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184608

AMA Style

Rasheed MT, Guo G, Shi D, Khan H, Cheng X. An Empirical Study on Retinex Methods for Low-Light Image Enhancement. Remote Sensing. 2022; 14(18):4608. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184608

Chicago/Turabian Style

Rasheed, Muhammad Tahir, Guiyu Guo, Daming Shi, Hufsa Khan, and Xiaochun Cheng. 2022. "An Empirical Study on Retinex Methods for Low-Light Image Enhancement" Remote Sensing 14, no. 18: 4608. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184608

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop