Next Article in Journal
Enhancing EEG-Based Emotion Detection with Hybrid Models: Insights from DEAP Dataset Applications
Previous Article in Journal
Displacement Sensing Using Bimodal Resonance in Over-Coupled Inductors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles

1
Department of Smart Vehicle Engineering, Konkuk University, Seoul 05029, Republic of Korea
2
Graduate School of Future Defense Technology Convergence, Konkuk University, Seoul 05029, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(6), 1825; https://doi.org/10.3390/s25061825
Submission received: 23 January 2025 / Revised: 28 February 2025 / Accepted: 13 March 2025 / Published: 14 March 2025
(This article belongs to the Section Vehicular Sensing)

Abstract

:
Intelligent vehicles, such as autonomous cars, drones, and robots, rely on sensors to gather environmental information and respond accordingly. RGB cameras are commonly used due to their low cost and high resolution but are limited in low-light conditions. While employing LiDAR or specialized cameras can address this issue, these solutions often incur high costs. Deep learning-based low-light image enhancement (LLIE) methods offer an alternative, but existing models struggle to adapt to road scenes. Furthermore, most LLIE models rely on supervised training but are heavily constrained by the lack of low-light and normal-light paired datasets. In particular, obtaining paired datasets for driving scenes is extremely challenging. To address these issues, this paper proposes Squeeze-EnGAN, a memory-efficient, GAN-based LLIE method capable of unsupervised learning without paired image datasets. Squeeze-EnGAN incorporates a fire module into a U-net architecture, substantially reducing the number of parameters and Multiply-Accumulate Operations (MACs) compared to its base model, EnlightenGAN. Additionally, Squeeze-EnGAN achieves real-time performance on devices like Jetson Xavier (0.061 s). Significantly, enhanced images improve object detection performance over original images, demonstrating the model’s potential to aid high-level vision tasks in intelligent vehicles.

1. Introduction

In autonomous driving systems, vehicles acquire information about their surroundings through various sensors. Among these sensors, RGB cameras are highly sought after due to their ability to capture high-resolution information and their relatively low cost. Furthermore, the color of an object is a critical feature for distinguishing objects, and cameras are the only sensors capable of capturing color information. For this reason, RGB cameras are widely utilized in tasks such as object detection, object tracking, semantic segmentation, and visual SLAM (Simultaneous Localization and Mapping) [1,2,3] in autonomous vehicles and intelligent mobility systems. Additionally, active research is being conducted on LiDAR-camera sensor fusion to identify object categories that are challenging to discern using LiDAR alone [4,5].
However, due to the characteristics of camera sensors, their performance is highly sensitive to ambient lighting conditions, which can lead to performance degradation. In particular, RGB cameras struggle to identify objects in low-light environments, such as during nighttime or in areas without light sources. This issue poses significant risks for intelligent mobility systems driving at night, as they may fail to respond appropriately to dynamic objects like pedestrians or other vehicles, as well as static obstacles such as poles, buildings, vegetation, and potholes [6]. Additionally, autonomous vehicles may fail to recognize visual cues such as lane markings or traffic signs, potentially resulting in violations of traffic regulations. While the vehicle’s headlights can serve as a built-in light source to detect objects ahead, their effective range is limited. Furthermore, in low-light conditions, the intensity of received light is diminished, leading to a reduced signal-to-noise ratio and increased susceptibility to noise.
To address this issue, this study developed a deep learning-based low-light image enhancement (LLIE) model to overcome the limitations of RGB cameras in autonomous driving environments. In particular, the main objective was to design lightweight LLIE models suitable for autonomous driving applications. Since reaction speed to surrounding objects is crucial in autonomous driving systems, all stages—sensor input, object recognition, decision-making and path generation, and control—must be executed in rapid cycles [7]. If the computational resources required for the driving process (e.g., computation time and memory) are excessive, the decision-making cycle may slow down, impairing the system’s ability to respond swiftly to sudden obstacles or environmental changes. The entire autonomous driving process typically consists of numerous sub-tasks, and LLIE can be considered one of these sub-tasks. If a single task consumes fewer resources while meeting the real-time computational constraints, other tasks can allocate more resources for complex calculations or incorporate new functionalities. This highlights the importance of reducing computation time for LLIE as a critical challenge in integrating it into the autonomous driving process. This challenge becomes even more significant given that most intelligent mobility systems operate within edge computing environments [8,9].
We identified the following challenges in applying existing low-light image enhancement (LLIE) research to autonomous driving environments:
  • General LLIE models struggle to adequately reflect the characteristics of driving environments. Issues such as noise, overexposure, and blur make it difficult to reliably apply these models to autonomous driving.
  • Most deep learning-based LLIE approaches rely on supervised learning methods that require paired normal and low-light images for training. However, obtaining natural image pairs for the same scene is challenging, particularly for driving scenarios where creating such pairs is even more constrained.
  • For LLIE to be applied to autonomous driving, computational resource usage must be minimized, but few models prioritize this requirement.
To address these issues, this study proposes Squeeze-EnGAN, an LLIE model designed to be reliably applied to autonomous driving systems. The foundation of Squeeze-EnGAN is EnlightenGAN [10], which first demonstrated high-performance LLIE without the need for paired images through the GAN [11] architecture. By inheriting the advantage of training on single images, Squeeze-EnGAN is robust across diverse driving scenarios using data captured in various environments.
Additionally, Squeeze-EnGAN incorporates the fire module proposed in SqueezeNet [12] to make the U-net [13] generator network more lightweight, addressing the computational constraints of edge computing environments. Compared to EnlightenGAN, Squeeze-EnGAN significantly reduces the number of parameters and computational overhead. Experimental results using an on-board computer demonstrate the effectiveness of this lightweight design by comparing inference times. Moreover, a comparative analysis of output image quality confirms that the model achieves minimal performance degradation despite its lightweight design.

2. Related Works

2.1. Traditional Methods

To overcome the limitations of RGB cameras and obtain high-quality images, LLIE techniques have been continuously studied. Traditional LLIE methods can be broadly categorized into hardware-based and software-based approaches.
Hardware-based methods include providing additional light sources, such as using a flash during capture, or increasing the exposure time to allow the camera to receive more light. However, in autonomous driving environments, adding external light sources is often impractical. Furthermore, increasing the exposure time can result in motion blur, which degrades object recognition performance due to afterimages of moving subjects.
In contrast, software-based methods modify the captured data rather than the imaging process itself. For instance, histogram equalization [14], which redistributes the pixel intensity distribution to achieve a more uniform histogram, is a commonly used technique. Another widely studied approach employs the Retinex theory [15], which postulates that an image can be decomposed into illumination and reflectance components. In this approach, the brightness of the illumination component is enhanced, and noise in the reflectance component is reduced. However, these methods often struggle to deliver consistent performance under varying low-light conditions, limiting their applicability to autonomous driving environments, where the system must adapt to diverse scenarios.

2.2. Deep Learning Approaches

Advancements in computing hardware have accelerated progress in the field of artificial intelligence, leading to the application of deep learning techniques in low-light image enhancement. Early studies incorporating deep learning, such as LLNet [16], demonstrated superior performance compared to traditional methods, and since then, a variety of neural network models have been proposed.
Supervised learning methods for low-light image enhancement require training datasets consisting of pairs of low-light images and their corresponding normal-light counterparts captured under identical conditions. This approach has been widely adopted in most neural networks due to its superior performance compared to other methods. However, in real-world scenarios, obtaining perfectly paired low-light and normal-light images of the same scene is highly challenging.
In contrast, unsupervised learning methods rely on independent images rather than paired datasets for training, thereby eliminating the data pairing constraints faced by supervised approaches. EnlightenGAN [10], the first neural network to utilize an unsupervised approach, employs a GAN [11]-based architecture. EnlightenGAN offers the significant advantage of enabling training with unpaired images.
For intelligent mobility systems, which operate in diverse environments, it is essential to apply robust models capable of adapting to various conditions. Therefore, this study adopts an unsupervised approach, leveraging unpaired and diverse environmental training datasets to enhance adaptability and robustness.

2.3. Adversarial Learning Without Supervision

The Generative Adversarial Network (GAN) [11], proposed by Ian Goodfellow, is a model that consists of two sub-network structures: a generator and a discriminator. GAN employs a training strategy where the generator and discriminator compete against each other. The generator takes random noise vectors as input and learns to produce outputs that mimic real images. In contrast, the discriminator receives either the generator’s output or a real image as input and learns to distinguish whether the input is a real image. As the two networks continue to compete during training, the generator’s outputs become increasingly similar to real images.
Subsequent developments, such as Deep Convolutional GAN (DCGAN) [17], applied convolutional layers and marked significant progress in tasks like image synthesis and style transfer. GANs have also demonstrated their utility in low-level vision tasks, including super-resolution [18], denoising [19], and dehazing [20]. Some studies [21,22] have proposed unsupervised models that learn domain mapping without the need for paired data. In this context, EnlightenGAN [10] has shown high performance in LLIE tasks by using an unsupervised GAN framework, eliminating the need for paired data.
Furthermore, EnlightenGAN utilizes a one-path GAN structure, which requires less computation compared to models employing cycle-consistency. However, EnlightenGAN was not designed with a focus on model lightweighting. To address this limitation, this study proposes a more resource-efficient and lightweight model to improve computational efficiency.

2.4. U-net with Fire Module

SqueezeNet [12] proposed the fire module to improve the memory efficiency of convolutional neural networks and reduce real-time inference latency. The fire module is a lightweight alternative to a single convolutional layer and is structured as shown in Figure 1. A standard convolution layer, illustrated in Figure 1a, contains filters equal to the number of specified output channels, with each filter applied across all input channels to produce a corresponding output channel. As a result, the number of channels increases to match the number of output channels. On the other hand, the fire module consists of two layers, as shown in Figure 1b. The first layer is a 1 × 1 convolutional layer that squeezes the input channels to a predefined number of intermediate channels. The number of intermediate channels is usually set to be smaller than the number of input channels. The feature maps produced by the first layer are simultaneously passed to two parallel convolutional layers: one is a 1 × 1 convolutional layer and the other is a 3 × 3 convolutional layer. The number of output channels from each of these two layers is set to half of the final output size and the final output of the fire module is obtained by concatenating the feature maps produced by both layers.
The fire module employs the following strategies to enable more efficient computation compared to conventional convolutional layers: (1) Replacing some 3 × 3 filters with 1 × 1 filters. This reduces the number of parameters to approximately one-ninth of the original. (2) Reducing the number of input channels for the 3 × 3 filters. Since the input channels are squeezed using 1 × 1 filters before being processed by the 3 × 3 filters, fewer channels require computation in the 3 × 3 layer. Using these strategies, SqueezeNet achieved AlexNet-level accuracy with a model size of less than 0.5 MB.
Squeeze U-net [23] is a study that applied the fire module to the U-net [13] architecture to create a lightweight version of U-net. This approach retained the original model’s semantic segmentation performance while reducing memory usage and computation time. Inspired by this approach, the present study replaces the U-net generator network in EnlightenGAN with a Squeeze U-net structure, developing a lightweight model that maintains performance and is more suitable for onboard systems.

3. Methods

Inspired by EnlightenGAN [10], the overall architecture of Squeeze-EnGAN comprises a generator with a U-net [13] structure and a global-local discriminator. This design enables training with unpaired images and allows the model to adapt to datasets from diverse environments. Furthermore, Squeeze-EnGAN incorporates the Squeeze U-net [23] approach, which utilizes fire modules in the generator network. This integration facilitates memory-efficient LLIE tasks.

3.1. Generator with Fire Modules

The generator is based on an encoder-decoder structure derived from U-net [13] and the overall structure is illustrated in Figure 2. U-net is a model that has achieved significant success in the field of image semantic segmentation. It consists of an encoder that extracts features from the input image and a decoder that reconstructs a semantic representation at the original resolution from the compressed features. U-net is effective due to its use of skip connections, which help preserve high-resolution information by directly connecting corresponding encoder and decoder layers. This innovative idea has led to the development of many derivative models and has also been applied in the fields of image generation and restoration. The generator of Squeeze-EnGAN also extracts features from the original low-light image and reconstructs it into a normal-light image. As shown in Figure 2, the architecture consists of an encoder that performs down-sampling and a decoder that performs up-sampling, with skip connections linking corresponding layers between the encoder and decoder.
However, the generator of Squeeze-EnGAN replaces the standard convolution layers with fire modules [12] throughout the encoder and decoder, except for the initial layers of the encoder and the final layers of the decoder. Each fire module consists of a 1 × 1 convolution layer followed by parallel 3 × 3 and 1 × 1 convolution layers, as shown in Figure 1. The initial 1 × 1 convolution layer compresses the input channels and passes them to the subsequent convolution layers. The feature maps produced by the 3 × 3 and 1 × 1 convolution layers are concatenated to form the output of the fire module.
In the encoder, average pooling is used to progressively reduce the size of the input feature maps. During the decoder’s upsampling process, the transpose convolution layers are replaced with an upsample operation followed by a fire module to avoid checkerboard artifacts [24]. The detailed layer specifications of the generator network are provided in Appendix A.
The generator in EnlightenGAN [10] uses a self-regularized attention map derived by inverting the illumination channel of the input image to focus more on locally dark regions. However, this approach requires additional computations for the attention module and incurs overhead during preprocessing, where the illumination map is generated for each input. Squeeze-EnGAN eliminates the use of the attention module, thereby saving the computational resources consumed in this process.
As shown in Table 1, we compared the Multiply-Accumulate Operations (MACs) and the number of parameters for each layer of the generators in EnlightenGAN and Squeeze-EnGAN. The generator with fire modules achieved a significant reduction in computational complexity compared to the original model, with the total MACs reduced by a factor of 0.39 and the number of parameters reduced by a factor of 0.15. This demonstrates that our model achieves a significantly more lightweight design compared to the previous model.

3.2. Global-Local Discriminator

The discriminators take real images (real-world images) and fake images (the output images from the generator) as input. The discriminator network is trained to output a value close to 1.0 for real images and a value close to 0.0 for fake images. The discriminator’s output is used not only to update the discriminator itself by minimizing classification error but also to guide the training of the generator through adversarial loss, encouraging the generator to produce images that are increasingly indistinguishable from real ones.
Conventional GAN models use the discriminator’s output to minimize the disparity between the distributions of the generator’s output and real images. In contrast, GAN models applied to LLIE aim to reduce the distribution gap between the generator’s output and normal-light images. Furthermore, as demonstrated in previous research [10], a local discriminator is additionally employed to mitigate the impact of locally exaggerated illumination (e.g., small light sources in a dark background). Therefore, we employ two discriminators to enhance the performance of the generator, as illustrated in Figure 3.
The global discriminator takes a full-resolution image as input and processes it through a series of convolutional layers with progressively increasing receptive fields. The local discriminator, on the other hand, uses randomly cropped regions from the global discriminator’s input images, enabling the generator to enhance local quality. Notably, object detection research [25,26] has leveraged local features to detect small objects effectively. Similarly, the local discriminator reinforces detailed image contexts, aiding the detection of objects located in local regions during high-level vision tasks in autonomous driving.
The adversarial loss computed from the output of the global discriminator trains both the generator and the global discriminator. In this process, the loss combines the Relativistic Discriminator [27] and Least Square GAN (LSGAN) [28] methods. The equations for the loss functions applied to the global discriminator D and the generator G are as follows:
L D G l o b a l = E x r ~ P r e a l [ ( D ( x r ) E x f ~ P f a k e [ D ( x f ) ] 1 ) 2 ] + E x f ~ P f a k e [ ( D ( x f ) E x r ~ P r e a l [ D ( x r ) ] ) 2 ]
L G G l o b a l = E x f ~ P f a k e [ ( D ( x f ) E x r ~ P r e a l [ D ( x r ) ] 1 ) 2 ] + E x r ~ P r e a l [ ( D ( x r ) E x f ~ P f a k e [ D ( x f ) ] ) 2 ]
In Equations (1) and (2), D represents the network of the global discriminator, while x r and x f denote samples of real images (natural normal-light images) and fake images (the output images from the generator), respectively. For the local discriminator, patches are randomly cropped from both the real images and the generated outputs. Unlike the global discriminator, the local discriminator utilizes the LSGAN loss, with the values calculated for each of the patches averaged and applied. The equations for the loss functions applied to the local discriminator D and the generator G are as follows:
L D L o c a l = E x r ~ P r e a l p a t c h e s D x r 1 2 + E x f ~ P f a k e p a t c h e s [ ( D ( x f ) ) 2 ]
L G L o c a l = E x r ~ P f a k e p a t c h e s D x r 1 2
In Equations (3) and (4), D represents the network of the local discriminator, while x r and x f denote randomly cropped sample patches from the real images and fake images, respectively. In summary, Equation (1) is used to train the global discriminator, Equation (3) is employed to train the local discriminator, and Equations (2) and (4) are used to train the generator.

3.3. Perceptual Loss

Johnson et al. [29] proposed perceptual loss for image transformation tasks to preserve style between images. Instead of relying on per-pixel differences, this method uses a pre-trained VGG-16 [30] model to calculate the distance between image features, thereby maintaining visual quality. Perceptual loss has been widely applied and proven effective in low-level vision tasks such as image super-resolution, low-light enhancement, and style transfer.
In this study, since unpaired data is used, the distribution of enhanced images during adversarial training may converge to that of unrelated images, independent of the input images. This limitation of unsupervised learning can cause the output images to appear contextually and semantically different from the originals. Perceptual loss constrains the distance between the features of the input and output images, playing a crucial role in ensuring that Squeeze U-net preserves the context of the original images. Notably, prior studies [10,31] have demonstrated the VGG model’s robustness to variations in image illumination, making it particularly suitable for this task. The corresponding equation is as follows:
L p e r c e p t u a l = 1 W i ,   j H i ,   j x = 1 W i ,   j y = 1 H i ,   j ϕ i ,     j I L x , y ϕ i ,     j G I L x , y 2
In Equation (5), G represents the generator, and I L denotes the low-light image input to the generator. ϕ refers to the feature map extracted from a pre-trained VGG-16 model, where i indicates the i-th max-pooling layer of VGG-16, and j represents the j-th convolutional layer following the i-th max-pooling layer. W i ,   j and H i ,   j denote the dimensions of the extracted feature map. By default, we set i = 1 and j = 5. Similarly, Equation (5) is also applied to cropped patches used as inputs for the local discriminator. Additionally, an instance normalization layer was added before feeding inputs into the VGG-16 model.

4. Experiments

4.1. Datasets and Two Training Sessions

We supplemented the datasets used in EnlightenGAN with additional datasets focused on road scenes to address driving scenarios. Specifically, we utilized the “High-Precision Data Collection Vehicle Night City Road Data” (HP-NCR) [32] and the “High-Precision Data Collection Vehicle Daytime City Road Data” (HP-DCR) [33]. Both datasets are publicly available on the AI-Hub website managed by the National Information Society Agency (NIA), a governmental organization in South Korea.
The HP-NCR and HP-DCR datasets consist of images extracted from videos captured at a resolution of 1920 × 1080 using cameras mounted on vehicles driving through urban roads at night and during the day, respectively. The images were extracted at 10 fps from 20-s videos, with 500 driving scenes resulting in over 100,000 images per dataset. However, this study utilized only a subset of these images. The driving scenes were recorded under various weather conditions (clear, cloudy, and rainy). Notably, the HP-NCR dataset exhibits extremely dark characteristics due to its low exposure settings.
The training dataset used for Squeeze-EnGAN is the same as that used in previous research on EnlightenGAN [34,35,36,37]. During the training process, 1013 low-light images were used as inputs for the generator, while 1102 normal-light images were used as reference targets. Additionally, we designed a new training session using a modified dataset. Thanks to the flexibility of unsupervised learning, various datasets can be effectively applied, leading to improved domain adaptation performance. This training session was conducted to verify whether the model could capture diverse real-world characteristics. Specifically, we aimed to learn domain-specific characteristics of road scenes by leveraging the HP-NCR and HP-DCR datasets. In this training session, the low-light input images for the generator consisted of 500 images from the low-light dataset used in EnlightenGAN and 1000 selected images from the HP-NCR dataset, resulting in a combined set of 1500 images. For normal-light images, 1102 images from the EnlightenGAN training dataset were combined with 398 images from the HP-DCR dataset. In this study, we name the model derived from this training session Squeeze-EnGAN-R. In all training sessions, images were resized to a resolution of 640 × 400.
For testing, we employed datasets used in previous studies (DICM [38], MEF [39], NPE [40], LIME [41], VV [42], and LOL [34]) along with 554 selected images from the HP-NCR dataset. The model outputs for the MEF, NPE, LIME, and VV datasets are shown in Figure 4, while the results for the LOL dataset are presented in Figure 5, and those for the HP-NCR dataset are shown in Figure 6.

4.2. Implementation Details

Our model was implemented using the PyTorch 1.12.1 library and trained in an environment configured with CUDA 11.3 and Python 3.9.17. The training of Squeeze-EnGAN was conducted over a total of 200 epochs. Initially, the model was trained for 100 epochs with a learning rate of 1 × 10−4, followed by a linearly decayed learning rate for the remaining 100 epochs. The local discriminator was configured to use 6 patches, each with a size of 64 × 64. In this study, we set the batch size to 24 and utilized the Adam Optimizer [43] for training. The entire training process was performed using a single NVIDIA A6000 GPU, requiring a total of 4 h 30 min to complete. In the case of Squeeze-EnGAN-R, the larger dataset increased the training time to 6 h 40 min.

4.3. Results

We performed comparative analyses with previous methods. Among them, URetinex-Net [44], CIDNet [45], and LLFlow [46] are supervised learning models, while ZeroDCE [47], PairLIE [48], and EnlightenGAN are unsupervised learning models. In Figure 4, which compares the result images of different models, the first to last columns represent the results for the LIME, DICM, MEF, NPE, and VV datasets, respectively. As shown in Figure 4f, replacing convolution layers with lightweight fire modules still yields high-quality results compared to other models. To compare the output quality of Squeeze-EnGAN with that of existing studies, we evaluated the average NIQE (Natural Image Quality Evaluator) score [49] for each test dataset. NIQE is a well-known no-reference image quality assessment metric, enabling quality evaluation without the need for ground-truth images. It measures how closely the test image resembles the statistical characteristics of natural images, with lower scores indicating better quality. Table 2 presents the NIQE scores measured for the datasets used in previous studies (DICM, MEF, NPE, LIME, and VV) as well as the HP-NCR dataset.
According to Table 2, the enhanced images produced by Squeeze-EnGAN demonstrated higher quality compared to the original images across all image sets. While EnlightenGAN achieved high quality in terms of the NIQE metric due to its large number of parameters, Squeeze-EnGAN maintained a comparable performance despite being lightweight. The NIQE score difference between the two models was 0.08 for the VV test set with the smallest gap, while it reached 0.24 for the DICM test set with the largest difference. These results demonstrate that Squeeze-EnGAN achieves competitive performance in enhancing image quality under various test conditions.
Additionally, the PSNR (Peak Signal-to-noise ratio) and SSIM (Structural Similarity Index Measure) [50] for the LOL dataset were evaluated and summarized in Table 3. A higher PSNR value indicates better reconstruction quality with lower distortion. Meanwhile, SSIM ranges from 0 to 1, with values closer to 1 indicating higher structural similarity to the reference image. Although our model achieves lower scores than supervised learning models in both metrics, Figure 5 shows that it produces visually high-quality results. Furthermore, our model achieved a higher SSIM score than the EnlightenGAN, even with fewer parameters.
For the HP-NCR dataset, we compared the results using the Squeeze-EnGAN-R model, which was trained to capture the characteristics of road scenes. As a result, the Squeeze-EnGAN-R model produced images with less noise and blur than both the original and other models. As shown in Figure 6, EnlightenGAN introduces artifacts and generates unnatural images for driving scenes. Other models also lead to unnatural color distortions and excessive blurring caused by over-enhancement of brightness. On the other hand, the results of Squeeze-EnGAN-R exhibit more natural colors, enabling the detection model to identify cars more effectively.
High-level vision tasks, such as identifying surrounding objects, play a critical role in camera-based perception modules for autonomous vehicles and intelligent mobility systems. To verify whether our low-light enhancement process contributes to object identification tasks, we conducted an effectiveness evaluation. Specifically, we compared the detection performance of the original images from the HP-NCR dataset with the enhanced outcomes. The 554 test images from the HP-NCR dataset contained a total of 823 vehicles, and we detected these vehicles using YOLO11 [51], a well-known object detection model. From the YOLO11 models available in different parameter sizes, we selected and utilized the small (s) and medium (m) models. The detection performance was evaluated using Precision, Recall, and mAP, with the results presented in the accompanying Table 4. These findings demonstrate the impact of low-light enhancement on improving object detection in challenging nighttime conditions.
According to Table 4, the detection performance improved across all metrics in the results of both EnlightenGAN and Squeeze-EnGAN-R compared to the original images. Notably, the results of Squeeze-EnGAN-R demonstrated better detection performance than those of EnlightenGAN across most metrics. In particular, the outcomes of Squeeze-EnGAN-R showed better performance across all metrics when using the YOLO11-s model. Even when compared to the original Squeeze-EnGAN model, Squeeze-EnGAN-R achieved higher performance in most metrics. This indicates that Squeeze-EnGAN-R adapts more effectively to driving environments and suggests that using diverse datasets in Squeeze-EnGAN can improve its performance in specific scenarios. In particular, unsupervised learning models are well suited for applications such as continual learning frameworks [52], where model weights are dynamically updated in real time to adapt to diverse driving environments.

4.4. Performance on Edge System

Most mobility systems face constraints in power supply and space, making it impractical to use computers equipped with high-performance GPUs. Additionally, installing expensive computing equipment in vehicles poses challenges in mass production due to cost concerns. With the recent advancements in compact onboard computers, GPU-equipped single-board computers are increasingly used in research for autonomous driving and robotics control to handle AI computations.
To demonstrate the utility of our model in mobility edge system environments, we measured inference times on NVIDIA’s GPU-equipped single-board computers, namely Jetson Xavier AGX (32 GB Developer Kit) and Jetson Nano (2 GB Developer Kit). The models were executed in a PyTorch 1.10 environment supported by the JetPack SDK. As shown in Table 5, Squeeze-EnGAN achieved faster inference speed than LLFlow, CIDNet, URetinex-Net, PairLIE, and EnlightenGAN on both devices. Additionally, due to memory limitations, LLFlow could not perform stable inference on the Jetson Nano. Notably, the Squeeze-EnGAN model achieves a 30.68% reduction in total inference time on the Jetson Xavier compared to EnlightenGAN. On the other hand, Squeeze-EnGAN demonstrated a 10% faster inference time on the Jetson Nano. During the model execution, the GPU memory usage measured on the Jetson Nano was 721 MB for Squeeze-EnGAN, while EnlightenGAN consumed 934 MB, which is 29.55% higher. In addition to these evaluations, the lower MACs observed in Table 1 further indicate a reduction in the model’s computational complexity. This provides insights into the model’s feasibility for deployment in resource-constrained, real-world mobility applications.

4.5. Ablation Study

We conducted an ablation study to investigate the effects of each component of Squeeze-EnGAN by performing new training sessions with certain components removed or modified. Specifically, experiments were conducted under the following four conditions:
  • To investigate the effect of the local discriminator, we conducted training without it.
  • All layers of the generator network shown in Figure 2, except for the final convolution layer, were replaced with fire modules.
  • Fire modules in the decoder were removed and replaced with standard convolution layers. Therefore, fire modules are used only in the encoder.
  • Fire modules in the encoder were removed and replaced with standard convolution layers. Therefore, fire modules are used only in the decoder.
The training dataset was not modified, and the experiments were conducted under the same conditions. The result images from each experiment are presented in Figure 7.
The results in Figure 7c demonstrate that excluding the local discriminator leads to a significant degradation in the quality of the output images. Although the brightness of the images has been enhanced, blurring and artifacts can be observed in all three images. In the case of Figure 7d, the image in the first row shows large artifacts on the structure. This indicates that using the fire module for the input image fails to extract sufficient features. Similarly, blurring and artifacts can be observed in the images of Figure 7e. This suggests that despite the high representational capacity of the decoder, significant information loss occurs during the encoding process. This results in suboptimal reconstruction, likely due to insufficient preservation of fine-grained spatial details in the encoder.
In contrast, Figure 7f exhibit realistic colors and visually high-quality images. This suggests that the rich information extracted by the encoder can be effectively processed by the fire module. However, as shown in Table 1, it requires approximately 3.9 M additional parameters compared to Squeeze-EnGAN. This indicates that the total number of parameters is more than approximately 3.8 times that of Squeeze-EnGAN. Overall, as shown in Figure 7b, Squeeze-EnGAN can generate high-quality images with a limited number of parameters. This supports the idea that our model achieves high performance while remaining lightweight by appropriately utilizing the local discriminator and strategically placing the fire modules.

5. Conclusions

This paper addresses methods to overcome the performance degradation of RGB cameras in low-light environments and enhance the perceptual capabilities of intelligent mobility systems. We focused on an unsupervised learning-based low-light image enhancement method using unpaired data. Furthermore, we proposed a lightweight deep learning model considering real-time applicability in driving environments. The model is designed as a GAN-based generator-discriminator network, enabling training without paired image data. To achieve model optimization, we incorporated fire modules into the U-net structure of the generator, significantly reducing the number of parameters and computational complexity.
Despite the lightweight design, the enhanced images produced by the model demonstrated visually superior quality compared to the original images and achieved competitive scores on the NIQE metric. Furthermore, our experiments confirmed an improvement in object detection rates in real driving scenes, demonstrating the model’s effectiveness for intelligent mobility applications. Notably, Squeeze-EnGAN achieved an inference time of 0.061 s on a compact onboard PC, delivering real-time performance while maintaining results comparable to those of existing studies. These findings suggest that the proposed method is well-suited for mobility environments requiring rapid decision-making to avoid accidents and provides a critical solution for enhancing the safety of future driving technologies. Nevertheless, to further optimize the system for autonomous driving applications, the overall framework must be refined. For instance, research should be conducted on middleware systems to ensure stable frame acquisition from cameras, process scheduling, and hardware accelerators to enhance computational efficiency.
Squeeze-EnGAN leverages an unsupervised learning approach, offering the significant advantage of training with unpaired image sets. However, the use of a GAN-based architecture presents challenges in balancing the generator and discriminator. Additionally, the model’s performance is highly dependent on the quality of the training dataset. In this study, we utilized the HP-NCR dataset, which contains various nighttime driving scenes and is publicly available through AI-Hub. However, noise and aliasing were observed in the source images, potentially impacting performance. To further enhance the model’s capabilities, it will be necessary to curate high-quality training datasets that effectively represent diverse driving scenarios or apply appropriate preprocessing steps to the data prior to model input. These improvements could significantly bolster the robustness and reliability of the proposed method.
Future research should focus on identifying optimal datasets while also exploring methods to further accelerate inference speed. Since intelligent driving systems are responsible for transporting people and goods, they must completely mitigate the risk of accidents. Enhancing low-light images should not become a double-edged sword that wastes resources and overloads the system. Achieving robust performance across diverse environmental conditions while minimizing resource consumption will significantly enhance the reliability of future driving technologies. These advancements are critical for ensuring both the efficiency and safety of intelligent mobility systems.

Author Contributions

Conceptualization, H.I. and C.M.; methodology, H.I.; software, H.I.; validation, H.I.; formal analysis, H.I. and J.K.; investigation, H.I.; resources, C.M.; data curation, H.I.; writing—original draft preparation, H.I.; writing—review and editing, H.I., J.K., and C.M.; visualization, H.I.; supervision, J.K. and C.M.; project administration, H.I.; funding acquisition, C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported by Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTIE) (P0020536, HRD Program for Industrial Innovation).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are available upon request to the authors.

Acknowledgments

This paper used datasets from ‘The Open AI Dataset Project (AI-Hub, S. Korea)’. All data information can be accessed through ‘AI-Hub, www.aihub.or.kr (accessed on 12 March 2025)’.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

This section describes the detailed architecture of the generator. Table A1 shows the layer specifications for the encoder part of the generator, while Table A2 contains the layer specifications for the decoder part of the generator. After every convolution operation, including those within the fire modules, LeakyReLU and batch normalization layers are applied. However, batch normalization is not applied to the single convolution operation or single fire module that follows an upsampling step. The final output is combined with the input and clamped to the range of −1.0 to 1.0.
Table A1. The layer specifications for the encoder part of the generator.
Table A1. The layer specifications for the encoder part of the generator.
LayerCinCmidCoutKernel Size
Convolution 2D3-323 × 3
Convolution 2D32-323 × 3
MaxPooling---2 × 2
FireModule3232643 × 3
FireModule6432643 × 3
MaxPooling---2 × 2
FireModule64481283 × 3
FireModule128481283 × 3
MaxPooling---2 × 2
FireModule128642563 × 3
FireModule256642563×3
MaxPooling---2×2
FireModule256805123×3
FireModule512805123×3
Table A2. The layer specifications for the decoder part of the generator.
Table A2. The layer specifications for the decoder part of the generator.
LayerCinCmidCoutKernel Size
Upsampling----
FireModule512802563 × 3
FireModule512642563 × 3
FireModule256642563 × 3
Upsampling----
FireModule256641283 × 3
FireModule256481283 × 3
FireModule128481283 × 3
Upsampling----
FireModule12848643 × 3
FireModule12832643 × 3
FireModule6432643 × 3
Upsampling----
Convolution 2D64-323 × 3
Convolution 2D64-323 × 3
Convolution 2D32-33 × 3

References

  1. Gruyer, D.; Magnier, V.; Hamdi, K.; Claussmann, L.; Orfila, O.; Rakotonirainy, A. Perception, information processing and modeling: Critical stages for autonomous driving applications. Annu. Rev. Control 2017, 44, 323–341. [Google Scholar] [CrossRef]
  2. Wang, P. Research on comparison of lidar and camera in autonomous driving. J. Phys. Conf. Ser. 2021, 2093, 012032. [Google Scholar] [CrossRef]
  3. Janai, J.; Güney, F.; Behl, A.; Geiger, A. Computer vision for autonomous vehicles: Problems, datasets and state of the art. Found. Trends Comput. Graph. Vis. 2020, 12, 1–308. [Google Scholar] [CrossRef]
  4. Song, H.; Cho, J.; Ha, J.; Park, J.; Jo, K. Panoptic-FusionNet: Camera-LiDAR fusion-based point cloud panoptic segmentation for autonomous driving. Expert Syst. Appl. 2024, 251, 123950. [Google Scholar] [CrossRef]
  5. Yeong, D.J.; Velasco-Hernandez, G.; Barry, J.; Walsh, J. Sensor and sensor fusion technology in autonomous vehicles: A review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef]
  6. Xiao, Y.; Jiang, A.; Ye, J.; Wang, M.W. Making of night vision: Object detection under low-illumination. IEEE Access 2020, 8, 123075–123086. [Google Scholar] [CrossRef]
  7. Muhammad, K.; Ullah, A.; Lloret, J.; Del Ser, J.; de Albuquerque, V.H.C. Deep learning for safe autonomous driving: Current challenges and future directions. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4316–4336. [Google Scholar] [CrossRef]
  8. Liu, S.; Liu, L.; Tang, J.; Yu, B.; Wang, Y.; Shi, W. Edge computing for autonomous driving: Opportunities and challenges. Proc. IEEE 2019, 107, 1697–1716. [Google Scholar] [CrossRef]
  9. Liu, L.; Lu, S.; Zhong, R.; Wu, B.; Yao, Y.; Zhang, Q.; Shi, W. Computing systems for autonomous driving: State of the art and challenges. IEEE Internet Things J. 2020, 8, 6469–6486. [Google Scholar] [CrossRef]
  10. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Wang, Z. EnlightenGAN: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  11. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial nets. arXiv 2014, arXiv:1406.2661. [Google Scholar]
  12. Iandola, F.N. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2015), Munich, Germany, 5–9 October 2015. [Google Scholar]
  14. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  15. Land, E.H.; McCann, J.J. Lightness and retinex theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef]
  16. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  17. Radford, A. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  18. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  19. Chen, Z.; Zeng, Z.; Shen, H.; Zheng, X.; Dai, P.; Ouyang, P. DN-GAN: Denoising generative adversarial networks for speckle noise reduction in optical coherence tomography images. Biomed. Signal Process. Control 2020, 55, 101632. [Google Scholar] [CrossRef]
  20. Li, R.; Pan, J.; Li, Z.; Tang, J. Single Image Dehazing via Conditional Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8202–8211. [Google Scholar]
  21. Yuan, Y.; Liu, S.; Zhang, J.; Zhang, Y.; Dong, C.; Lin, L. Unsupervised Image Super-Resolution Using Cycle-in-Cycle Generative Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 701–710. [Google Scholar]
  22. Chen, Y.S.; Wang, Y.C.; Kao, M.H.; Chuang, Y.Y. Deep Photo Enhancer: Unpaired Learning for Image Enhancement from Photographs with GANs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 6306–6314. [Google Scholar]
  23. Beheshti, N.; Johnsson, L. Squeeze U-Net: A Memory and Energy Efficient Image Segmentation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Virtual Event, 14–19 June 2020; pp. 364–365. [Google Scholar]
  24. Odena, A.; Dumoulin, V.; Olah, C. Deconvolution and checkerboard artifacts. Distill 2016, 1, e3. [Google Scholar] [CrossRef]
  25. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  26. Lim, J.S.; Astrid, M.; Yoon, H.J.; Lee, S.I. Small Object Detection Using Context and Attention. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 13–16 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 181–186. [Google Scholar]
  27. Jolicoeur-Martineau, A. The relativistic discriminator: A key element missing from standard GAN. arXiv 2018, arXiv:1807.00734. [Google Scholar]
  28. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Smolley, S.P. Least Squares Generative Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  29. Johnson, J.; Alahi, A.; Li, F.-F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the 14th European Conference on Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar]
  30. Simonyan, K. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  31. RichardWebster, B.; Anthony, S.E.; Scheirer, W.J. Psyphy: A psychophysics driven evaluation framework for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2280–2286. [Google Scholar] [CrossRef] [PubMed]
  32. AIHub. High-Precision Data Collection Vehicle Night City Road Data. Available online: https://www.aihub.or.kr/aihubdata/data/view.do?dataSetSn=71580 (accessed on 13 January 2025).
  33. AIHub. High-Precision Data Collection Vehicle Daytime City Road Data. Available online: https://www.aihub.or.kr/aihubdata/data/view.do?dataSetSn=71577 (accessed on 13 January 2025).
  34. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  35. Dang-Nguyen, D.T.; Pasquini, C.; Conotter, V.; Boato, G. RAISE: A Raw Images Dataset for Digital Image Forensics. In Proceedings of the 6th ACM Multimedia Systems Conference, Portland, OR, USA, 18–20 March 2015; pp. 219–224. [Google Scholar]
  36. Kalantari, N.K.; Ramamoorthi, R. Deep high dynamic range imaging of dynamic scenes. ACM Trans. Graph. 2017, 36, 144. [Google Scholar] [CrossRef]
  37. Cai, J.; Gu, S.; Zhang, L. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef]
  38. Lee, C.; Lee, C.; Kim, C.S. Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef]
  39. Ma, K.; Zeng, K.; Wang, Z. Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef]
  40. Wang, S.; Zheng, J.; Hu, H.M.; Li, B. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef]
  41. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef]
  42. Vonikakis, V. Busting image enhancement and tone-mapping algorithms. Available online: https://sites.google.com/site/vonikakis/datasets (accessed on 13 January 2025).
  43. Kingma, D.P. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  44. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. URetinex-Net: Retinex-Based Deep Unfolding Network for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 5901–5910. [Google Scholar]
  45. Yan, Q.; Feng, Y.; Zhang, C.; Wang, P.; Wu, P.; Dong, W.; Zhang, Y. You only need one color space: An efficient network for low-light image enhancement. arXiv 2024, arXiv:2402.05809. [Google Scholar]
  46. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.P.; Kot, A. Low-Light Image Enhancement with Normalizing Flow. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 22 February–1 March 2022; Volume 36, No. 3. pp. 2604–2612. [Google Scholar]
  47. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Event, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  48. Fu, Z.; Yang, Y.; Tu, X.; Huang, Y.; Ding, X.; Ma, K.K. Learning a Simple Low-Light Image Enhancer from Paired Low-Light Instances. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 22252–22261. [Google Scholar]
  49. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  50. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  51. Jocher, G.; Qiu, J. Ultralytics YOLO11. Available online: https://github.com/ultralytics/ultralytics (accessed on 13 January 2025).
  52. Shaheen, K.; Hanif, M.A.; Hasan, O.; Shafique, M. Continual learning for real-world autonomous systems: Algorithms, challenges and frameworks. J. Intell. Robot. Syst. 2022, 105, 9. [Google Scholar] [CrossRef]
Figure 1. (a) represents the input and output of a conventional convolution layer with a kernel size of 3 × 3; (b) represents the structure of the fire module, which can replace the 3 × 3 convolution layer.
Figure 1. (a) represents the input and output of a conventional convolution layer with a kernel size of 3 × 3; (b) represents the structure of the fire module, which can replace the 3 × 3 convolution layer.
Sensors 25 01825 g001
Figure 2. The generator network of Squeeze-EnGAN. The contracting path consists of four downsampling units (DS), while the expansive path comprises four upsampling units (US). The features extracted in the contracting path are concatenated with the corresponding layers in the expansive path via skip connections.
Figure 2. The generator network of Squeeze-EnGAN. The contracting path consists of four downsampling units (DS), while the expansive path comprises four upsampling units (US). The features extracted in the contracting path are concatenated with the corresponding layers in the expansive path via skip connections.
Sensors 25 01825 g002
Figure 3. The structure of the global and local discriminator networks. The local discriminator uses randomly cropped regions from the input image of the global discriminator as its input.
Figure 3. The structure of the global and local discriminator networks. The local discriminator uses randomly cropped regions from the input image of the global discriminator as its input.
Sensors 25 01825 g003
Figure 4. Visualization of comparisons between the original images and the outcomes of the models for the LIME, DICM, MEF, NPE, and VV image sets. (a) Input. (b) URetinex-Net. (c) ZeroDCE. (d) PairLIE. (e) EnlightenGAN. (f) Squeeze-EnGAN (Ours).
Figure 4. Visualization of comparisons between the original images and the outcomes of the models for the LIME, DICM, MEF, NPE, and VV image sets. (a) Input. (b) URetinex-Net. (c) ZeroDCE. (d) PairLIE. (e) EnlightenGAN. (f) Squeeze-EnGAN (Ours).
Sensors 25 01825 g004aSensors 25 01825 g004b
Figure 5. Visualization of comparisons between the original images and the outcomes of the models for the LOL dataset. (a) input. (b) Ground Truth. (c) CIDNet. (d) LLFlow. (e) EnlightenGAN. (f) Squeeze-EnGAN (Ours).
Figure 5. Visualization of comparisons between the original images and the outcomes of the models for the LOL dataset. (a) input. (b) Ground Truth. (c) CIDNet. (d) LLFlow. (e) EnlightenGAN. (f) Squeeze-EnGAN (Ours).
Sensors 25 01825 g005
Figure 6. Visualization of comparisons between the original images and the outcomes of the models for the HP-NCR image sets. (a) Input. (b) URetinexNet. (c) LLFlow. (d) EnlightenGAN. (e) Squeeze-EnGAN. (f) Squeeze-EnGAN-R.
Figure 6. Visualization of comparisons between the original images and the outcomes of the models for the HP-NCR image sets. (a) Input. (b) URetinexNet. (c) LLFlow. (d) EnlightenGAN. (e) Squeeze-EnGAN. (f) Squeeze-EnGAN-R.
Sensors 25 01825 g006aSensors 25 01825 g006b
Figure 7. Comparison of result images for the original, Squeeze-EnGAN, and ablation experiments. Each column represents the results for the following: (a) input, (b) Squeeze-EnGAN, (c) training without the local discriminator, (d) replacing all generator layers with fire modules except the final convolution layer, (e) using fire modules only in the encoder, and (f) using fire modules only in the decoder.
Figure 7. Comparison of result images for the original, Squeeze-EnGAN, and ablation experiments. Each column represents the results for the following: (a) input, (b) Squeeze-EnGAN, (c) training without the local discriminator, (d) replacing all generator layers with fire modules except the final convolution layer, (e) using fire modules only in the encoder, and (f) using fire modules only in the decoder.
Sensors 25 01825 g007aSensors 25 01825 g007b
Table 1. Comparison of #MACs and the number of parameters between the generator with the fire module (ours) and the generator of EnlightenGAN. The input image size for each network is the same at 640 × 400.
Table 1. Comparison of #MACs and the number of parameters between the generator with the fire module (ours) and the generator of EnlightenGAN. The input image size for each network is the same at 640 × 400.
Layer NameSqueeze-EnGAN (Ours)EnlightenGAN *
Components#MACs (M)#ParamsComponents#MACs (M)#Params
Conv1Conv2D × 22646.01610272Conv2D × 22719.74410560
DS1Maxpooling × 1
Fire module × 2
1556.48024128Maxpooling × 1
Conv2D × 2
3571.71255680
DS2Maxpooling × 1
Fire module × 2
1153.02471712Maxpooling × 1
Conv2D × 2
3555.328221952
DS3Maxpooling × 1
Fire module × 2
763.904190336Maxpooling × 1
Conv2D × 2
3547.136886272
DS4Maxpooling × 1
Fire module × 2
475.776474592Maxpooling × 1
Conv2D × 2
3543.0403542016
US1Upsampling × 1
Fire module × 3
1459.520358608Upsampling × 1
Conv2D × 3
11828.2242950912
US2Upsampling × 1
Fire module × 3
2266.112138464Upsampling × 1
Conv2D × 3
11859.968738176
US3Upsampling × 1
Fire module × 3
3226.62448816Upsampling × 1
Conv2D × 3
11923.456184768
US4Upsampling × 1
Conv2D × 3
12034.04846240Upsampling × 1
Conv2D × 3
12034.04846240
Conv2Conv2D × 125.34499Conv2D × 125.34499
Total 25606.8481363267 64608.0008636675
* The generator in EnlightenGAN incorporates an attention module.
Table 2. Comparison of NIQE metrics for different models on the DICM, MEF, NPE, LIME, VV, and HP-NCR dataset.
Table 2. Comparison of NIQE metrics for different models on the DICM, MEF, NPE, LIME, VV, and HP-NCR dataset.
MethodDataset
DICMMEFNPELIMEVVHP-NCR
Input4.254.274.324.353.527.02
URetinex-Net4.203.794.694.343.036.34
CIDNet3.793.563.744.133.215.82
LLFlow3.633.464.093.983.015.69
ZeroDCE4.584.934.535.824.815.95
PairLIE4.094.184.214.513.665.91
EnlightenGAN3.503.234.113.722.585.48
Ours3.743.374.293.892.665.68
Table 3. Comparison of PSNR and SSIM metrics for different models for the LOL dataset. SL means supervised learning, while UL means unsupervised learning.
Table 3. Comparison of PSNR and SSIM metrics for different models for the LOL dataset. SL means supervised learning, while UL means unsupervised learning.
MethodLOL Dataset
PSNRSSIM
SLURetinex-Net21.3280.835
CIDNet23.8090.857
LLFlow21.1490.854
ULPairLIE19.5100.736
ZeroDCE14.8610.559
EnlightenGAN17.4800.651
Ours16.1740.658
Table 4. Vehicle detection performance on the HP-NCR dataset and enhanced outcomes measured using YOLO11.
Table 4. Vehicle detection performance on the HP-NCR dataset and enhanced outcomes measured using YOLO11.
MetricMethod
OriginalEnlightenGANSqueeze-EnGANSqueeze-EnGAN-R
YOLO11-sPrecision0.3640.5360.3770.537
Recall0.3690.3290.4360.375
mAP500.3350.3720.3870.399
mAP50-950.1830.1920.1950.208
YOLO11-mPrecision0.6220.6450.7100.623
Recall0.3940.4180.4100.453
mAP500.4500.4850.4900.497
mAP50-950.2640.2820.2880.290
Table 5. Comparison of inference times for each model on two devices with a 600 × 400 image input.
Table 5. Comparison of inference times for each model on two devices with a 600 × 400 image input.
MethodInference Time (s)
Jetson XavierJetson Nano
LLFlow1.612-
CIDNet0.2411.760
URetinex-Net0.3713.332
PairLIE0.1571.350
EnlightenGAN0.0880.651
Squeeze-EnGAN (ours)0.0610.590
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

In, H.; Kweon, J.; Moon, C. Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles. Sensors 2025, 25, 1825. https://doi.org/10.3390/s25061825

AMA Style

In H, Kweon J, Moon C. Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles. Sensors. 2025; 25(6):1825. https://doi.org/10.3390/s25061825

Chicago/Turabian Style

In, Haegyo, Juhum Kweon, and Changjoo Moon. 2025. "Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles" Sensors 25, no. 6: 1825. https://doi.org/10.3390/s25061825

APA Style

In, H., Kweon, J., & Moon, C. (2025). Squeeze-EnGAN: Memory Efficient and Unsupervised Low-Light Image Enhancement for Intelligent Vehicles. Sensors, 25(6), 1825. https://doi.org/10.3390/s25061825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop