Highlights

In brief

The DehazeGAN model is a generative adversarial network that can recover clean images from hazy ones.

© A*STAR Institute of High Performance Computing

Seeing clearly through the haze

2 Nov 2020

A promising new computational model can recreate clean images out of hazy ones.

If you’ve ever spent a day out in the haze, you might have noticed how difficult it is to see anything. But your human eyes aren’t the only ones that struggle in such conditions—digital vision sensors and computer vision algorithms take a hit too, with potentially serious implications for systems that rely on a clear vision, such as video surveillance cameras or autonomous vehicles.

Because the smoke or dust particles that make up haze create a kind of non-additive noise, hazy images can’t be resolved with just simple contrast enhancement methods. Instead, haze removal relies on the accurate estimation of two factors: global atmospheric light and the transmission map, which is the path of light that is not scattered.

Unlike most existing methods that estimate these two parameters separately, which reduces efficiency and accuracy, researchers led by Hongyuan Zhu, a Research Scientist at the A*STAR Institute for Infocomm Research (I2R) on secondment to the Institute of High Performance Computing (IHPC), have created a new model that applies a generative adversarial network (GAN) to single-image dehazing for the first time.

With their two-network architecture, GANs can be used to produce high-quality images in tasks such as image generation and object detection. The resulting model, aptly named DehazeGAN, has been shown to reliably recover clean images from hazy ones and outperform state-of-the-art methods.

“The noise created by haze is material and distance-dependent, according to the atmospheric scattering model. DehazeGAN is the first end-to-end method that solves the image 'dehazing' problem by embracing this model,” Zhu said.

DehazeGAN’s success lies in its two components: a novel compositional generator, which enables DehazeGAN to directly learn the physical parameters from data, and a novel deeply supervised discriminator, which ensures clean image output.

“Our method achieves superior performance in all metrics thanks to physical modeling and adversarial learning,” he shared. “Moreover, it models the recovery process as a highly efficient, fully convolutional neural network with real-time performance.”

According to the Zhu, DehazeGAN can be used to enhance the quality of vision sensors in autonomous vehicles or mobile phones, as well as the robustness and accuracy of existing computer vision systems under adverse weather conditions.

To test DehazeGAN’s performance, the researchers created the HazeCOCO dataset of synthesized haze images, which they have shared for use in other single-image 'dehazing' efforts.

“HazeCOCO is currently the largest haze dataset with various diverse visual patterns for learning discriminative 'dehazing' features, which can benefit further research in this field,” said Zhu.

The A*STAR-affiliated researchers contributing to this research are from the Institute for Infocomm Research (I2R) and Institute of High Performance Computing (IHPC).

Want to stay up to date with breakthroughs from A*STAR? Follow us on Twitter and LinkedIn!

References

Zhu, H. et al. Single-Image Dehazing via Compositional Adversarial Network. IEEE Transactions on Cybernetics 10.1109 (2019) | article

About the Researcher

View articles

Hongyuan Zhu

Senior Scientist and Unit Lead, Satellite Sensing

A*STAR Institute for Infocomm Research (A*STAR I2R)
Hongyuan Zhu is the Unit Head and PI of Satellite Sensing at the A*STAR Institute for Infocomm Research (A*STAR I2R). He leads the Advanced Perception and Reasoning Lab, focusing on developing autonomous agents with super-large-scale multimodal sensing and reasoning capabilities for sustainability, climate/weather and defence solutions. He was selected as a Top 2% Scientist by Stanford in 2023~2025 and received the A*STAR Career Award in 2022. His team achieved first place in the Scene2Cap challenge ICCV23 and third place in the EPIC Text-Video Retrieval Challenge in CVPR22, as well as being the only Asian team to achieve 1st Prize Finalist in the KUKA Innovation Challenge 2021. Zhu has been an associate editor of Visual Computer since 2020. He also served as the Senior Program Committee member of IJCAI, the Area Chair of ACM MM Asia, and the Guest Editor of IET Image Processing. Hs has published around 120 papers in top-tier journals and conferences, including CVPR, ICCV, NeurIPS, ICML, AAAI, IJCAI, ACL and TPAMI.

This article was made for A*STAR Research by Wildtype Media Group