LEDNet: Joint Low-light Enhancement and Deblurring in the Dark


Shangchen Zhou           Chongyi Li           Chen Change Loy
S-Lab, Nanyang Technological University

Visual results on real-world night blurry images (MouseOut: input, MouseOver: result)



Visual results on real-world night blurry videos

Materials




Paper


Dataset


Codes

Abstract


Night photography typically suffers from both low light and blurring issues due to the dim environment and the common use of long exposure. While existing light enhancement and deblurring methods could deal with each problem individually, a cascade of such methods cannot work harmoniously to cope well with joint degradation of visibility and textures. Training an end-to-end network is also infeasible as no paired data is available to characterize the coexistence of low light and blurs. We address the problem by introducing a novel data synthesis pipeline that models realistic low-light blurring degradations. With the pipeline, we present the first large-scale dataset for joint low-light enhancement and deblurring. The dataset, LOL-Blur, contains 12,000 low-blur/normal-sharp pairs with diverse darkness and motion blurs in different scenarios. We further present an effective network, named LEDNet, to perform joint low-light enhancement and deblurring. Our network is unique as it is specially designed to consider the synergy between the two inter-connected tasks. Both the proposed dataset and network provide a foundation for this challenging joint task. Extensive experiments demonstrate the effectiveness of our method on both synthetic and real-world datasets.

Dataset

Examples in LOL-Blur dataset, showing diverse darkness and motion blurs in dark dynamic scenes. The dataset also provides realistic blurs in the simulated regions, like light streaks shown in the last two columns.

We provide a data synthesis pipeline that models low-light blur degradation realistically, hence allowing us to generate a large-scale dataset (LOL-Blur) data for this joint task. The pipeline contains five main steps, as shown in the above figure. Based on this synthesis pipeline, we construct a large-scale dataset LOL-Blur, consisting of a total of 10,200 pairs {x,y} for training and 1,800 pairs {x,y} for testing.

Method

Overview of our proposed LEDNet.

It contains an Encoder for Light Enhancement, LE-Encoder, and a Decoder for Deblurring, D-Decoder. They are connected by three Filter Adaptive Skip Connections (FASC). The PPM and the proposed CurveNLU layers are inserted in LE-Encoder, making light enhancement more stable and powerful. LEDNet applies spatially-adaptive transformation to D-Decoder using filters generated by FASC from enhanced features. CurveNLU and FASC enable LEDNet to perform spatially-varying feature transformation for both intensity enhancement and blur removal.

Results


Visual comparison on synthetic image.


Visual comparison on real-world night blurry image.

Citation

If you find our dataset and paper useful for your research, please consider citing our work:

@article{zhou2022lednet,
          author = {Zhou, Shangchen and Li, Chongyi and Loy, Chen Change},
          title = {LEDNet: Joint Low-light Enhancement and Deblurring in the Dark},
          journal={arXiv preprint arXiv:2202.03373},
          year = {2022}
          }

Contact

If you have any question, please contact us via shangchenzhou@gmail.com.



visitors