Inspired by recent advancements in conditional diffusion models and cold diffusion models, we introduces a new noise modeling approach, EL2NM, building upon the work of Kristina Monakhova. EL2NM employs a series of refinement steps to convert complex noise distributions into empirical data distributions, akin to Langevin dynamics. At its core lies the U-Net architecture, utilized to train the noise model and iteratively produce noise outputs. The U-Net architecture, adapted from SR3, is modified in this work to accommodate conditional image generation. The sampling technique of the cold diffusion model is applied iteratively to generate the final noise image.
Our main contributions are summarized as follows:
Comparison
This study compared the proposed noise model with previous works, with each row representing a different noise modeling method. The experiments demonstrated that the proposed method is capable of generating noise images similar to the Starlight method, yielding higher quality noise images.
This study compared commonly used noise modeling methods and demonstrated the noise images generated by the proposed method and baseline methods.
We found that our method EL2NM is more stable than the starlight and can generate realistic noisy image in low light environments.
We compared the baseline Starlight model based on GAN networks, the non-deep low-light noise model ELD, as well as two deep learning-based noise models, CA-GAN and Noise Flow.
Both Noise Flow and CA-GAN miss the significant banding noise present in real noisy clips. ELD miss the quantizaion noise.
The EL2NM method exhibited good performance on this dataset.
Qualitative performance indicators are presented in Table, indicating that KL divergence computed by EL2NM was comparable to the baseline.
Ablation Study
We compared the method with only conditional diffusion model and the method with only cold diffusion model with our method, and the results show that our method can better establish the noise model.
We can see that when only the conditional diffusion model is used, the image does not recover noise information and some image information is lost. When there is only a cold diffusion model, the noise in image restoration is not complete enough, and the visual effect quality is not high. When combined, it can generate noise images in weak light environments more completely while ensuring the stability of the generation.
@InProceedings{Qin_2024_CVPR,
author = {Qin, Jiahao and Qin, Pinle and Chai, Rui and Qin, Jia and Jin, Zanxia},
title = {EL2NM: Extremely Low-light Noise Modeling Through Diffusion Iteration},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2024},
pages = {1085-1094}
}