PairLIE论文阅读笔记

PairLIE论文阅读笔记

论文为2023CVPR的Learning a Simple Low-light Image Enhancer from Paired Low-light Instances.论文链接如下:

openaccess.thecvf.com/content/CVPR2023/papers/Fu_Learning_a_Simple_Low-Light_Image_Enhancer_From_Paired_Low-Light_Instances_CVPR_2023_paper.pdf

文章目录

  • PairLIE论文阅读笔记
    • 出发点
    • 创新点
    • 模型
    • 设计及其损失

出发点

1.However, collecting high-quality reference maps in real-world scenarios is time-consuming and expensive.

出发点1:在低光照领域,从现实世界中获取高质量的参考照片进行监督学习,既费时又困难,成本昂贵。

因为获得低光环境的照片是容易的,而此低光照片对应的亮度较大的参考图片是难得的。

2.To tackle the issues of limited information in a single low-light image and the poor adaptability of handcrafted priors, we propose to leverage paired low-light instances to train the LIE network.

Additionally, twice-exposure images provide useful information for solving the LIE task. As a result, our solution can reduce the demand for handcrafted priors and improve the adaptability of the network.

出发点2:为了解决手动设置的先验的低适应性,减少手动设置先验的需求,同时提升模型对陌生环境的适应性。

创新点

The core insight of our approach is to sufficiently exploit priors from paired low-light images.

Those low-light image pairs share the same scene content but different illumination. Mathematically, Retinex decomposition with low-light image pairs can be expressed as:

在这里插入图片描述

创新点1:作者利用两张低光图片进行训练,以充分提取低光图片的信息。

instead of directly imposing the Retinex decomposition on original low-light images, we adopt a simple self-supervised mechanism to remove inappropriate features and implement the Retinex decomposition on the optimized image.

创新点2:作者基于Retinex理论,但是并不循旧地直接运用Retinex的分解。作者采用一个简单的自监督机制以实现不合理特征的去除(通常是一些噪音)以及更好地实现Retinex理论。

模型

PairLIE论文阅读笔记_第1张图片

将两张同一场景不同曝光的低光图片送入训练中,图片I1与I2先经过P-Net去除噪音,得到i1与i2,然后利用L-Net与R-Net分解为照度L1与反射R1(对应有L2与R2)。

在测试,只需要输入一张低光照图片I,经过P-Net的噪音去除,得到i,然后用L-Net与R-Net分解为照度和反射,然后对照度L进行增强,操作为g(L),把增强结果与反射R进行元素乘法,得到增强后的图片Enhanced Image。

设计及其损失

Note that, this paper does not focus on designing modernistic network structures. L-Net and R-Net are very similar and simple,

1.模型使用的L-Net与R-Net十分简单。整体架构只是单纯的卷积神经网络。

Apart from L-Net and R-Net, we introduce P-Net to remove inappropriate features from the original image. Specifically, the structure of the P-Net is identical to the R-Net.

2,P-Net被设计用于去除不合理特征。
L p = ∣ ∣ I 1 − i 1 ∣ ∣ 2 2 L_p = \mid\mid I_1 - i_1 \mid\mid^2_2 Lp=∣∣I1i122

Note that the projection loss needs to cooperate with the other constraints to avoid a trivial solution.i,e.,i1 = I1.

3.Projection Loss:最大程度限制去除不合理特征后的i1和原始低光图片I1的区别。

这个损失需要避免一个特例,即降噪后图片与原图相同,即未降噪。
L c = ∣ ∣ R 1 − R 2 ∣ ∣ 2 2 (1) L_c = \mid\mid R_1 - R_2 \mid\mid^2_2\tag{1} Lc=∣∣R1R222(1)

Since sensor noise hidden in dark regions will be amplified when the contrast is improved.

In our method, the sensor noise can be implicitly removed by Eq. 1.

4.Reflection Loss:通常用传感或摄影设备拍摄低光场景照片会携带一定的设备噪音,这个损失最大限度保证两张图片的反射是相同的,减少传感或摄影设备的影响,这是因为图片场景的内容相同。

这个损失是确保反射的一致性。
L R = ∣ ∣ R ∘ L − i ∣ ∣ 2 2 + ∣ ∣ R − i / s t o p g r a d ( L ) ∣ ∣ 2 2 + ∣ ∣ L − L 0 ∣ ∣ 2 2 + ∣ ∣ ∇ L ∣ ∣ 1 L_R = \mid\mid R \circ L - i \mid\mid^2_2 + \mid\mid R - i / stopgrad(L)\mid\mid^2_2 + \mid\mid L - L_0 \mid\mid^2_2 + \mid\mid \nabla L \mid\mid_1 LR=∣∣RLi22+∣∣Ri/stopgrad(L)22+∣∣LL022+∣∣L1

∣ ∣ R ∘ L − i ∣ ∣ 2 2 \mid\mid R \circ L - i \mid\mid^2_2 ∣∣RLi22 is applied to ensure a reasonable decomposition.

∣ ∣ R − i / s t o p g r a d ( L ) ∣ ∣ 2 2 \mid\mid R - i / stopgrad(L) \mid\mid^2_2 ∣∣Ri/stopgrad(L)22 is to guide the decomposition.

Specifically, the initialized illumination L0 is calculated via the maximum of the R, G, and B channels: L 0 = m a x c ∈ R , G , B I c ( x ) . L_0 = \underset{c \in{R, G, B}}{max} I^c(x). L0=cR,G,BmaxIc(x).

5.Retinex Loss:Retinex损失是为了限制分解组块L-Net和R-Net以满足Retinex的理论要求。

本文毕

你可能感兴趣的:(低光增强,论文阅读,笔记)