TY - JOUR
T1 - Multi-Attention Guided SKFHDRNet For HDR Video Reconstruction
AU - Ullah, Ehsan
AU - Pedersen, Marius
AU - Waaseth, Kjartan Sebastian
AU - Baltzersen, Bernt Erik
N1 - Publisher Copyright:
© 2023 Society for Imaging Science and Technology.
PY - 2023/9
Y1 - 2023/9
N2 - We propose a three stage learning-based approach for High Dynamic Range (HDR) video reconstruction with alternating exposures. The first stage performs alignment of neighboring frames to the reference frame by estimating the flows between them, the second stage is composed of multi-attention modules and a pyramid cascading deformable alignment module to refine aligned features, and the final stage merges and estimates the final HDR scene using a series of dilated selective kernel fusion residual dense blocks (DSKFRDBs) to fill the over-exposed regions with details. The proposed model variants give HDR-VDP-2 values on a dynamic dataset of 79.12, 78.49, and 78.89 respectively, compared to Chen et al. [“HDR video reconstruction: A coarse-to-fine network and a real-world benchmark dataset, ” Proc. IEEE/CVF Int'l. Conf. on Computer Vision (IEEE, Piscataway, NJ, 2021), pp. 2502-2511] 79.09, Yan et al. [“Attention-guided network for ghost-free high dynamic range imaging, ” Proc. IEEE/CVF Conf. on Computer Vision and Pattern Recognition (IEEE, Piscataway, NJ, 2019), pp. 1751-1760] 78.69, Kalantari et al. [“Patch-based high dynamic range video, ” ACM Trans. Graph. 32 (2013) 202-1] 70.36, and Kalantari et al. [“Deep hdr video from sequences with alternating exposures, ” Computer Graphics Forum (Wiley Online Library, 2019), Vol. 38, pp. 193-205] 77.91. We achieve better detail reproduction and alignment in over-exposed regions compared to state-of-the-art methods and with a smaller number of parameters.
AB - We propose a three stage learning-based approach for High Dynamic Range (HDR) video reconstruction with alternating exposures. The first stage performs alignment of neighboring frames to the reference frame by estimating the flows between them, the second stage is composed of multi-attention modules and a pyramid cascading deformable alignment module to refine aligned features, and the final stage merges and estimates the final HDR scene using a series of dilated selective kernel fusion residual dense blocks (DSKFRDBs) to fill the over-exposed regions with details. The proposed model variants give HDR-VDP-2 values on a dynamic dataset of 79.12, 78.49, and 78.89 respectively, compared to Chen et al. [“HDR video reconstruction: A coarse-to-fine network and a real-world benchmark dataset, ” Proc. IEEE/CVF Int'l. Conf. on Computer Vision (IEEE, Piscataway, NJ, 2021), pp. 2502-2511] 79.09, Yan et al. [“Attention-guided network for ghost-free high dynamic range imaging, ” Proc. IEEE/CVF Conf. on Computer Vision and Pattern Recognition (IEEE, Piscataway, NJ, 2019), pp. 1751-1760] 78.69, Kalantari et al. [“Patch-based high dynamic range video, ” ACM Trans. Graph. 32 (2013) 202-1] 70.36, and Kalantari et al. [“Deep hdr video from sequences with alternating exposures, ” Computer Graphics Forum (Wiley Online Library, 2019), Vol. 38, pp. 193-205] 77.91. We achieve better detail reproduction and alignment in over-exposed regions compared to state-of-the-art methods and with a smaller number of parameters.
UR - http://www.scopus.com/inward/record.url?scp=85176334957&partnerID=8YFLogxK
U2 - 10.2352/J.IMAGINGSCI.TECHNOL.2023.67.5.050409
DO - 10.2352/J.IMAGINGSCI.TECHNOL.2023.67.5.050409
M3 - Article
AN - SCOPUS:85176334957
SN - 1062-3701
VL - 67
JO - Journal of Imaging Science and Technology
JF - Journal of Imaging Science and Technology
IS - 5
ER -