DSFormer: Leveraging Transformer with Cross-Modal Attention for Temporal Consistency in Low-Light Video Enhancement

被引:0
|
作者
Xu, JiaHao [1 ,2 ]
Mei, ShuHao [2 ]
Chen, ZiZheng [2 ]
Zhang, DanNi [2 ]
Shi, Fan [1 ,2 ]
Zhao, Meng [1 ,2 ]
机构
[1] Tianjin Univ Technol, Minist Educ, Engn Res Ctr Learning Based Intelligent Syst, Tianjin 300384, Peoples R China
[2] Tianjin Univ Technol, Sch Comp Sci & Engn, Tianjin 300384, Peoples R China
基金
中国国家自然科学基金;
关键词
Low-Light Video Enhancement; Transformer; Optical flow;
D O I
10.1007/978-981-97-5612-4_3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in deep learning have significantly impacted low-light video enhancement, sparking great interest in the field. However, while these techniques have proven effective for enhancing individual static images, they struggle with temporal instability when applied to videos, leading to artifacts and flickering. This challenge is further compounded by the difficulty of obtaining dynamic low-light/high-light video pairs in real-world scenarios. Our proposed solution tackles these issues by integrating a cross-attention mechanism with optical flow. This approach helps mitigate temporal inconsistencies, often found when training with static images, by using optical flow to infer motion in individual frames. We have also developed a Transformer model (DSFormer) that leverages spatial and channel features to enhance visual quality and temporal stability in videos. Additionally, we have created a novel dual path feed-forward network (DPFN) that improves our method's ability to capture and maintain local contextual information, which is crucial for low-light enhancement. Through extensive comparative and ablation studies, we demonstrate that our approach delivers high luminance and temporal consistency in enhancement sequences.
引用
收藏
页码:27 / 38
页数:12
相关论文
共 50 条
  • [31] Attention Guided Low-Light Image Enhancement with a Large Scale Low-Light Simulation Dataset
    Lv, Feifan
    Li, Yu
    Lu, Feng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (07) : 2175 - 2193
  • [32] Attention Guided Low-Light Image Enhancement with a Large Scale Low-Light Simulation Dataset
    Feifan Lv
    Yu Li
    Feng Lu
    International Journal of Computer Vision, 2021, 129 : 2175 - 2193
  • [33] Coherent Event Guided Low-Light Video Enhancement
    Liang, Jinxiu
    Yang, Yixin
    Li, Boyu
    Duan, Peiqi
    Xu, Yong
    Shi, Boxin
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 10581 - 10591
  • [34] A Novel Framework for Extremely Low-light Video Enhancement
    Kim, Minjac
    Park, Dubok
    Han, David K.
    Ko, Hanscok
    2014 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS (ICCE), 2014, : 93 - 94
  • [35] EFCANet: Exposure Fusion Cross-Attention Network for Low-Light Image Enhancement
    Yang, Zhe
    Liu, Fangjin
    Li, Jinjiang
    APPLIED SCIENCES-BASEL, 2023, 13 (01):
  • [36] Low-Light Video Enhancement with Synthetic Event Guidance
    Liu, Lin
    An, Junfeng
    Liu, Jianzhuang
    Yuan, Shanxin
    Chen, Xiangyu
    Zhou, Wengang
    Li, Houqiang
    Wang, Yan Feng
    Tian, Qi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1692 - 1700
  • [37] Adaptive lightweight Transformer network for low-light image enhancement
    Meng, De
    Lei, Zhichun
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (6-7) : 5365 - 5375
  • [38] Cross-Modal Attention Mechanism for Weakly Supervised Video Anomaly Detection
    Sun, Wenwen
    Cao, Lin
    Guo, Yanan
    Du, Kangning
    BIOMETRIC RECOGNITION, CCBR 2023, 2023, 14463 : 437 - 446
  • [39] Low-light images enhancement via a dense transformer network
    Huang, Yi
    Fu, Gui
    Ren, Wanchun
    Tu, Xiaoguang
    Feng, Ziliang
    Liu, Bokai
    Liu, Jianhua
    Zhou, Chao
    Liu, Yuang
    Zhang, Xiaoqiang
    DIGITAL SIGNAL PROCESSING, 2024, 148
  • [40] A Transformer Network Combing CBAM for Low-Light Image Enhancement
    Sun, Zhefeng
    Wang, Chen
    CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (03): : 5205 - 5220