Deep Matching Online Video Stabilization Using TSNet

被引:0
|
作者
Wang, Fangyi [1 ]
Zhong, Bihua [1 ]
Liu, Tao [2 ,3 ]
Kong, Xiaofang [1 ]
Bai, Hongyang [4 ]
Wan, Gang [1 ]
机构
[1] Nanjing Univ Sci & Technol, Natl Key Lab Transient Phys, Nanjing 210094, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Elect & Opt Engn, Nanjing 210094, Peoples R China
[3] Nanjing Univ Sci & Technol, Jiangsu Key Lab Spectral Imaging & Intelligent Sen, Nanjing 210094, Peoples R China
[4] Nanjing Univ Sci & Technol, Sch Energy & Power Engn, Nanjing 210094, Peoples R China
基金
中国国家自然科学基金;
关键词
Trajectory; Cameras; Streaming media; Smoothing methods; Real-time systems; Motion estimation; Jitter; Three-dimensional displays; Optical flow; Neural networks; Analytic hierarchy process (AHP); trajectory smoothing; video stabilization;
D O I
10.1109/TIM.2024.3476524
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the continuous progress of unmanned aerial vehicle (UAV) target recognition and detection technology, the application of UAVs in measurement and inspection is becoming more and more extensive. However, UAVs are affected by external disturbance during flight, which leads to the degradation of the video quality recorded by the camera and thus affects the effect of target detection and recognition. To solve this problem, this article proposes a video stabilization framework based on deep matching and TSNet, which consists of three phases: motion estimation, trajectory smoothing, and image warping. In the motion estimation stage, this article adopts the keypoint matching method to estimate the camera motion, instead of using optical flow to estimate the trajectory as most of the video stabilization methods. In the trajectory smoothing phase, this article designs a plug-and-play trajectory smoothing network called TSNet and extracts the camera paths of stable and unstable videos from a public video stabilization dataset as a training set. To better evaluate the stabilization methods, this article proposes an evaluation metric called Smooth Ratio by attaching weights to several video stabilization evaluation metrics followed in most papers through the analytic hierarchy process (AHP). The experimental results show that the average stabilization time of our method is 0.067 s for one frame, and the stabilization effect is significantly compared to representative methods.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Video matching using binary signature
    Li, YJ
    Jin, JS
    Zhou, XF
    ISPACS 2005: PROCEEDINGS OF THE 2005 INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING AND COMMUNICATION SYSTEMS, 2005, : 317 - 320
  • [42] Video compression using matching pursuits
    Al-Shaykh, OK
    Miloslavsky, E
    Nomura, T
    Neff, R
    Zakhor, A
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 1999, 9 (01) : 123 - 143
  • [43] Video tracking using block matching
    Hariharakrishnan, K
    Schonfeld, D
    Raffy, P
    Yassa, F
    2003 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOL 3, PROCEEDINGS, 2003, : 945 - 948
  • [44] Video Stabilization Using ORB Detector
    Bansal, Parth
    Dinesh, Jahanvi B.
    Kumar, Shravan V. R.
    Krishna, Sujay B.
    Chandar, T. S.
    2022 14TH INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2022), 2022, : 50 - 55
  • [45] Video Stabilization Using Epipolar Geometry
    Goldstein, Amit
    Fattal, Raanan
    ACM TRANSACTIONS ON GRAPHICS, 2012, 31 (05):
  • [46] Stabilization and registration of full-motion video data using deep convolutional neural networks
    Walvoord, Derek J.
    Couwenhoven, Doug W.
    Bayer, Michael A.
    SIGNAL PROCESSING, SENSOR/INFORMATION FUSION, AND TARGET RECOGNITION XXVII, 2018, 10646
  • [47] Video stabilization using space-time video completion
    Voronin, V.
    Frantc, V.
    Marchuk, V.
    Shrayfel, I.
    Gapon, N.
    Agaian, S.
    MOBILE MULTIMEDIA/IMAGE PROCESSING, SECURITY, AND APPLICATIONS 2016, 2016, 9869
  • [48] Video scene retrieval using online video annotation
    Masuda, Tomoki
    Yamamoto, Daisuke
    Ohira, Shigeki
    Nagao, Katashi
    NEW FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2008, 4914 : 54 - +
  • [49] A Deep-Intelligence Framework for Online Video Processing
    Zhang, Weishan
    Xu, Liang
    Li, Zhongwei
    Lu, Qinghua
    Liu, Yan
    IEEE SOFTWARE, 2016, 33 (02) : 44 - 51
  • [50] TSNet: Deep Ne work for Human Action Recognition in Hazy Videos
    Chaudhary, Sachin
    Murala, Subrahmanyam
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 3981 - 3986