Improving AI-assisted video editing: Optimized footage analysis through multi-task learning

被引:0
|
作者
Li, Yuzhi [1 ]
Xu, Haojun [1 ]
Cai, Feifan [1 ]
Tian, Feng [1 ]
机构
[1] Shanghai Univ, Shanghai, Peoples R China
关键词
Footage analysis; Multi-task learning; AI-assisted video editing;
D O I
10.1016/j.neucom.2024.128485
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, AI-assisted video editing has shown promising applications. Understanding and analyzing camera language accurately is fundamental in video editing, guiding subsequent editing and production processes. However, many existing methods for camera language analysis overlook computational efficiency and deployment requirements in favor of improving classification accuracy. Consequently, they often fail to meet the demands of scenarios with limited computing power, such as mobile devices. To address this challenge, this paper proposes an efficient multi-task camera language analysis pipeline based on shared representations. This approach employs a multi-task learning architecture with hard parameter sharing, enabling different camera language classification tasks to utilize the same low-level feature extraction network, thereby implicitly learning feature representations of the footage. Subsequently, each classification sub- task independently learns the high-level semantic information corresponding to the camera language type. This method significantly reduces computational complexity and memory usage while facilitating efficient deployment on devices with limited computing power. Furthermore, to enhance performance, we introduce a dynamic task priority strategy and a conditional dataset downsampling strategy. The experimental results demonstrate that achieved a comprehensive accuracy surpassing all previous methods. Moreover, training time was reduced by 66.33%, inference cost decreased by 59.85%, and memory usage decreased by 31.95% on the 2-task dataset MovieShots; on the 4-task dataset AVE, training time was reduced by 95.34%, inference cost decreased by 97.23%, and memory usage decreased by 61.21%.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] VHS to HDTV Video Translation Using Multi-task Adversarial Learning
    Luo, Hongming
    Liao, Guangsen
    Hou, Xianxu
    Liu, Bozhi
    Zhou, Fei
    Qiu, Guoping
    MULTIMEDIA MODELING (MMM 2020), PT I, 2020, 11961 : 77 - 86
  • [42] Improving Weakly Supervised Lesion Segmentation using Multi-Task Learning
    Chu, Tianshu
    Li, Xinmeng
    Vo, Huy V.
    Summers, Ronald M.
    Sizikova, Elena
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 143, 2021, 143 : 60 - 73
  • [43] Improving generalization ability of neural networks ensemble with multi-task learning
    State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210093, China
    不详
    J. Comput. Inf. Syst., 2006, 4 (1235-1240):
  • [44] A multi-task learning approach for improving travel recommendation with keywords generation
    Chen, Lei
    Cao, Jie
    Zhu, Guixiang
    Wang, Youquan
    Liang, Weichao
    KNOWLEDGE-BASED SYSTEMS, 2021, 233 (233)
  • [45] Advancing translational research in neuroscience through multi-task learning
    Cao, Han
    Hong, Xudong
    Tost, Heike
    Meyer-Lindenberg, Andreas
    Schwarz, Emanuel
    FRONTIERS IN PSYCHIATRY, 2022, 13
  • [46] Supervised Shallow Multi-task Learning: Analysis of Methods
    Abhadiomhen, Stanley Ebhohimhen
    Nzeh, Royransom Chimela
    Ganaa, Ernest Domanaanmwi
    Nwagwu, Honour Chika
    Okereke, George Emeka
    Routray, Sidheswar
    NEURAL PROCESSING LETTERS, 2022, 54 (03) : 2491 - 2508
  • [47] Multi-Task Integration as a Strategy for Improving Teacher Performance and Student Learning
    Thongnin, Panya
    Wongwanich, Suwimon
    Piromsombat, Chayut
    5TH WORLD CONFERENCE ON EDUCATIONAL SCIENCES, 2014, 116 : 1919 - 1924
  • [48] Prediction of drug–target interactions through multi-task learning
    Chaeyoung Moon
    Dongsup Kim
    Scientific Reports, 12
  • [49] A Multi-task Kernel Learning Algorithm for Survival Analysis
    Meng, Zizhuo
    Xu, Jie
    Li, Zhidong
    Wang, Yang
    Chen, Fang
    Wang, Zhiyong
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2021, PT III, 2021, 12714 : 298 - 311
  • [50] Supervised Shallow Multi-task Learning: Analysis of Methods
    Stanley Ebhohimhen Abhadiomhen
    Royransom Chimela Nzeh
    Ernest Domanaanmwi Ganaa
    Honour Chika Nwagwu
    George Emeka Okereke
    Sidheswar Routray
    Neural Processing Letters, 2022, 54 : 2491 - 2508