Encoder-Decoder Architectures based Video Summarization using Key-Shot Selection Model

With the exponential growth of video data, video summarization has become a challenging task. In this article, we propose a deep learning framework for video summarization that utilizes a sequence learning cum encoder-decoder network architecture with a key-shot selection model. We develop two RNN-b...

Celý popis

Uloženo v:
Podrobná bibliografie
Vydáno v:Multimedia tools and applications Ročník 83; číslo 11; s. 31395 - 31415
Hlavní autoři: Yashwanth, Kolli, Soni, Badal
Médium: Journal Article
Jazyk:angličtina
Vydáno: New York Springer US 01.03.2024
Springer Nature B.V
Témata:
ISSN:1573-7721, 1380-7501, 1573-7721
On-line přístup:Získat plný text
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
Popis
Shrnutí:With the exponential growth of video data, video summarization has become a challenging task. In this article, we propose a deep learning framework for video summarization that utilizes a sequence learning cum encoder-decoder network architecture with a key-shot selection model. We develop two RNN-based deep models, Additive Attentive Summariser (AAS) and Multiplicative Attentive Summariser (MAS), as well as a CNN-based model named - Sequential CNN Summariser (SCS). Our SCS and MAS model displays state-of-the-art performance in semantic segmentation, which we leverage to achieve superior performance in video summarization. We evaluate our models on two well-known datasets, SumMe and TVSum, and show that our proposed MAS and SCS models outperform state-of-the-art models such as DR-DSN. The proposed MAS model achieved an average F1 score of 44.1% and 60.7% on SumMe and TVSum datasets, respectively. Further, our contributions include the development of novel RNN-based and CNN-based models for video summarization and comprehensive experimental evaluations on multiple datasets that demonstrate the effectiveness of our proposed models.
Bibliografie:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1573-7721
1380-7501
1573-7721
DOI:10.1007/s11042-023-16700-3