MULTI-MODAL GRAPHNET LEARNING-BASED FEATURE EXTRACTION FOR SPATIOTEMPORAL SALIENCY DETECTION ON MULTIMEDIA VIDEOS
Abstract
In multimedia content analysis, spatiotemporal saliency detection plays a crucial role in understanding visual data. However, existing methods often struggle with efficiently capturing complex patterns in videos. To address this, we propose a Multi-modal GraphNet Learning-Based Feature Extraction approach. Our method integrates multi-modal information from both spatial and temporal domains to enhance saliency detection accuracy. By leveraging GraphNet, we effectively model the intricate relationships among video frames. We validate our approach on a diverse set of multimedia videos, demonstrating significant improvements in saliency detection performance. Specifically, our method achieves an average precision of 0.85 and a recall of 0.78, outperforming state-of-the-art techniques. Furthermore, our approach exhibits robustness across various video types and scenarios. Through experimental evaluation, we confirm the efficacy of our proposed method in enhancing spatiotemporal saliency detection. This work contributes to advancing the field of multimedia analysis, offering a promising solution for understanding visual content in videos.

Authors
S.V. Prabhakar, M.D. Ambika
Maharani’s Science College for Women, India

Keywords
Multi-Modal, GraphNet, Feature Extraction, Spatiotemporal Saliency Detection, Multimedia Videos
Yearly Full Views
JanuaryFebruaryMarchAprilMayJuneJulyAugustSeptemberOctoberNovemberDecember
000007000000
Published By :
ICTACT
Published In :
ICTACT Journal on Image and Video Processing
( Volume: 14 , Issue: 4 , Pages: 3264 - 3272 )
Date of Publication :
May 2024
Page Views :
54
Full Text Views :
7

Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.