Please use this identifier to cite or link to this item: https://gnanaganga.inflibnet.ac.in:8443/jspui/handle/123456789/16523
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBhagawat, Vasanth C-
dc.contributor.authorKalaiarasan, C-
dc.contributor.authorChinnaiyan, R-
dc.date.accessioned2024-08-29T05:41:24Z-
dc.date.available2024-08-29T05:41:24Z-
dc.date.issued2023-
dc.identifier.citationpp. 1-6en_US
dc.identifier.isbn9798350317060-
dc.identifier.urihttps://doi.org/10.1109/ICCAMS60113.2023.10526025-
dc.identifier.urihttps://gnanaganga.inflibnet.ac.in:8443/jspui/handle/123456789/16523-
dc.description.abstractAction recognition is an active research field in computer vision that has gained significant attention due to its practical applications. While deep learning approaches have shown promise, many existing methods fail to effectively capture spatiotemporal features from videos, mainly because they do not adequately consider the diversity of temporal scales. In this paper, we propose a novel approach, termed Long-Short-Term Spatiotemporal Features CNN (LSF CNN), to address this limitation. The LSF CNN consists of two subnetworks: the Long-Term Spatiotemporal Features Extraction Network (LT-Net) and the Short-Term Spatiotemporal Features Extraction Network (ST-Net). The LT-Net extracts long-term spatiotemporal features from stacked RGB images, while the ST-Net utilizes optical flow computed from adjacent frames to extract short-term spatiotemporal features. Additionally, we introduce a novel expression for the optical flow field, which has been empirically shown to outperform traditional methods in action recognition tasks. The spatiotemporal features from both scales are combined in a fully-connected layer and fed into a linear Support Vector Machine (SVM) for classification. By leveraging the two-stream architecture and the improved optical flow expression, our proposed approach enables comprehensive learning of deep features in both spatial and temporal domains. Extensive experiments conducted on the HMDB51 and UCF101 datasets validate the effectiveness of our approach, as it significantly improves action recognition accuracy by effectively incorporating long-short-term spatiotemporal information. This research contributes to advancing the state-of-the-art in human action recognition, opening up new possibilities for enhanced performance in various real-world applications. © 2023 IEEE.en_US
dc.language.isoenen_US
dc.publisher2023 International Conference on New Frontiers in Communication, Automation, Management and Security, ICCAMS 2023en_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.subjectCnnen_US
dc.subjectMachine Learningen_US
dc.titleMotion-Classification Using Cnn and Lstmen_US
dc.typeArticleen_US
Appears in Collections:Conference Papers

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.