Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.13091/1008
Full metadata record
DC FieldValueLanguage
dc.contributor.authorMuhammad, Khan-
dc.contributor.authorMustaqeem-
dc.contributor.authorUllah, Amin-
dc.contributor.authorImran, Ali Shariq-
dc.contributor.authorSajjad, Muhammad-
dc.contributor.authorKıran, Mustafa Servet-
dc.contributor.authorde Albuquerque, Victor Hugo C.-
dc.date.accessioned2021-12-13T10:32:19Z-
dc.date.available2021-12-13T10:32:19Z-
dc.date.issued2021-
dc.identifier.issn0167-739X-
dc.identifier.issn1872-7115-
dc.identifier.urihttps://doi.org/10.1016/j.future.2021.06.045-
dc.identifier.urihttps://hdl.handle.net/20.500.13091/1008-
dc.description.abstractHuman action recognition in videos is an active area of research in computer vision and pattern recognition. Nowadays, artificial intelligence (AI) based systems are needed for human-behavior assessment and security purposes. The existing action recognition techniques are mainly using pre-trained weights of different AI architectures for the visual representation of video frames in the training stage, which affect the features' discrepancy determination, such as the distinction between the visual and temporal signs. To address this issue, we propose a bi-directional long short-term memory (BiLSTM) based attention mechanism with a dilated convolutional neural network (DCNN) that selectively focuses on effective features in the input frame to recognize the different human actions in the videos. In this diverse network, we use the DCNN layers to extract the salient discriminative features by using the residual blocks to upgrade the features that keep more information than a shallow layer. Furthermore, we feed these features into a BiLSTM to learn the long-term dependencies, which is followed by the attention mechanism to boost the performance and extract the additional high-level selective action related patterns and cues. We further use the center loss with Softmax to improve the loss function that achieves a higher performance in the video-based action classification. The proposed system is evaluated on three benchmarks, i.e., UCF11, UCF sports, and J-HMDB datasets for which it achieved a recognition rate of 98.3%, 99.1%, and 80.2%, respectively, showing 1%-3% improvement compared to the state-of-the-art (SOTA) methods. (C) 2021 Elsevier B.V. All rights reserved.en_US
dc.description.sponsorshipERCIM 'Alain Benoussan' Fellowship Programme [2019-40]; Brazilian National Council for Research and Development (CNPq)Conselho Nacional de Desenvolvimento Cientifico e Tecnologico (CNPQ) [304315/2017-6, 430274/2018-1]en_US
dc.description.sponsorshipThis work was carried out during the tenure of an ERCIM 'Alain Benoussan' Fellowship Programme under the Contract 2019-40, Color and Visual Computing Lab at the Department of Computer Science, NTNU, Gjovik, Norway. The work of Victor Hugo C. de Albuquerque was supported in part by the Brazilian National Council for Research and Development (CNPq) under Grant 304315/2017-6 and Grant 430274/2018-1.en_US
dc.language.isoenen_US
dc.publisherELSEVIERen_US
dc.relation.ispartofFUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCEen_US
dc.rightsinfo:eu-repo/semantics/closedAccessen_US
dc.subjectArtificial Intelligenceen_US
dc.subjectAction Recognitionen_US
dc.subjectAttention Mechanismen_US
dc.subjectBig Dataen_US
dc.subjectDilated Convolutional Neural Networken_US
dc.subjectDeep Bi-Directional Lstmen_US
dc.subjectMultimedia Data Securityen_US
dc.subjectBig Dataen_US
dc.subjectFrameworken_US
dc.subjectSecurityen_US
dc.subjectInterneten_US
dc.subjectMachineen_US
dc.subjectFusionen_US
dc.subjectSystemen_US
dc.subjectThingsen_US
dc.titleHuman action recognition using attention based LSTM network with dilated CNN featuresen_US
dc.typeArticleen_US
dc.identifier.doi10.1016/j.future.2021.06.045-
dc.identifier.scopus2-s2.0-85111316846en_US
dc.departmentFakülteler, Mühendislik ve Doğa Bilimleri Fakültesi, Bilgisayar Mühendisliği Bölümüen_US
dc.authoridMuhammad, Khan/0000-0003-4055-7412-
dc.authorwosidMuhammad, Khan/L-9059-2016-
dc.authorwosid, Mustaqeem/AAM-9396-2020-
dc.authorwosidde Albuquerque, Victor Hugo C./C-3677-2016-
dc.authorwosidUllah, Amin/AAH-5034-2020-
dc.authorwosidSannino, Giovanna/N-1319-2013-
dc.identifier.volume125en_US
dc.identifier.startpage820en_US
dc.identifier.endpage830en_US
dc.identifier.wosWOS:000687315100009en_US
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanıen_US
dc.authorscopusid56651946700-
dc.authorscopusid57212930354-
dc.authorscopusid57195399776-
dc.authorscopusid56109077100-
dc.authorscopusid57215455402-
dc.authorscopusid54403096500-
dc.authorscopusid36239105500-
dc.identifier.scopusqualityQ2-
item.languageiso639-1en-
item.fulltextWith Fulltext-
item.cerifentitytypePublications-
item.openairetypeArticle-
item.grantfulltextembargo_20300101-
item.openairecristypehttp://purl.org/coar/resource_type/c_18cf-
crisitem.author.dept02.03. Department of Computer Engineering-
Appears in Collections:Mühendislik ve Doğa Bilimleri Fakültesi Koleksiyonu
Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collections
WoS İndeksli Yayınlar Koleksiyonu / WoS Indexed Publications Collections
Files in This Item:
File SizeFormat 
1-s2.0-S0167739X21002405-main.pdf
  Until 2030-01-01
2.39 MBAdobe PDFView/Open    Request a copy
Show simple item record



CORE Recommender

SCOPUSTM   
Citations

22
checked on Apr 20, 2024

WEB OF SCIENCETM
Citations

104
checked on Apr 20, 2024

Page view(s)

174
checked on Apr 22, 2024

Download(s)

6
checked on Apr 22, 2024

Google ScholarTM

Check




Altmetric


Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.