Skip navigation

Please use this identifier to cite or link to this item: http://10.10.120.238:8080/xmlui/handle/123456789/824
Full metadata record
DC FieldValueLanguage
dc.contributor.authorSingh H.en_US
dc.contributor.authorSuman S.en_US
dc.contributor.authorSubudhi B.N.en_US
dc.contributor.authorJakhetiya V.en_US
dc.contributor.authorGhosh A.en_US
dc.date.accessioned2023-11-30T08:51:08Z-
dc.date.available2023-11-30T08:51:08Z-
dc.date.issued2022-
dc.identifier.issn2691-4581-
dc.identifier.otherEID(2-s2.0-85142827234)-
dc.identifier.urihttps://dx.doi.org/10.1109/TAI.2022.3221912-
dc.identifier.urihttp://localhost:8080/xmlui/handle/123456789/824-
dc.description.abstractSeveral research works have been developed in the area of action recognition. Unfortunately, when these algorithms are applied to low-light or dark videos, their performances are highly affected and found to be very poor or fall rapidly. To address the issue of improving the performance of action recognition in dark or low-light videosen_US
dc.description.abstractin this article, we have developed an efficient deep 3D CNN based action recognition model. The proposed algorithm follows two-stages for action recognition. In the first stage, the low-light videos are enhanced using Zero-Reference Deep Curve Estimation (Zero-DCE), followed by the min-max sampling algorithm. In the latter stage, we propose an action classification network to recognize the actions in the enhanced videos. In the proposed action classification network, we explored the capabilities of the <inline-formula><tex-math notation="LaTeX">$R(2+1)D$</tex-math></inline-formula> for spatio-temporal feature extraction. The model&#x0027en_US
dc.description.abstracts overall generalization performance depends on how well it can capture long-range temporal structure in videos, which is essential for action recognition. So we have used a Graph convolutional network (GCN) on the top of R(2+1)D as our video feature encoder which captures long-term temporal dependencies of the extracted features. Finally, a Bidirectional Encoder Representations from Transformers (BERT) is adhered to classify the actions from the 3D features extracted from the enhanced video scenes. The effectiveness of the proposed action recognition scheme is verified on ARID V1.0 and ARID V1.5 datasets. It is observed that the proposed algorithm is able to achieve <inline-formula><tex-math notation="LaTeX">$96.60\%$</tex-math></inline-formula> and <inline-formula><tex-math notation="LaTeX">$99.88\%$</tex-math></inline-formula> as Top-1 and Top-5 accuracy, respectively, on ARID V1.0 dataset. Similarly, on ARID V1.5, the proposed algorithm is able to achieve <inline-formula><tex-math notation="LaTeX">$86.93\%$</tex-math></inline-formula> and <inline-formula><tex-math notation="LaTeX">$99.35\%$</tex-math></inline-formula> as Top-1 and Top-5 accuracies, respectively. To corroborate our findings, we have compared the results obtained by the proposed scheme with those of fifteen state-of-the-art action recognition techniques. IEEEen_US
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.sourceIEEE Transactions on Artificial Intelligenceen_US
dc.subjectAction recognitionen_US
dc.subjectArtificial intelligenceen_US
dc.subjectBit error rateen_US
dc.subjectConvolutional neural networksen_US
dc.subjectDark videoen_US
dc.subjectFeature extractionen_US
dc.subjectImage processingen_US
dc.subjectThree-dimensional displaysen_US
dc.subjectTransformersen_US
dc.subjectVideosen_US
dc.titleAction Recognition in Dark Videos using Spatio-temporal Features and Bidirectional Encoder Representations from Transformersen_US
dc.typeJournal Articleen_US
Appears in Collections:Journal Article

Files in This Item:
There are no files associated with this item.
Show simple item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.