Skip navigation

Please use this identifier to cite or link to this item: http://10.10.120.238:8080/xmlui/handle/123456789/658
Full metadata record
DC FieldValueLanguage
dc.contributor.authorMumtaz D.en_US
dc.contributor.authorJakhetiya V.en_US
dc.contributor.authorNathwani K.en_US
dc.contributor.authorSubudhi B.N.en_US
dc.contributor.authorGuntuku S.C.en_US
dc.date.accessioned2023-11-30T08:44:50Z-
dc.date.available2023-11-30T08:44:50Z-
dc.date.issued2022-
dc.identifier.issn1551-3203-
dc.identifier.otherEID(2-s2.0-85122283743)-
dc.identifier.urihttps://dx.doi.org/10.1109/TII.2021.3139010-
dc.identifier.urihttp://localhost:8080/xmlui/handle/123456789/658-
dc.description.abstractWith the boom of social media communication, teleconferencing, and online classes, audiovisual communication over bandwidth strained networks has become an integral part of our lives. Consequently, the growing demand for the quality of experience necessitates developing algorithms to measure and enrich user experience. Prior studies have mainly focused on assessing speech quality and intelligibility with reference to audio quality assessment, while other categories in user-generated multimedia (UGM) are less explored. Moreover, frequency-domain properties of speech and UGM audio are significantly different from each other. Furthermore, there is a lack of a standard dataset for the quality assessment of UGM. Considering these limitations, in this article, we first develop the IIT-JMU-UGM audio dataset consisting of 1150 audio clips, with diverse context, content, and types of degradation commonly observed in real-world scenarios and annotated with the subjective quality scores. Finally, we propose a non-intrusive audio quality assessment metric using a stacked gated-recurrent-unit-based deep learning framework. The proposed model outperforms several baseline methods, including state-of-the-art non-intrusive and intrusive approaches. The resulting Pearson's correlation coefficient of 0.834 indicates that the proposed method efficiently mirrors human auditory perception. © 2005-2012 IEEE.en_US
dc.language.isoenen_US
dc.publisherIEEE Computer Societyen_US
dc.sourceIEEE Transactions on Industrial Informaticsen_US
dc.subjectAudio quality assessmenten_US
dc.subjectdeep learningen_US
dc.subjectgated recurrent unit (GRU)en_US
dc.subjectnon-intrusive quality metricen_US
dc.subjectuser-generated multimedia (UGM)en_US
dc.titleNonintrusive Perceptual Audio Quality Assessment for User-Generated Content Using Deep Learningen_US
dc.typeJournal Articleen_US
Appears in Collections:Journal Article

Files in This Item:
There are no files associated with this item.
Show simple item record


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.