Study of Subjective and Objective Quality Assessment of Audio-Visual Signals. 2020

Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik

The topics of visual and audio quality assessment (QA) have been widely researched for decades, yet nearly all of this prior work has focused only on single-mode visual or audio signals. However, visual signals rarely are presented without accompanying audio, including heavy-bandwidth video streaming applications. Moreover, the distortions that may separately (or conjointly) afflict the visual and audio signals collectively shape user-perceived quality of experience (QoE). This motivated us to conduct a subjective study of audio and video (A/V) quality, which we then used to compare and develop A/V quality measurement models and algorithms. The new LIVE-SJTU Audio and Video Quality Assessment (A/V-QA) Database includes 336 A/V sequences that were generated from 14 original source contents by applying 24 different A/V distortion combinations on them. We then conducted a subjective A/V quality perception study on the database towards attaining a better understanding of how humans perceive the overall combined quality of A/V signals. We also designed four different families of objective A/V quality prediction models, using a multimodal fusion strategy. The different types of A/V quality models differ in both the unimodal audio and video quality prediction models comprising the direct signal measurements and in the way that the two perceptual signal modes are combined. The objective models are built using both existing state-of-the-art audio and video quality prediction models and some new prediction models, as well as quality-predictive features delivered by a deep neural network. The methods of fusing audio and video quality predictions that are considered include simple product combinations as well as learned mappings. Using the new subjective A/V database as a tool, we validated and tested all of the objective A/V quality prediction models. We will make the database publicly available to facilitate further research.

UI MeSH Term Description Entries

Related Publications

Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
June 2010, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
December 2019, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
December 2015, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
August 2020, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
January 2024, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
January 2023, IEEE transactions on image processing : a publication of the IEEE Signal Processing Society,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
May 1994, Journal of digital imaging,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
December 2018, Otolaryngologia polska = The Polish otolaryngology,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
July 2018, Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference,
Xiongkuo Min, and Guangtao Zhai, and Jiantao Zhou, and Mylene C Q Farias, and Alan Conrad Bovik
January 2021, Frontiers in neuroscience,
Copied contents to your clipboard!