ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.10331
111
0

Research on Audio-Visual Quality Assessment Dataset and Method for User-Generated Omnidirectional Video

12 June 2025
F. Zhao
Da Pan
Zelu Qi
Ping Shi
ArXiv (abs)PDFHTML
Main:5 Pages
10 Figures
Bibliography:1 Pages
1 Tables
Abstract

In response to the rising prominence of the Metaverse, omnidirectional videos (ODVs) have garnered notable interest, gradually shifting from professional-generated content (PGC) to user-generated content (UGC). However, the study of audio-visual quality assessment (AVQA) within ODVs remains limited. To address this, we construct a dataset of UGC omnidirectional audio and video (A/V) content. The videos are captured by five individuals using two different types of omnidirectional cameras, shooting 300 videos covering 10 different scene types. A subjective AVQA experiment is conducted on the dataset to obtain the Mean Opinion Scores (MOSs) of the A/V sequences. After that, to facilitate the development of UGC-ODV AVQA fields, we construct an effective AVQA baseline model on the proposed dataset, of which the baseline model consists of video feature extraction module, audio feature extraction and audio-visual fusion module. The experimental results demonstrate that our model achieves optimal performance on the proposed dataset.

View on arXiv
@article{zhao2025_2506.10331,
  title={ Research on Audio-Visual Quality Assessment Dataset and Method for User-Generated Omnidirectional Video },
  author={ Fei Zhao and Da Pan and Zelu Qi and Ping Shi },
  journal={arXiv preprint arXiv:2506.10331},
  year={ 2025 }
}
Comments on this paper