Video Quality Assessment Model for Exposure Correction with Vision-Language Guidance
Light-VQA+ is a video quality assessment model specialized in evaluating the performance of video exposure correction algorithms. It utilizes vision-language guidance from CLIP to extract brightness, noise, and brightness consistency features, and fuses them with semantic and motion features via a cross-attention module. The model also incorporates a trainable attention mechanism to align with the Human Visual System.