论文标题
顺式和盲目嘈杂的学生,以改进图像质量评估
Conformer and Blind Noisy Students for Improved Image Quality Assessment
论文作者
论文摘要
图像恢复,增强和产生的生成模型已显着提高了生成的图像的质量。令人惊讶的是,这些模型比其他方法产生的图像更令人愉悦,但是,使用传统的感知质量指标(例如PSNR或SSIM),它们可能获得较低的感知质量评分。因此,有必要开发一个定量度量标准来反映新算法的性能,该算法应与该人的意见分数(MOS)良好结合。基于学习的感知图像质量评估(IQA)的方法通常需要扭曲和参考图像,以准确测量感知质量。但是,通常只有扭曲或生成的图像可用。在这项工作中,我们探讨了基于变压器的全参考IQA模型的性能。我们还基于使用噪声伪标记的数据从全参考教师模型到盲人学生模型的半监督知识蒸馏提出了一种IQA方法。我们的方法在NTIRE 2022感知图像质量评估挑战中取得了竞争成果:我们的全参考模型排名第四,我们的盲目嘈杂学生在70名参与者中排名第三,每个参与者在各自的轨道中。
Generative models for image restoration, enhancement, and generation have significantly improved the quality of the generated images. Surprisingly, these models produce more pleasant images to the human eye than other methods, yet, they may get a lower perceptual quality score using traditional perceptual quality metrics such as PSNR or SSIM. Therefore, it is necessary to develop a quantitative metric to reflect the performance of new algorithms, which should be well-aligned with the person's mean opinion score (MOS). Learning-based approaches for perceptual image quality assessment (IQA) usually require both the distorted and reference image for measuring the perceptual quality accurately. However, commonly only the distorted or generated image is available. In this work, we explore the performance of transformer-based full-reference IQA models. We also propose a method for IQA based on semi-supervised knowledge distillation from full-reference teacher models into blind student models using noisy pseudo-labeled data. Our approaches achieved competitive results on the NTIRE 2022 Perceptual Image Quality Assessment Challenge: our full-reference model was ranked 4th, and our blind noisy student was ranked 3rd among 70 participants, each in their respective track.