Subjective IQA of the interpolated images from the Middlebury benchmarkCurrent benchmarks for optical flow algorithms evaluate the estimation either directly by comparing the predicted flow fields with the ground truth or indirectly by using the predicted flow fields for frame interpolation and then comparing the interpolated frames with the actual frames. In the latter case, objective quality measures such as the mean squared error are typically employed. However, it is well known that for image quality assessment, the actual quality experienced by the user cannot be fully deduced from such simple measures. Hence, we conducted StudyMB 2.0, a subjective quality assessment crowdsourcing study for the interpolated frames provided by one of the optical flow benchmarks, the Middlebury interpolation benchmark.
StudyMB 2.0 contains interpolated frames from 155 methods applied to each of 8 contents. We collected forced-choice paired comparisons between interpolated images and corresponding ground truth. To increase the sensitivity of observers when judging minute difference in paired comparisons we introduced a new method to the field of full-reference quality assessment, called artefact amplification. From the crowdsourcing data (3720 comparisons of 20 votes each) we reconstructed absolute quality scale values according to Thurstone's model.
|
Cite usStudyMB 2.0 is freely available to the research community. If you use our database in your research, please cite the following reference:
@article{men2020subjective, title={Subjective Annotation for a Frame Interpolation Benchmark using Artifact Amplification}, author={Men, Hui and Hosu, Vlad and Lin, Hanhe and Bruhn, Andr{\'e}s and Saupe, Dietmar}, journal={arXiv preprint arXiv:2001.06409}, year={2020} } |
Downloads
346.3 MB download
23.3 kB download
43 kB download
|
The scores file
|