Skip to main content

EDITORIAL article

Front. Neurosci., 10 May 2023
Sec. Perception Science
This article is part of the Research Topic Scene-Dependent Image Quality and Visual Assessment View all 7 articles

Editorial: Scene-dependent image quality and visual assessment

\r\nJohn Jarvis
John Jarvis1*Sophie TriantaphillidouSophie Triantaphillidou1Mylene FariasMylene Farias2Robin Jenkin,Robin Jenkin1,3
  • 1Department of Computer Science and Engineering, University of Westminster, London, United Kingdom
  • 2Department of Computer Science, Texas State University, San Marcos, TX, United States
  • 3NVIDIA, Santa Clara, CA, United States

Editorial on the Research Topic
Scene-dependent image quality and visual assessment

Image quality is defined as the degree of excellence the image conveys. It is assessed using visual (image) psychophysics and predictive modeling. Engineering type image quality models, employed in imaging system design and optimization, are based on the assumption that the perceived image quality is a function of both the imaging system performance and the human visual system performance. Such models employ imaging system performance parameters related to color, sharpness, resolution, noise, etc. as input parameters, along with models of the human visual system. System performance parameters are typically derived from simple test charts depicting well-characterized signals captured under strict laboratory conditions. Such signals include, for example, uniform color patches with known lightness and chroma, sinusoidal exposures, edges and random noise with known spatial frequency contents and contrast. Similarly, vision models employed in image quality modeling are largely based on visual measurements obtained from simple test stimuli, such as sine-waves, Gabor functions, random noise, and uniform color patches. The modeled image quality is, as a result, system-dependent but scene content-independent. Computational models have also been used in the determination of image quality and these base their predictions on the contents of captured and processed images. The modeled image quality, in these cases, is scene content dependent, but the system performance parameters are undifferentiated.

There are many unresolved challenges when it comes to the assessment of image and video quality. To begin with, engineering-type quality models that account for imaging system performance, fail to account for scene content, since they do not address the action of advanced algorithms which are incorporated into commercial imaging systems. These algorithms are image content aware and become more or less active depending on the individual image structures, local tones and colors. In contrast, computational quality models derived directly from images of individual natural scenes fail to directly relate to component parameters in the imaging system hardware and signal image processes. They are therefore unfit for system design and optimization.

Recently, advances have been made in measuring and modeling human spatial vision, as well as spatial camera performance directly from natural scene images. In human vision studies, contrast detection and discrimination sensitivity (directly relating to perceived sharpness and resolution) were determined from the images' narrow spatial frequency bands, viewed within the context of the complete image contents (Triantaphillidou et al., 2019; Jarvis et al., 2022). Therefore, they provide visual system performance estimates which are scene content dependent. In camera measurement studies, scene-and-process-dependent camera Modulation Transfer Functions and the Noise Power Spectra were derived from printed natural images (Fry et al., 2019a), and more recently directly from natural scene captures (van Zwanenberg et al., 2021a,b). Initial research indicated that such scene content- dependent visual and camera functions were of benefit to “purer in form” quality metrics that do not rely on observers' data fittings (Fry et al., 2019b).

Nonetheless, studies such as those mentioned above are only just starting to appear in the literature, and there are still many more questions to answer. With respect to the human observer, how do psychological factors such as aesthetic appreciation, individual preferences and attention span impact quality? Cultural differences may also impact quality estimations, particularly with respect to preference for color hue and saturation. Further research on scene- and process- dependent image quality assessment is also necessary in the evaluation of the robustness of non-human (machine) perception systems; such research is still in its infancy.

The six articles presented in this Research Topic address a wide range of challenges in assessing image quality, embracing topics relating to color perception, human decision-making processes, aesthetics, the impact of optical glare and the application of neural network theory.

On the subject of aesthetics, Leder et al. present a taxonomy of factors likely to contribute to image quality and perceived beauty in mobile phone images. A model for visual processing is described, which addresses important psychological factors that are not incorporated in traditional quality models (which focus mainly on image properties such as resolution and pictorial noise). The model is an extension of a previous version describing aesthetic appreciation in art and could offer an important future framework for our understanding of quality and beauty in reproduced images in general.

Further psychological factors are covered in a study of the decision–making heuristics involved in quality estimations outlined by Leisti et al.. The authors present three experiments designed to investigate psychological factors underlying quality appraisal and the length of time taken in the process. The results obtained highlight a number of important factors to be taken into consideration when experimentally measuring human image quality preferences.

Umair Arif et al. investigate the perception of 3D virtual images appearing within a real visual environment using a reaction time experiment. The work provides some interesting results and quantifies the accuracy of the reaction time data by measuring and analyzing omitted, completed and anticipated responses. The development of commercial products employing 3D objects immersed within an image, is expanding and the results presented by the authors will provide useful background in the evolution of this technology.

Color science features in the contribution from Xie et al. who examine the concept that a larger display color gamut volume (CGV) produces higher perceived brightness and chroma. In this study, RGBW (RGB plus white channel) displays were investigated in a paired comparison experiment to obtain scales of color gamut expressed as perceived brightness and colorfulness. The results show that the actual structure of a reproduced image is important when optimizing CGV.

Artificial neural networks form the basis of a study by Prabhushankar and AlRegib who outline a thorough review of problems associated with deep neural network models when noise is introduced either from the imaging stages or from environmental factors. A new concept termed “stochastic surprisal” is introduced, which relates to the network and the input and this is tested through image quality assessments and recognition.

Finally, the visual mechanism relating to glare as a component in image quality modeling is outlined by McCann et al.. The authors give a comprehensive review of previous work on glare and its impact on brightness perception and provide a computational model of glare, with potential to provide a useful component for incorporation in future image quality models.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Conflict of interest

RJ was employed by NVIDIA.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Fry, E. W. S., Triantaphillidou, S., Jenkin, R. B., Jacobson, R. E., and Jarvis, J. R. (2019a). Validation of modulation transfer functions and noise power spectra from natural scenes. J. Im. Sci. Tech. 9, 60406-1–60406-11. doi: 10.2352/J.ImagingSci.Technol.2019.63.6.060406

CrossRef Full Text | Google Scholar

Fry, E. W. S., Triantaphillidou, S., Jenkin, R. B., Jacobson, R. E., and Jarvis, J. R. (2019b). Scene-and-process-dependent spatial image quality metrics. J. Im. Sci. Tech., 9, 60407-1–60407-13. doi: 10.2352/J.ImagingSci.Technol.2019.63.6.060407

CrossRef Full Text | Google Scholar

Jarvis, J., Trianraphillidou, S., and Gupta, G. (2022). Contrast discrimination in images of natural scenes. J. Opt. Soc. Am. A 39, B50–B64. doi: 10.1364/JOSAA.447390

PubMed Abstract | CrossRef Full Text | Google Scholar

Triantaphillidou, S., Jarvis, J, Psarrou, A., and Gupta, G. (2019). Contrast sensitivity in images of natural scenes. Signal Process Image Commun. 75, 64-75. doi: 10.1016/j.image.2019.03.002

CrossRef Full Text | Google Scholar

van Zwanenberg, O., Triantaphillidou, S., Jenkin, R., and Psarrou, A. (2021a). Estimation of ISO12233 edge spatial frequency response from natural scene derived step-edge data. J. Im. Sci. Tech. 65, 60402-1–60402-16. doi: 10.2352/J.ImagingSci.Technol.2021.65.6.060402

CrossRef Full Text | Google Scholar

van Zwanenberg, O., Triantaphillidou, S., Psarrou, A., and Jenkin, R. (2021b). Analysis of natural scene derived spatial frequency responses for estimating camera ISO12233 slanted-edge performance. J. Im. Sci. Tech. 65, 60405-1–60405-12. doi: 10.2352/J.ImagingSci.Technol.2021.65.6.060405

CrossRef Full Text | Google Scholar

Keywords: image quality, visual modeling, visual assessment, scene and process dependency, imaging system performance, image quality metrics

Citation: Jarvis J, Triantaphillidou S, Farias M and Jenkin R (2023) Editorial: Scene-dependent image quality and visual assessment. Front. Neurosci. 17:1205341. doi: 10.3389/fnins.2023.1205341

Received: 13 April 2023; Accepted: 24 April 2023;
Published: 10 May 2023.

Edited and reviewed by: Rufin VanRullen, Centre National de la Recherche Scientifique (CNRS), France

Copyright © 2023 Jarvis, Triantaphillidou, Farias and Jenkin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: John Jarvis, ai5qYXJ2aXMxJiN4MDAwNDA7d2VzdG1pbnN0ZXIuYWMudWs=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.