Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 14 March 2023
Sec. Perception Science
This article is part of the Research Topic Scene-Dependent Image Quality and Visual Assessment View all 7 articles

Stochastic surprisal: An inferential measurement of free energy in neural networks

  • Omni Lab for Intelligent Visual Engineering and Science (OLIVES), Georgia Institute of Technology, Electrical and Computer Engineering, Atlanta, GA, United States

This paper conjectures and validates a framework that allows for action during inference in supervised neural networks. Supervised neural networks are constructed with the objective to maximize their performance metric in any given task. This is done by reducing free energy and its associated surprisal during training. However, the bottom-up inference nature of supervised networks is a passive process that renders them fallible to noise. In this paper, we provide a thorough background of supervised neural networks, both generative and discriminative, and discuss their functionality from the perspective of free energy principle. We then provide a framework for introducing action during inference. We introduce a new measurement called stochastic surprisal that is a function of the network, the input, and any possible action. This action can be any one of the outputs that the neural network has learnt, thereby lending stochasticity to the measurement. Stochastic surprisal is validated on two applications: Image Quality Assessment and Recognition under noisy conditions. We show that, while noise characteristics are ignored to make robust recognition, they are analyzed to estimate image quality scores. We apply stochastic surprisal on two applications, three datasets, and as a plug-in on 12 networks. In all, it provides a statistically significant increase among all measures. We conclude by discussing the implications of the proposed stochastic surprisal in other areas of cognitive psychology including expectancy-mismatch and abductive reasoning.

1. Introduction

The human visual system is the resultant of an evolutionary process influenced and constrained by the natural visual stimuli present in the outside environment (Geisler, 2008; Sebastian et al., 2017). The free energy principle is an over-arching theory that provides a mathematical framework for this evolutionary process (Friston, 2009). The principle provides a theory of cognition that can unify and discuss relationships among fundamental psychological concepts such as memory, attention, value, reinforcement, and salience (Friston, 2009). It decomposes the visual system into perception and action modalities and argues that the visual system is an inference engine whose objective is to perceive the outside environment as best as it can. If this perception is insufficient for making an inference, an action is taken to achieve the objective by influencing the outside environment. While the action is dependent on the type of inference that is to be made, perception is dependent on the natural visual stimuli. Hence, a study of the human visual system warrants a study of the patterns that it is sensitive to. Broadly, these patterns are classified under natural scene statistics (Geisler, 2008). Color, luminance, spatio-temporal structures and spectral residues are some statistics that are useful in performing fundamental visual tasks including image quality assessment (Zhang and Li, 2012), visual saliency detection (Hou and Zhang, 2007), and object detection and recognition (Sebastian et al., 2017).

Image quality assessment is the objective assessment of subjective quality of images. Visual saliency detection finds those regions in an image that attract significant human attention. Object recognition attempts to recognize any given object in an image. Methods like Hou and Zhang (2007) and Murray et al. (2013) use spectral residue to detect salient regions. Hou and Zhang (2007) extend their spectral residue-based saliency detection algorithm to show that object detection is possible. The spectral residual concept is used in SR-SIM (Zhang and Li, 2012) and BleSS (Temel and AlRegib, 2016a) to utilize the frequency characteristics to quantify residuals for IQA. All three disparate applications share commonalities in their spectral residual statistics that are used to show comparable performance within each application. Hence, natural scene statistics and their governing visual system principles are building blocks of computational machine vision systems that attempt to mimic human perception.

One such a principle is the consistency in spatial structures that allows for a sparse set of convolutional kernels to represent natural scenes. Large-scale neural networks are built on this principle. Neural networks are empowered to mimic human vision by performing the same tasks as the human visual system including image quality assessment (Temel et al., 2016), visual saliency detection (Sun et al., 2020), and object recognition (Krizhevsky et al., 2012) among others. Recently the generalization capabilities of neural networks has led to their widespread adoption in a number of computational fields. Neural networks have produced state-of-the-art results on multifarious data ranging from natural images (Krizhevsky et al., 2012), computed seismic (Shafiq M. A. et al., 2018; Shafiq M. et al., 2018), and biomedical images (Prabhushankar and AlRegib, 2021b; Prabhushankar et al., 2022). In object recognition on Imagenet dataset (Deng et al., 2009), He et al. (2016) surpassed the top five human accuracy of 94.9%. In the application of image quality assessment, Bosse et al. (2017) extracted patch-wise distortion characteristics from images using deep neural networks before fusing them to obtain an objective quality score. The authors in Liu et al. (2017) device a sparse representation-based entropic measure of quality that is inspired by the free energy principle. This is extended in Liu et al. (2019) where the authors use the free energy principle as a plug-in on top of existing blind image quality assessment techniques. In both these works, free energy principle is seen as a technique that measures the disparity between an outside environment and the expectation of that environment through some biologically plausible mechanism. Other existing works, including Zhai et al. (2011) and Gu et al. (2014), quantify this disparity to estimate quality.

Hence, from the perspective of free energy principle, neural networks act as biologically plausible mechanisms to perceive the outside environment. This is done by supervising the networks to learn particular tasks. Prabhushankar and AlRegib (2021a) describe supervised learning as associative learning where a set of learned features is associated with any given class. This class can be an objective score in image quality assessment or an object class from recognition. The learned features are associated with a specific dataset and application, and are not easily transferable (Temel et al., 2018). A number of recent works including (Goodfellow et al., 2014; Temel et al., 2017; Hendrycks and Dietterich, 2019) describe the fallibility of neural networks to adversarial noise and slight perturbations in data arising from acquisition or environmental errors. The feature representation space can be altered significantly by noise that is sometimes non-noticeable in data. This is in contrast with the spectral residual feature which is used to infer both object (Hou and Zhang, 2007) and image quality (Zhang and Li, 2012; Temel and AlRegib, 2016a).

We posit that these shortcomings of supervised neural networks are a resultant of neural networks exclusively utilizing the perception modality of free energy principle. In other words, the passivity of neural networks during inference leads to their non-robust nature. This view is corroborated by Demekas et al. (2020) who identify three challenges in supervised learning. Firstly, they claim that neural networks lack an explicit control mechanism of incorporating prior beliefs into predictions. Secondly, neural networks train via a scalar loss function that does not allow for incorporating uncertainty in action. Lastly, neural networks do not perform any action during inference that would elicit changes in the input from the outside environment.

In this paper, we tackle the above challenges by introducing a framework for action during inference. This is opposed to the free energy principle based works in Liu et al. (2017, 2019) where the methodology does not require actions at inference. Based on the free energy principle, we treat any trained neural network as an inference engine. We define a quantity called stochastic surprisal that is a function of a neural network's inference and some action performed on this inference. Reducing surprisal is generally seen as a single action that reduces the distributional difference between two quantities. However, during inference, we have access to only a single data point. We overcome this challenge by considering that all possible actions that the network can undertake are equally likely. The term stochastic is derived based on this assumption of action-randomness. Stochastic surprisal acts on top of any existing neural networks to address the challenge of passive inference. Existing neural networks can either be generative or discriminative. We evaluate stochastic surprisal on two applications including image quality assessment and robust object recognition. In image quality assessment, we evaluate our technique to assess the quality of distorted images at different levels of distortions. Similarly, in robust object recognition, we recognize distorted images when the original neural network is only trained on pristine images. In other words, we propose a concept that is able to assess the noise characteristics in images to assign objective quality, as well as ignore the same noise characteristics to robustly classify images. The contributions of this paper include,

• We unify the concepts of image quality assessment and robust object recognition. We show that the features that are extracted from neural networks simultaneously characterize the scene and context within the image for recognition as well as the noise perturbing it's quality.

• We term our features as stochastic surprisal and relate them to the free energy principle. We provide a mathematical framework to extract stochastic surprisal from both discriminative and generative neural networks as a function of some action.

• We discuss the implications of our proposed method from an abductive reasoning as well as expectancy-mismatch perspective. Both these concepts lead to separate applications including context and relevance based contrastive visual explanations and human visual saliency detection.

We first describe the free energy principle in Section 2.1.1. The free energy principle is then related to neural networks in Section 2.1.2 before describing stochastic surprisal. The generation of stochastic surprisal in generative and discriminative networks is described in Sections 2.1.2.1 and 2.1.2.2, respectively. Finally, the applications of image quality assessment and robust recognition and our methodology is discussed in Section 2.3. The results are provided in Section 3. We further discuss the implications of the proposed stochastic surprisal on other cognitive concepts and conclude in Section 4.

2. Theoretical overview and methodology

In this section, we provide a thorough background of the free energy principle and its application in neural networks, both generative and discriminative. We then define and detail the framework for the extraction of stochastic surprisal. This is followed by the application of stochastic surprisal in image quality assessment and robust recognition.

2.1. Background

2.1.1. Free energy principle

The Free Energy Principle (FEP) proposes a theory to explain the self-organizing capability of any intelligent and adaptive system (Friston, 2009). FEP assumes the demarcation of a system that exists in an environment through a functional Markov Blanket. The Markov Blanket (Hipólito et al., 2021) provides statistical independence to the system from its environment, thereby imbuing the system with a sense of self . A consequence of this separation is that the system only experiences the environment through the Markov Blanket based on a limited set of sensory inputs. These sensory inputs are used to create a generative model of the outside environment within the system. The system then performs a limited set of actions affecting the outside environment while updating its internal model of the outside environment. The FEP provides a mathematically concrete set of principles to bound the long-term entropy of the internal generative model that is confined in the set of all possible sensory inputs and its possible performative actions. Friston (2019) argues that the assumption of the Markov Blanket and the ensuing FEP is an overarching theory that provides a tool to study and explain self-organization at any spatio-temporal scale from infinitesimal quantum mechanics to generational biological evolution.

In this paper, we are interested in the FEP's application to visual processes related to the human brain. The applicability of FEP across concepts such as memory, attention, value, and reinforcement (Friston, 2009) is possible because of the central assumption that the limited sensory inputs from the outside environment to the brain are also likely sensory inputs. In other words, the human brain only allows for a limited set of likely encounters (Demekas et al., 2020). The term likely is a function of the expectation set by the internal generative model within the brain. Hence, the brain is considered to encode a Bayesian recognition density that predicts the sensory inputs based on some hypothesis regarding their cause. This leads to the proposition that the brain is an inverse generative model where it expects to sense only a limited set of likely inputs from the environment. Any mismatch to this expectation is handled in two stages. Firstly, the internal model is updated with the mismatched sensory input to improve the perception. Secondly, an action is performed to change the environment. This way, the environment and the model are made to fit each other by reducing the mismatched input. A mismatched input is typically termed as a surprising event (Buckley et al., 2017). Self-organization in the brain creates an imperative to minimize the surprisal of any event and the FEP provides a mathematical theory of this minimization by providing a tractable upper bound to the surprisal. Mathematically, average surprisal is the entropy of the distribution of all events. More unlikely an event, more surprisal it creates in the internal model. The free energy decomposed using surprisal (Demekas et al., 2020) is given by,

Free Energy=Divergence+Surprisal.    (1)

Here, divergence is the difference between the variables representing the outside environment that generate the sensory inputs and the variables in the internal generative model that mimic the outside world.

2.1.2. Free energy principle in neural networks

The assumption of the existence of an internal tractable generative model that is an inference engine has been adopted in the construction of early neural networks. Hinton and Zemel (1993) describe the Helmholtz free energy that is used to construct autoencoders as agents that minimize the reconstruction cost and the code cost. The code cost is a function of the entropy of the probability distribution given a vector. In FEP, this code cost is the surprisal. Variational Autoencoders (Kingma and Welling, 2019) minimize Variational Free Energy (VFE) and consequently surprisal. VFE is a generalization of the Helmholtz free energy where the divergence of the approximate and true probabilities are minimized (Gottwald and Braun, 2020). While the generative models of autoencoders lend themselves directly to the FEP, the discriminative models also train themselves using some variation of a loss function that resembles free energy. In this paper, we use both generative and discriminative models and we introduce them in terms of the free energy principle.

2.1.2.1. Generative networks

In this section, we consider a general autoencoder as our generative model. An autoencoder is an unsupervised learning network which learns a regularized representation of inputs to reconstruct them as its output (Hinton and Zemel, 1993; Kwon et al., 2019). Since Hinton and Zemel (1993), a number of variations have been proposed to autoencoders to construct either application-specific or property-specific networks. These variations generally deal with constraining the latent representations learned by an autoencoder. For instance, Ng (2011) constrain the latent representation to be sparse, thereby constructing sparse autoencoders. Kingma and Welling (2013) constrain the latent representation to follow a Gaussian distribution. These are termed as variational autoencoders. These are two instances of property-specific autoencoders. Application-specific autoencoders include fully-connected networks used for image compression (Gedeon and Harris, 1992), and convolutional autoencoders for image denoising (Mao et al., 2016).

All these autoencoders consist of the same base architecture as shown in Figure 1. They consist of an encoder fθ(·), parameterized by θ to map inputs x to a latent representation zg. These latent representations zg are used to reconstruct the same input x^ using a decoder gϕ(·). This operation is mathematically represented as,

z=fθ(x)x^=gϕ(z)=gϕ(fθ(x)),    (2)

For a natural image input, x ∈ ℝH × W × C, where H, W, C are height, width, channel of input image, respectively. The encoder and decoder are trained jointly by minimizing a loss function J(θ, ϕ) defined as:

J(θ,ϕ)=L(x,gϕ(fθ(x)))+Ω(zg;θ,ϕ),    (3)

where L is a reconstruction error which measures the dissimilarity between the input x, and the reconstructed image x^. Ω is a regularization term added to avoid overfitting the network to the training set and to imbue the required constraints. For a sparse autoencoder, Ω is an l1 sparsity constraint. However, since the l1 constraint is not differentiable, a practical solution for constructing this sparsity constraint is to use KL-Divergence on zg. Specifically, the sum of zg is constrained to either zero or a very small value using a distance metric like KL-Divergence. This is shown in Figure 1 in blue.

FIGURE 1
www.frontiersin.org

Figure 1. Block diagram for perception (in blue) and proposed action (in red) for a sparse autoencoder. The image x is taken from the CURE-TSR dataset (Temel et al., 2017). The training loss function is J(θ, ϕ). The latent representation z = fθ(·) is zg. The reconstructed image is shown as x^. This forms the perception pipeline. The action pipeline is shown in red where the action Ag is backpropagated through the decoder to the latent representation space. The learned blue perception representation space zg changes to the action space zga as a consequence of Ag. This change is stochastic surprisal, given by Ag(x,x^)ϕ.

During training, the network parameters, θ and ϕ are updated by backpropagating the gradients of J(θ, ϕ) w.r.t. the parameters. The update rule is given by,

θ=θ-J(θ,ϕ)θ,ϕ=ϕ-J(θ,ϕ)ϕ,    (4)

The two gradients provide the change in the network parameters required to incorporate better perception capabilities as measured by the loss function J(θ, ϕ).

Consider Equation (3) and compare this against the free energy decomposition in Equation (1). The L reconstruction error measures the divergence. The regularization is the surprisal. Technically, regularization prevents the network from reconstructing x exactly. Hence, surprisal is added in generative networks to make them generalizable. A thorough analysis of regularization for reconstruction and feature transfer of autoencoders to multiple tasks is provided in Prabhushankar et al. (2018). While regularization impacts the reconstruction negatively, it enhances the adaptability and usability of features for generalized tasks and test sets.

2.1.2.2. Discriminative networks

Discriminative networks are neural networks whose function is to assign labels to input data. While the required training data in generative networks are images x ∈ ℝH × W × C, the training data for discriminative networks are (x, y), where x ∈ ℝH × W × C and y ∈ [1, N]. Here, y is an integer label assigned to x, ranging between 1 and the total number of classes N. The goal of a discriminative network is to assign the label y, given x at inference. The simplest discriminative network is an image classification network. Consider an L-layered network f(·) trained to classify images on a domain X. For the task of classification, where f(·) is trained to classify between N classes, the last layer is commonly a fully connected layer consisting of N weights or filters. During inference, the representation space zd = fL−1(x) after the first (L−1) layers are projected independently onto each of the N filters. The filter with the maximum projection is inferred as the class ŷ to which x belongs. Mathematically, zd and ŷ are related as,

zd=fL-1(x),    (5)
=argmax(WLTzd+bL),ŷ=argmax()    (6)
WLdL-1×N, zdL-1×1,bLN×1,N×1,ŷ[1,N],    (7)

where WL and bL are the parameters of the final fully connected layer. Note our choice of the variable zd. This is a similar variable that is used to denote the latent representation in Equation (2). Similar to the decoder gϕ(·) acting on zg in generative networks, we have the final fully connected layer WL and bL acting on zd. This forms the perception pipeline that classifies x as ŷ. This is shown in blue in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Block diagram for perception (in blue) and proposed action (in red) for a classification network. The image x is taken from the CURE-TSR dataset (Temel et al., 2017). The perception pipeline is shown in blue where the network assigns a class 3 to x. The action pipeline is shown in red where the action Ad,a=1 is backpropagated through the final fully connected layer to the learned blue perception manifold zd. zd changes to the action manifold zda as a consequence of Ad. This change is stochastic surprisal, given by Ad(ŷ,i;W)WL.

Training an image classification technique requires a loss function J(ŷ, y; θ), where θ are the network parameters and (x, y) are the image-label pairs required for training. A common choice of J(·) is the cross-entropy loss. Considering σ(ỹ) to be the softmax probability distribution of the output vector from f(·), the cross-entropy loss interms of KL-Divergence and entropy can be expressed as,

J(·)=KL(y||σ())-i=1N(σ(yi~))ln(σ(yi~)).    (8)

Here, KL(||) refers to the KL-divergence between the probability output of the network and the label vector y expressed as a one-hot probability distribution. Notice the similarity between Equations (1) and (8). The divergence in the FEP is the KL divergence and the surprisal is the entropy given by the second term in Equation (8). Unlike the generative networks, surprisal is not introduced into the network. Rather, the existing surprisal is minimized. A number of foundational works in FEP (Friston, 2009, 2019) use the entropy of a distribution to describe free energy. The network is then trained by backpropagating the errors w.r.t θ similar to Equation (4).

2.1.2.3. Terminologies

Before describing our contributions, we summarize a few key terminologies that are extensively used within the FEP setup and how they relate to neural networks.

2.1.2.3.1. External state of the world

X is the observed distribution of the outside world and each xX is an instance of this distribution. When describing discriminative systems, data is denoted as (x, y) where x is the data point and y is its label. When dealing with generative models, data is x only. When there is some distortion associated with the outside environment, the sampled data is x′ and the distribution is X. We will see X and x′ in IQA and recognition experiments when input data are distorted by noise.

2.1.2.3.2. System

A neural network f(·) trained on a distribution X. A trained system is one that does not take in any external inputs to change or update its weights. We consider that a trained system is at NESS density. For a discriminative network, f(·) is the entire system and its training data is denoted by (x, y). For a generative network, fθ(·) is an encoder trained to produce a latent representation space zg given data denoted by x and gϕ(·) is the decoder trained to reconstruct the image given a latent representation zg.

2.1.2.3.3. Markov blanket

The part of the system that produces the latent representation z. In a generative system the Markov blanket is the encoder fθ(·) and in a discriminative system, the Markov blanket is the initial part of the network from Equation (5), fL−1(·).

2.1.2.3.4. Internal state of the system

Let z denote the internal state of the latent representation within a system. Given a generative network, the latent representation after the encoder, zg = fθ(x) is the internal state. Given a discriminative network, the internal state is zd = fL−1(x). The internal states of both the networks are interchangeably referred to as latent representations or as perception manifolds. Note that similar to external state, if an input x is distorted to x′, its internal state is also distorted and we will use either zd or zg to denote the internal state of the system. Given any action, a, the internal state shifts to za to accommodate this action without necessarily changing x. All these states are shown in Figures 1, 2.

2.2. Stochastic surprisal

During inference, the networks are passive. As discussed in Section 1 and noted by Demekas et al. (2020), there is no mechanism to include a non-scalar surprisal that allows for an action during inference. In this paper, we alleviate this challenge by defining a new quantity called stochastic surprisal as a function of a hypothetical action. Consider the differences in the existing definitions of surprisal. In generative networks from Equation (3), surprisal is the induced regularization that prevents overfitting and creates specific constraints for a latent representation zg. In discriminative networks from Equation (8), surprisal is the entropy of the network's predicted distribution obtained from a linear combination on zd. While the surprisal in Equation (1) deals with bounding the system's surprise of the distributional divergence between the internal model and external environment, the regularization-based and entropy-based definitions provide a mathematically-tractable definition in neural networks. In this paper, we provide a new mathematically-tractable definition of surprisal that is inherently a function of an action A and its effect on the network. A formal definition is provided first.

Definition 2.1 (Stochastic Surprisal). Given a trained neural network fθ(·) parameterized by θ, the gradient change Aθ with respect to the network parameters for all possible actions A from the perspective of fθ(·) is termed stochastic surprisal.

Stochastic surprisal measures the change required in the trained perception network to measure any given action A. It is stochastic since it does not measure the divergence between distributions but rather a single data point's influence on the network. It is a non-scalar value that acts on the network parameters according to Equation (4). Note that we do not actually update the network. Rather, we only measure the network update and use it as a surprisal quantity. This update is possible based on some action all of which are considered equally likely. A thorough discussion of the naming is provided in Section 4.1.

2.2.1. Action and stochastic surprisal

Action is a function of any application. We first define it in a general fashion for generative and discriminative networks. In Section 2.3, we define it specifically for image quality assessment and robust recognition.

2.2.1.1. Generative networks

The action in generative networks is straightforward. Given an image x and its reconstructed image x^, the possible action is to change the weight parameters in a way that reduces the disparity between x and x^. In this paper, we quantify this disparity as the Mean Square Error given by x-x~22. However, as described in Section 2.1.2.1, the surprisal is present in the regularization terms. Hence, any action performed has to account for this surprisal. In this paper, we use the elastic net regularization. The overall action that induces a change in the network is given by,

Ag=x-x^22+βj=1hKL(zj||ρ^j)+λW22.    (9)

where Ag is a generative action. x-x^22 is the MSE loss function, and W22 is the regularization on the weights. j=1hKL(zj||ρ^j) is the sparsity constraint denoted as the divergence between the latent representation and some small value ρ^j,j[1,h] where h is the size of the latent representation. By minimizing the KL divergence, the latent variables zj, j ∈ [1, h] are made sparse. β and λ are hyperparameters controlling the regularization.

Stochastic surprisal is the gradient of this action Ag with respect to the decoder weights. The action pipeline along with the stochastic surprisal generation is shown in Figure 1 in red. At inference, a test image is passed through a trained network and reconstructed. The action from Equation (9) is calculated and backpropagated to the latent representation space zd. The change, measured as the gradients, creates a change in zd and the new action manifold is termed zda. A toy example of the geometric interpretation of this change is also shown Figure 1. The blue perception manifold zg that reconstructs x^ is acted on by Ag to obtain a new red action manifold zda. The decoder can use this space to reconstruct x exactly. In Section 3, we show how these generated gradients can be used as features for image quality assessment. Note that we keep the perception pipeline as is and make no changes to the training process.

2.2.1.2. Discriminative networks

The action Ad in discriminative networks is more involved than generative networks. While in generative networks, the possible action is to reconstruct the image with higher fidelity, in discriminative networks, the action can take any one of N outcomes. At inference, discriminative networks are given an image x and asked to predict its label y. Assuming that ŷ is the prediction, the action we use to elicit change in the network parameters is by backpropagating an action class a in the loss function J(ŷ, a; W), a ∈ [1, N].

Ad=ai-22,i[1,N].    (10)

Here ai is the action class defined as a Kronecker delta function given by,

ai={1,if i=class,0,otherwise    (11)

There is no regularization added to the discriminative action since the probability distribution σ(ỹ) derived from ỹ is a function of its surprisal entropy. Note that we use an MSE function for Ad in Equation (10) similar to Ag from Equation (9). An important difference between Equations (9) and (10) is the number of possible actions. In discriminative networks that classify between N classes, there are N possible i in Equation (10). Hence, there are N possible actions Ad and N possible surprisals AdiWL,i[1,N]. The action pipeline for discriminative network for a toy example where the predicted class is 3 and the action class is 1 is shown in Figure 2 in red. The surprisals are the red gradients from the final fully connected layer. We also show the geometric interpretation of a given action on the learned representation space zd. The blue perception manifold is acted upon by Ad1 through Ad1WL to obtain the red action manifold. Note that there are N such possible red zda due to the N possible actions. This idea of N separate gradients to characterize data is not new. In Settles et al. (2007), the authors construct positive and negative instance labels for a given input x in a binary decision setting. This is done to quantify uncertainty in an active learning setting. In this paper, we extend this characterization to N-label settings and use the image-label pairs to extract stochastic surprisal from the network.

Notice the difference in the definitions of action. In FEP, the generative model acts on the outside world creating a change that reduces its surprisal. Our definition in Equation (10) is the same one that is used in I-FGSM (Goodfellow et al., 2014) adversarial generation technique. Equation (10) is continuously applied and a gradient w.r.t. the input, i.e., Adx, is added to x until the prediction changes adversarially. Changing the input would be a true action from the FEP sense. However, in this paper, we do not explicitly change the outside world or x. Rather, we measure the effect of such a change on the network using AdWL without making said change.

2.3. Methodology

We validate the effectiveness of stochastic surprisal during inference on two applications: Image Quality Assessment (IQA) and Robust Classification. The action gradients, Aϕ are used in two ways. The first approach is to use the surprisal gradients as error directions. This is done by projecting images with and without distortions onto the gradient space and comparing them. In this case, the surprisal acts as a measurement between the images and acts as a Full-Reference IQA metric. The second approach is to directly use surprisal gradients as feature vectors. The directional change caused by the actions is dependent on the network, the input and the action class. By keeping the network same across action classes, surprisal becomes a characteristic of the data. This approach is explored for the application of robust classification.

2.3.1. Image quality assessment

Image quality assessment is a field of image processing that objectively estimates the perceptual quality of a degraded image. Multiple methods have been proposed to predict the subjective quality of images (Wang et al., 2003, 2004; Sampat et al., 2009; Ponomarenko et al., 2011; Wang and Li, 2011; Zhang et al., 2011; Mittal et al., 2012; Zhang and Li, 2012; Prabhushankar et al., 2017a, 2018; Temel and AlRegib, 2019). All these methods extract structure related hand-crafted features from both reference and distorted images and compare them to predict the quality. Recently, machine learning models directly extract features from images (Temel et al., 2016; Bosse et al., 2017; Prabhushankar et al., 2017b). The authors in Bosse et al. (2017) propose to do so in either the presence or absence of the original pristine image. In Ma et al. (2021), the authors propose a free energy inspired technique to predict the quality. They use a Generative-Adversarial Network as the base perception module and an additional CNN to model content and degradation dependent characteristics. In this paper, we approach the action module in FEP as a function of the perception module itself. We do so by extracting stochastic surprisal from the same perception network. Hence, our method acts as a plug-in on top of existing quality estimators. In this paper, we show quantitative results by plugging-in on top of UNIQUE (Temel et al., 2016) and qualitative results on top of Bosse et al. (2017). We first describe and motivate the usage of UNIQUE for quantitative results.

2.3.1.1. UNIQUE

We choose UNIQUE as the base technique since it follows the generative process described in Section 2.1.2.1 and Figure 1. This allows for the generation of stochastic surprisal from Equation (3) based on the Action in Equation (9). The authors in Temel et al. (2016) train a sparse autoencoder with a one layer encoder and decoder and a sigmoid non-linearity on 100, 000 patches of size 8 × 8 × 3 extracted from ImageNet (Deng et al., 2009) testset. The autoencoder is trained with MSE reconstruction loss. This network is f(·) from Equation (3). UNIQUE follows a full reference IQA workflow which assumes access to both reference and distorted images while estimating quality. The reference and distorted images are converted to YGCr color space and converted to 8 × 8 × 3 patches. These patches are mean subtracted and ZCA whitened before being passed through the trained encoder. The activations of all reference patches in the latent space are extracted and concatenated. Activations lesser than a threshold of 0.025 are suppressed to 0. The choice of threshold 0.025 is made based on the sparsity coefficient used during training. Similar procedure is followed for distorted image patches. The suppressed and concatenated features of both the reference and distorted images are compared using Spearman correlation. The resultant is the estimated quality of the distorted image.

2.3.1.2. Proposed methodology

We provide the block diagram for the proposed methodology in Figure 3. Both the pristine and distorted images go through the same pre-processing steps detailed in UNIQUE (Temel et al., 2016) and are projected onto the stochastic surprisal gradients of the decoder. The gradients Agϕ are extracted by backpropagating Equation (9). In this paper, we use the same hyperparameters β = 3, λ = 3e−3, and ρj = 0.035 as used in Temel et al. (2016). Once projected, the resultant is passed through an inverse sigmoidal layer to obtain the latent representation. Note that the latent representation is zg for the pristine image and zg for the distorted image. Once passed through the inversion layer, both the magnitude and phase of each latent representation is concatenated and their spearman correlation coefficient is taken to estimate the quality score of the image.

FIGURE 3
www.frontiersin.org

Figure 3. Block diagram of the proposed framework of IQA as a plug-in on top of Temel et al. (2016).

2.3.2. Robust classification

The goal is to characterize an image x using all N actions. Consider an image x whose class as predicted by fθ(·) is ŷ. Stochastic surprisal of x against class 1 is provided by backpropagating a loss between ŷ and 1 and obtaining corresponding gradients. The gradient is proportional to Ad(ŷ,1;WL), where W is the weight parameters and 1 is the action class. Specifically, it is WLAd(ŷ,1;WL) for weights in layer L and class i ∈ [1, N]. We backpropagate over all N classes to obtain the overall surprisal features across all classes. The final feature, rx for an image x, is given by concatenating all individual features and rx is characteristic of image x. Hence,

ri=(WLAd(ŷ,i;WL))),i[1,N],rx=[r1,r2rN].    (12)

Given a trained feed-forward network f(·) and image x, we extract gradients using Equation (12) which serve as our features. Gradients as features are used in diverse applications including visual explanations (Selvaraju et al., 2017; Prabhushankar et al., 2020; Prabhushankar and AlRegib, 2021b), adversarial attacks (Goodfellow et al., 2014), anomaly detection (Kwon et al., 2020), and image quality assessment (Kwon et al., 2019) among others. In this work, we use gradients as features to characterize data.

2.3.2.1. MLP [H(·)]

Once rx is obtained for all N classes, the surprisal feature is now analogous to zd from Equation (5). However, rx is of dimensionality (N×dL-1)×1 since it is a concatenation of N gradients. To account for the larger dimension size, we classify rx by training an MLP H(·) on top of rx derived from training data. In this paper, we use a simple three layered MLP as H(·) with sigmoid activations. The exact structure of the MLP is dependent on dL−1 of the base f(·) network and is given in Table 1 for ResNets 18,34,50, and 101 (He et al., 2016) that are considered in Section 3.

TABLE 1
www.frontiersin.org

Table 1. Structure of H(·) for different ResNet architectures as f(·).

2.3.2.2. Training H(·)

The concatenated rx features for all training data are extracted and normalized. H(·) is trained on all training rx using the same training procedure as the perception network f(·). H(·) is trained for 200 epochs with SGD optimizer and cross-entropy loss, momentum = 0.9, weight decay = 5e−4, and adaptive learning rates of 0.1, 0.02, 0.004 changed at epochs 60, 120, 160, respectively.

2.3.2.3. Testing using f(·) and H(·)

During test time, the proposed framework operates in three steps. In step 1, as shown in Equation (13), the given image passes through the perception network to provide a coarse estimation ŷ. In step 2, the stochastic surprisal features rx are extracted according to Equation (14) and concatenated. In step 3, rx is normalized and passed through the MLP H(·) to obtain the final prediction ỹ. This is shown in Equation (15).

ŷ=argmaxf(x),    (13)
rx=[(WLMSE(ŷ,δii)),i[1,N]],    (14)
=H(rx),    (15)

Note that we substituted Ad in Equation (14) with the MSE formulation of action from Equation (10).

3. Results

3.1. Image quality assessment

We report the results of the our proposed method in comparison with commonly cited methods in this section. We first discuss the datasets used for comparison as well as the evaluation metrics. We finally show the results in Table 2 and discuss these results.

TABLE 2
www.frontiersin.org

Table 2. Overall performance of image quality estimators.

3.1.1. Datasets

We compare our proposed quality estimation technique on three datasets—MULTI-LIVE (Jayaraman et al., 2012), TID2013 (Ponomarenko et al., 2015), and DR IQA (Athar and Wang, 2023). We choose MULTI-LIVE and TID2013 datasets for two reasons. Firstly, our proposed technique is a plug-in approach on top of an existing technique (Temel et al., 2016). Hence, it is imperative to compare against and show results on datasets that were used in Temel et al. (2016). Secondly, the two datasets provide access to seven categories of distortion among five levels. This is useful in comparison against the recognition experiments discussed in Section 3.2 which follows a similar setup. The complex distortions can either be a combination of multiple distortions such as distortions generated in the MULTI-LIVE dataset (Jayaraman et al., 2012) or the human visual system (HVS) specific peculiar distortions such as the ones presented in the TID2013 (Ponomarenko et al., 2015) dataset. A more challenging scenario is presented in DR IQA dataset, where the authors conjecture a degraded reference setting for image quality assessment. In this setting, pristine images are unavailable as a reference. Instead, singly distorted images are used as reference to construct IQA metrics for multiply distorted images. In Table 2, we provide results for DR IQA dataset as DRv1 and DRv2 based on the author's division of the dataset. Each of DRv1 and DRv2 have 31, 790 multiply distorted images and 1, 122 singly distorted images. Additionally, this dataset does not have true subjective quality scores from humans but is derived from a synthetic quality benchmark. This synthetic score uses existing Full Reference metrics for quality generation including some of comparisons in Table 2.

3.1.2. Evaluation metrics

The performance is validated using outlier ratio (consistency), root mean square error (accuracy), Pearson correlation (linearity), Spearman correlation (rank), and Kendall correlation (rank). Arrows next to each metric in Table 2 indicate the desirability of a higher number (↑) or a lower number (↓). Statistical significance between correlation coefficients is measured with the formulations suggested in ITU-T Rec. P.1401 (ITU-T, 2012) and provided below each correlation coefficient. A 0 value corresponds to statistically similar performance, -1 means the method is statistically inferior to proposed method, and 1 indicates that the method is statistically superior to proposed method. Two best performing methods for each metric are highlighted.

3.1.3. Results

We compare our proposed stochastic surprisal-based UNIQUE against other image quality estimators based only on handcrafted features and perception pipeline in Table 2. These compared full reference estimators include PSNR-HA (Ponomarenko et al., 2011), SSIM (Wang et al., 2004), MS-SSIM (Wang et al., 2003), CW-SSIM (Sampat et al., 2009), IW-SIM (Wang and Li, 2011), SR-SIM (Zhang and Li, 2012), FSIM (Zhang et al., 2011), FSIMc (Zhang et al., 2011), PerSIM (Temel and AlRegib, 2015), CSV (Temel and AlRegib, 2016b), UNIQUE (Temel et al., 2016). We also compare against no reference metrics including BRISQUE (Mittal et al., 2012), BIQI (Moorthy and Bovik, 2010), and BLIINDS2 (Saad et al., 2012). All these techniques were also compared against the base UNIQUE algorithm in Temel et al. (2016). In addition to these, we compare against new estimators including COHERENSI (Temel and AlRegib, 2019) and SUMMER (Temel and AlRegib, 2019). SUMMER beats UNIQUE among six of the 10 categories. Note that we do not show results for BRISQUE, BIQI, and BLIINDS2 for DR IQA dataset since NR methods, that are generally trained on singly distorted images, exhibit a large performance gap on multiply distorted images (Athar and Wang, 2023).

The proposed stochastic surprisal-based method plugs on top of UNIQUE and its results are provided under the last column in Table 2. It is always in the top two methods for MULTI-LIVE and TID2013 datasets in all evaluation metrics. In particular, the proposed method achieves the best performance for all the categories except in OR and KRCC in TID2013 dataset. UNIQUE, by itself, does not achieve the best performance for any of the metrics in MULTI dataset. However, the same network using the proposed gradient features significantly improves the performance and achieves the best performance on all metrics. For instance, UNIQUE is the third best performing method in MULTI dataset in terms of RMSE, PLCC, SRCC, and KRCC. However, the action-based features improve the performance for those metrics by 1.315, 0.036, 0.020, and 0.023, respectively and achieve the best performance for all metrics. This further reinforces the plug-in capability of the proposed method during inference. On DR IQA dataset, FSIM and FSIMc perform the best across all categories. The authors in Athar and Wang (2023) used FSIMc to construct DR IQA models. However, the proposed algorithm remains competitive among all evaluation metrics. The results are statistically significant in 53 of the 78 compared metrics across both DRv1 and DRv2. Note that a number of these compared FR-IQA metrics have been utilized to construct the synthetic ground truth quality scores.

3.2. Robust classification

Neural networks are sensitive to distortions in test that the network was not privy to during training (Temel et al., 2017, 2018; Hendrycks and Dietterich, 2019). These distortions include image acquisition errors, environmental conditions during acquisition, transmission, and storage errors among others. CIFAR-10C (Hendrycks and Dietterich, 2019) dataset consists of 19 real world distortions each of which has five levels of degradation that distort the 10, 000 images in CIFAR-10 testset. Neural networks that use perception-only mechanics suffer performance accuracy drops on CIFAR-10C. Current techniques that alleviate the drop in perception-only accuracy require additional training data. The authors in Vasiljevic et al. (2016) show that finetuning or retraining networks using distorted images increases the performance of classification under the same distortion. However, performance between different distortions is not generalized well. For instance, training on gaussian blurred images does not guarantee a performance increase in motion blur images (Geirhos et al., 2018b). Other proposed methods include training on style-transferred images (Geirhos et al., 2018a), training on adversarial images (Hendrycks and Dietterich, 2019), training on simulated noisy virtual images (Temel et al., 2017), and self-supervised methods like SimCLR (Chen et al., 2020) that train by augmenting distortions. Augmix (Hendrycks et al., 2019) creates multiple chains of augmentations to train the base network. All these works require additional training data. Our proposed stochastic surprisal-based technique is a plug-in on top of any existing method that increases the base network's robustness to distortions without any need for new data.

3.2.1. Experimental setup and dataset

We use CIFAR-10C (Hendrycks and Dietterich, 2019) as our dataset of choice with all its 95 distortions and degradation levels. ResNet-18,34,50, and 101 (He et al., 2016) architectures are used as the base f(·) perception-only networks. These are trained from scratch on CIFAR-10 dataset. Following the terminologies established in Section 2, X is the training set of CIFAR-10 and X are the 19 distorted domains in which the testing set of CIFAR-10C reside. Each of the 19 corruptions have five levels of distortions. Higher the level, higher is the distortion. The distortions include blur characteristics like gaussian blur, zoom blur, glass blur, and environmental distortions like rain, snow, fog, haze among others.

3.2.2. Comparison against existing state of the art methods

In Table 3, we compare the Top-1 accuracy between perception-only inference and our proposed stochastic surprisal-based inference. All the state-of-the-art techniques require additional training data—noisy images (Vasiljevic et al., 2016), adversarial images (Hendrycks and Dietterich, 2019), self-supervision SimCLR augmentations (Chen et al., 2020), and augmentation chains (Hendrycks et al., 2019). We term these perception-only techniques as f′(·) and we actively infer on top of them. For all f′(·) other than Augmix, the base network is a ResNet-18. For Augmix, we use WideResNet architecture following the authors in Hendrycks et al. (2019). Another commonly used robustness technique is to pre-process the noisy images to denoise them. Denoising 19 distortions is, however, not a viable strategy assuming that the characteristics of the distortions are unknown. We use Non-Local Means (Buades et al., 2011) denoising and the results obtained are lower than the perception-only accuracy by almost 3%. However, the proposed technique on this model increases the results by 3.84%. We create untargeted adversarial images using FGSM attack (Goodfellow et al., 2014) and use them to train a ResNet-18 architecture. In the experimental setup of augmenting noise (Vasiljevic et al., 2016), we augment the training data of CIFAR-10 with six distortions provided by Temel et al. (2018) to randomly distort 500 CIFAR-10 training images to train f′(·). For all techniques, the proposed technique plugs on top of f′(·) and increases the accuracy to create robust networks. Note that in all the perception-only methods in Table 3, we do not use the augmented data to train H(·). The gain obtained is by creating actions on only the undistorted data. Even when the augmented network f′(·) gains on non-augmented f(·), the proposed technique plugs on top of f′(·) to provide additional gains.

TABLE 3
www.frontiersin.org

Table 3. Stochastic surprisal-based plug-in on top of existing robustness techniques.

3.2.3. Analyzing distortion-wise accuracy gains

The results of all four ResNet architectures for each of the 19 distortions is shown in Figure 4. X-Axis in each plot shows 19 distortions averaged over all 5 distortion levels. Y-Axis shows Top-1 accuracy. The bars in blue show perception-only inference results and the red region in each bar represents the performance gain obtained by stochastic surprisal-based inference. There is an increase in performance across distortions and networks. In 9 of the 19 distortions, the proposed method averages 4% more than its perception-only counterpart. These include gaussian blur, gaussian noise, glass blur, impulse noise, motion blur, pixelate, shot noise, speckle noise, and zoom blur. The highest increase is 8.22% for glass blur. In 2 of the distortions, brightness and saturate, the results increase by < 0.4% averaged over all levels. This is because of the statistics that the distortions affect. Distortions can change either the local or global statistics within images. Distortions like saturate, brightness, contrast, fog, and frost change the low level or global statistics in the image domain. Neural networks are actively trained to ignore such changes so that their effects are not propagated beyond the first few layers. Hence, gradients derived from the final fully connected layer do not capture the necessary changes required within f(·) to compensate for these distortions. Therefore, both the proposed and perception-only inference follow each other closely in distortions like brightness and saturate.

FIGURE 4
www.frontiersin.org

Figure 4. Visualization of accuracy gains (in red) of using the proposed stochastic surprisal-based inference over perception-only inference (in red) on CIFAR-10C dataset (Hendrycks and Dietterich, 2019) for four networks across 19 distortions.

3.2.4. Level-wise recognition on CIFAR-10C

In Figure 5B, the proposed performance gains for the four networks are categorized based on the distortion levels. All 19 categories of distortion on CIFAR-10C are averaged for each level and their respective perception-only accuracy and stochastic surprisal-based gains are shown. Note that the levels are progressively more distorted. Hence, level 1 distribution X is similar to the training distribution X when compared to level 5 distributions. As the distortion level increases, the proposed method's accuracy gains also increase. This is because, with a larger distributional shift, more characteristic is the action required w.r.t. the network parameters. In Figure 5A, we show the distortion-wise and level-wise accuracy gains for each network. Note that, a stochastic surprisal-based ResNet-18 performs similarly to a perception-only ResNet-50.

FIGURE 5
www.frontiersin.org

Figure 5. Visualization of accuracy gains (in red) of using the proposed stochastic surprisal-based inference over perception-only inference (in blue) on CIFAR-10C dataset (Hendrycks and Dietterich, 2019) for four networks (A) averaged across 19 distortions and 5 levels (B) shown across 5 levels of distortion.

4. Discussion

We conclude this paper by considering the terminology of stochastic surprisal as well as some of the broader implications of the proposed technique. These include the abductive reasoning module and expectancy-mismatch hypothesis in cognitive science.

4.1. Choice of the terminology of stochastic surprisal

We motivate the terminology of stochastic surprisal in two ways:

• As an analogy to gradient descent and stochastic gradient descent: Gradient descent requires the gradients from the all available training data to update the weights. However, since this is computationally infeasible for large neural networks, stochastic gradient descent allows using a single training datapoint to estimate gradients, repeated across all data. In stochastic surprisal, we use the single data point, available at inference, under all allowable actions to estimate surprisal.

• Meaning of stochastic: The word stochastic implies some randomness within the setting. This randomness is derived from the possible set of all actions. In discriminative networks in Equation (10), ai, i ∈ [1, N] is the set of all possible actions with N being the number of trained classes. This suggests that we allow a datapoint to be any available class, all of which are equally likely. Similarly, in generative networks in Equation (9), we add random perturbations at the output of the autoencoder. Hence, there is an inherent randomness within the actions that allow for the usage of the word stochastic.

4.2. Abductive reasoning

The free energy principle postulates that the brain encodes a Bayesian recognition density that predicts sensory data based upon some hypotheses about their causes. This mode of inference is called inference to the best explanation. The underlying reasoning model is abductive reasoning. Abductive reasoning was introduced by the philosopher Peirce (1931), who saw abduction as a reasoning process from effect to cause (Paul, 1993). An abductive reasoning framework creates a hypothesis and tests its validity without considering the cause. A hypothesis can be considered as an answer to one of the three following questions: a causal “Why P?” question, a counterfactual “What if?” question, and a contrastive “Why P, rather than Q?” question (AlRegib and Prabhushankar, 2022). Here P is the prediction and Q is any contrast class. The action considered in this paper is the latter contrastive question of the form “Why P, rather than Q?”. Stochastic surprisal measures the answer to this question. We explore this further in Prabhushankar et al. (2020) and AlRegib and Prabhushankar (2022). We borrow the visualization procedure from Prabhushankar et al. (2020) to visually analyze stochastic surprise in the applications of IQA and recognition in Figure 6. We do so to illustrate the broader impact of action at inference time. As in Section 2.3.1.2, we use stochastic surprisal as a plug-in approach.

FIGURE 6
www.frontiersin.org

Figure 6. Stochastic surprisal answers contrastive questions. (A) Grad-CAM and contrastive explanations for the application of image quality assessment and (B) Grad-CAM and contrastive explanations for the application of recognition. The highlighted regions in each image provide a visual explanation to the question beneath it. While Grad-CAM (Selvaraju et al., 2017) shows all the perceived regions in the image, the stochastic surprisal provides fine-grained answers to contrastive questions. Best viewed in color.

For IQA visualizations, we use a trained full-reference metric DIQaM-FR Bosse et al. (2017) as our perception model. In Figure 6A, the pretrained network from Bosse et al. (2017) provides a quality score of 0.58 to the distorted lighthouse image. Here 0.58 acts as P in the contrastive question. We use MSE loss function as Ad and a real number Q ∈ [0, 1] to calculate stochastic surprisal. Contrastive explanations of Q values including 0.25, 0.75, and1 along with Grad-CAM results are shown in Figure 6A. Grad-CAM highlights the entire image indicating that the network estimates the quality based on the whole image. While this builds trust in the network, it does not help us understand the network decision. The stochastic surprisal, however, provides fine-grained explanations. Consider the contrastive questions asking why the quality is neither 1nor0.75. The network estimates this to be primarily due to distortions concentrating in the foreground portion of the image. This explanation is inline with previous works in IQA that posit that distortions in the more salient foreground or edge features cause a larger drop in perceptual quality than that in color or background (Chandler, 2013; Prabhushankar et al., 2017b). When the contrastive question asks why the prediction is not 0.25, the network highlights the sky indicating its good quality for a higher score of 0.58.

Figure 6B shows the contrastive questions answered by the stochastic surprisal for the application of recognition. Given an image of a spoonbill from ImageNet dataset (Deng et al., 2009), a VGG-16 network highlights the body, feathers, legs and beak of the bird in the Grad-CAM (Selvaraju et al., 2017) explanation. Consider a more fine grained contrastive question regarding the difference between a spoonbill and flamingo. The stochastic surprisal highlights regions in the neck of the spoonbill indicating that the contrast between the input spoonbill image and the network's notion of a flamingo lies in the spoonbill's lack of S-shaped neck. Similarly, the contrast between a spoonbill and a crane is in the color of the spoonbill's feathers. The contrast between a pig and a spoonbill is in the shape of neck and legs in the spoonbill which is emphasized. All these visualizations serve to illustrate the stochastic nature of the proposed method. It is stochastic in the sense that it individually depends on the network, the data, as well as the action. In this case, the action of not predicting a flamingo has a different explanation compared to the action of not predicting a pig.

4.3. Expectancy-mismatch

The expectancy-mismatch hypothesis in cognitive science is a way to quantify and analyze human attention. According to this hypothesis, human attention mechanism suppresses expected messages and focuses on the unexpected ones (Horstmann, 2002; Summerfield and Egner, 2009; Becker and Horstmann, 2011; Krebs et al., 2012; Horstmann et al., 2016; Sun et al., 2020). Becker and Horstmann (2011) shows that a message which is unexpected, captures human attention. Then, the human visual system establishes whether the input matches the observers' expectation. If they are conflicting, error neurons in the human brain encode the prediction error and pass the error message back to the representational neurons. The proposed method uses gradients with respect to the network parameters to measure an action. In both the generative and discriminative networks, this action takes the form of a change in the output thereby creating a mismatch with the network's expected result. Hence, the proposed method can act as a framework for exploring expectancy-mismatch in future works.

4.4. Related learning paradigms

The proposed stochastic surprisal decomposes the decision making and training process of a neural network into perception and action phases. A number of other machine learning paradigms including continual and lifelong learning (Parisi et al., 2019), online learning (Hoi et al., 2021), and introspective learning (Prabhushankar and AlRegib, 2022) also have multiple stages. Online learning assumes an exploration and exploitation stage in a neural network's training process. Hence, the differentiation in the training stages is based on time rather than the proposed action. Continual and lifelong learning is a research paradigm that tackles the topic of catastrophic forgetting when a neural network is trained to perform multiple tasks. Introspective learning conjectures reasons in the form of counterfactual or contrastive questions in its two stages to make predictions. Hence, while there are multiple machine learning paradigms that conjecture decomposition of neural network's training and decision processes, the proposed framework that is based on the FEP is unique in its decomposition. The field of active learning (Benkert et al., 2022; Logan et al., 2022) involves actions within the training and decision making processes. However, active learning requires actions from the users while the considered actions in the proposed methodology are with respect to the neural network.

5. Conclusion

In this paper, we examine supervised learning from the perspective of Free Energy Principle. The learning process of both generative and discriminative models can be decomposed into divergence and surprisal measures. Surprisal is introduced in generative models via regularization and constraints that allow a generative aspect to their functionality. While this complicates the action itself, the set of possible actions is still limited. Discriminative networks follow the traditional route of free energy minimization by defining surprisal in terms of recognition entropy and minimizing it. This allows the action itself to be a simple fidelity-based reconstruction error. However, in discriminative networks, there are N set of possible actions, N being the number of classes in the recognition density. We account for both these peculiarities in defining our action space. We use a fidelity-based MSE loss for both generative and discriminative networks. In addition, generative networks are reinforced with KL-divergence based elastic net regularization, and in discriminative networks we backpropagate N possible actions. We measure this scalar action quantity in terms of a vector quantity called stochastic surprisal that is a function of the network parameters and an individual data point rather than a distribution. We use stochastic surprisal to assess distortions in image quality assessment and disregard distortions in robust recognition. We then discuss the implications of stochastic surprisal in other areas of cognitive science including abductive reasoning and expectancy-mismatch. A computational bottleneck within the framework is the consideration of all N possible actions to estimate the surprisal feature rx. rx scales linearly with N thereby becoming prohibitive on datasets with a large number of classes. Selecting only a subset of the most likely actions is one plausible solution to the challenge of scalability.

Data availability statement

Publicly available datasets were analyzed in this study. This data can be found at: CIFAR-10C: https://zenodo.org/record/2535967; CIFAR-10: https://www.cs.toronto.edu/~kriz/cifar.html; TID2013: https://www.ponomarenko.info/tid2013.htm; MULTI-LIVE: http://live.ece.utexas.edu/research/Quality/live_multidistortedimage.html.

Author contributions

MP wrote the first draft of the manuscript. GA and MP had multiple rounds of revisions and approved the submitted version. All authors contributed to the article and approved the submitted version.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

AlRegib, G., and Prabhushankar, M. (2022). “Explanatory paradigms in neural networks: Towards relevant and contextual explanations,” in IEEE Signal Processing Magazine, Vol. 39 (IEEE), 59–72. doi: 10.1109/MSP.2022.3163871

CrossRef Full Text | Google Scholar

Athar, S., and Wang, Z. (2023). “Degraded reference image quality assessment,” in IEEE Transactions in Image Processing. doi: 10.1109/TIP.2023.3234498

CrossRef Full Text | Google Scholar

Becker, S. I., and Horstmann, G. (2011). Novelty and saliency in attentional capture by unannounced motion singletons. Acta Psychol. 136, 290–299. doi: 10.1016/j.actpsy.2010.12.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Benkert, R., Prabhushankar, M., and AlRegib, G. (2022). “Forgetful active learning with switch events: efficient sampling for out-of-distribution data,” in 2022 IEEE International Conference on Image Processing (ICIP), 2196–2200. doi: 10.1109/ICIP46576.2022.9897514

CrossRef Full Text | Google Scholar

Bosse, S., Maniry, D., Müller, K.-R., Wiegand, T., and Samek, W. (2017). Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. 27, 206–219. doi: 10.1109/TIP.2017.2760518

PubMed Abstract | CrossRef Full Text | Google Scholar

Buades, A., Coll, B., and Morel, J.-M. (2011). Non-local means denoising. Image Process. On Line 1, 208–212. doi: 10.5201/ipol.2011.bcm_nlm

CrossRef Full Text | Google Scholar

Buckley, C. L., Kim, C. S., McGregor, S., and Seth, A. K. (2017). The free energy principle for action and perception: a mathematical review. J. Math. Psychol. 81, 55–79. doi: 10.1016/j.jmp.2017.09.004

CrossRef Full Text | Google Scholar

Chandler, D. M. (2013). Seven challenges in image quality assessment: past, present, and future research. ISRN Signal Process. 2013, 905685. doi: 10.1155/2013/905685

CrossRef Full Text | Google Scholar

Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. (2020). A simple framework for contrastive learning of visual representations. arXiv [Preprint] arXiv:2002.05709.

Google Scholar

Demekas, D., Parr, T., and Friston, K. J. (2020). An investigation of the free energy principle for emotion recognition. Front. Comput. Neurosci. 14, 30. doi: 10.3389/fncom.2020.00030

PubMed Abstract | CrossRef Full Text | Google Scholar

Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009). “Imagenet: a large-scale hierarchical image database,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition, 248–255. doi: 10.1109/CVPR.2009.5206848

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. (2009). The free-energy principle: a rough guide to the brain? Trends Cogn. Sci. 13, 293–301. doi: 10.1016/j.tics.2009.04.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. (2019). A free energy principle for a particular physics. arXiv [Preprint] arXiv:1906.10184.

Google Scholar

Gedeon, T., and Harris, D. (1992). “Progressive image compression,” in [Proceedings 1992] IJCNN International Joint Conference on Neural Networks, Vol. 4, 403–407. doi: 10.1109/IJCNN.1992.227311

CrossRef Full Text | Google Scholar

Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F. A., and Brendel, W. (2018a). “ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness,” in International Conference on Learning Representations.

Google Scholar

Geirhos, R., Temme, C. R., Rauber, J., Schütt, H. H., Bethge, M., and Wichmann, F. A. (2018b). “Generalisation in humans and deep neural networks,” in Advances in Neural Information Processing Systems, 7538–7550.

Google Scholar

Geisler, W. S. (2008). Visual perception and the statistical properties of natural scenes. Annu. Rev. Psychol. 59, 167–192. doi: 10.1146/annurev.psych.58.110405.085632

PubMed Abstract | CrossRef Full Text | Google Scholar

Goodfellow, I. J., Shlens, J., and Szegedy, C. (2014). Explaining and harnessing adversarial examples. arXiv [Preprint] arXiv:1412.6572.

Google Scholar

Gottwald, S., and Braun, D. A. (2020). The two kinds of free energy and the bayesian revolution. PLoS Comput. Biol. 16, e1008420. doi: 10.1371/journal.pcbi.1008420

PubMed Abstract | CrossRef Full Text | Google Scholar

Gu, K., Zhai, G., Yang, X., and Zhang, W. (2014). Using free energy principle for blind image quality assessment. IEEE Trans. Multim. 17, 50–63. doi: 10.1109/TMM.2014.2373812

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770–778. doi: 10.1109/CVPR.2016.90

PubMed Abstract | CrossRef Full Text | Google Scholar

Hendrycks, D., and Dietterich, T. (2019). “Benchmarking neural network robustness to common corruptions and perturbations,” in International Conference on Learning Representations.

PubMed Abstract | Google Scholar

Hendrycks, D., Mu, N., Cubuk, E. D., Zoph, B., Gilmer, J., and Lakshminarayanan, B. (2019). “AugMix: A simple data processing method to improve robustness and uncertainty,” in International Conference on Learning Representations.

Google Scholar

Hinton, G. E., and Zemel, R. (1993). “Autoencoders, minimum description length and Helmholtz free energy,” in Advances in Neural Information Processing Systems, 6.

Google Scholar

Hipólito, I., Ramstead, M. J., Convertino, L., Bhat, A., Friston, K., and Parr, T. (2021). Markov blankets in the brain. Neurosci. Biobehav. Rev. 125, 88–97. doi: 10.1016/j.neubiorev.2021.02.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoi, S. C., Sahoo, D., Lu, J., and Zhao, P. (2021). Online learning: a comprehensive survey. Neurocomputing 459, 249–289. doi: 10.1016/j.neucom.2021.04.112

CrossRef Full Text | Google Scholar

Horstmann, G. (2002). Evidence for attentional capture by a surprising color singleton in visual search. Psychol. Sci. 13, 499–505. doi: 10.1111/1467-9280.00488

PubMed Abstract | CrossRef Full Text | Google Scholar

Horstmann, G., Becker, S., and Ernst, D. (2016). Perceptual salience captures the eyes on a surprise trial. Attent. Percept. Psychophys. 78, 1889–1900. doi: 10.3758/s13414-016-1102-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Hou, X., and Zhang, L. (2007). “Saliency detection: a spectral residual approach,” in 2007 IEEE Conference on Computer Vision and Pattern Recognition, 1–8. doi: 10.1109/CVPR.2007.383267

CrossRef Full Text | Google Scholar

ITU-T (2012). P. 1401: Methods, Metrics and Procedures for Statistical Evaluation, Qualification and Comparison of Objective Quality Prediction Models. Technical report, ITU Telecommunication Standardization Sector.

Google Scholar

Jayaraman, D., Mittal, A., Moorthy, A. K., and Bovik, A. C. (2012). “Objective quality assessment of multiply distorted images,” in Asilomar Conference on Signals, Systems, and Computers, 1693–1697. doi: 10.1109/ACSSC.2012.6489321

CrossRef Full Text | Google Scholar

Kingma, D. P., and Welling, M. (2013). Auto-encoding variational bayes. arXiv:1312.6114.

PubMed Abstract | Google Scholar

Kingma, D. P., and Welling, M. (2019). An introduction to variational autoencoders. arXiv [Preprint] arXiv:1906.02691. doi: 10.1561/9781680836233

CrossRef Full Text | Google Scholar

Krebs, R., Fias, W., Achten, E., and Boehler, C. (2012). “Stimulus conflict and stimulus novelty trigger saliency signals in locus coeruleus and anterior cingulate cortex,” in Frontiers in Human Neuroscience Conference Abstract: Belgian Brain Council. doi: 10.3389/conf.fnhum

CrossRef Full Text | Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). “Imagenet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems, Vol. 25.

Google Scholar

Kwon, G., Prabhushankar, M., Temel, D., and AlRegib, G. (2019). “Distorted representation space characterization through backpropagated gradients,” in 2019 IEEE International Conference on Image Processing (ICIP), 2651–2655. doi: 10.1109/ICIP.2019.8803228

CrossRef Full Text | Google Scholar

Kwon, G., Prabhushankar, M., Temel, D., and AlRegib, G. (2020). “Backpropagated gradient representations for anomaly detection,” in European Conference on Computer Vision (Springer), 206–226. doi: 10.1007/978-3-030-58589-1_13

CrossRef Full Text | Google Scholar

Liu, Y., Gu, K., Zhang, Y., Li, X., Zhai, G., Zhao, D., et al. (2019). Unsupervised blind image quality evaluation via statistical measurements of structure, naturalness, and perception. IEEE Trans. Circuits Syst. Video Technol. 30, 929–943. doi: 10.1109/TCSVT.2019.2900472

CrossRef Full Text | Google Scholar

Liu, Y., Zhai, G., Gu, K., Liu, X., Zhao, D., and Gao, W. (2017). Reduced-reference image quality assessment in free-energy principle and sparse representation. IEEE Trans. Multim. 20, 379–391. doi: 10.1109/TMM.2017.2729020

CrossRef Full Text | Google Scholar

Logan, Y.-y., Prabhushankar, M., and AlRegib, G. (2022). Decal: deployable clinical active learning. arXiv [Preprint] arXiv:2206.10120.

Google Scholar

Ma, J., Wu, J., Li, L., Dong, W., Xie, X., Shi, G., et al. (2021). Blind image quality assessment with active inference. IEEE Trans. Image Process. 30, 3650–3663. doi: 10.1109/TIP.2021.3064195

PubMed Abstract | CrossRef Full Text | Google Scholar

Mao, X., Shen, C., and Yang, Y.-B. (2016). “Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections,” in Advances in Neural Information Processing Systems, 29.

Google Scholar

Mittal, A., Moorthy, A. K., and Bovik, A. C. (2012). No-reference image quality assessment in the spatial domain. IEEE Trans. Image Proc. 21, 4695–4708. doi: 10.1109/TIP.2012.2214050

PubMed Abstract | CrossRef Full Text | Google Scholar

Moorthy, A. K., and Bovik, A. C. (2010). A two-step framework for constructing blind image quality indices. IEEE Signal Proc. Lett. 17, 513–516. doi: 10.1109/LSP.2010.2043888

CrossRef Full Text | Google Scholar

Murray, N., Vanrell, M., Otazu, X., and Parraga, C. A. (2013). Low-level spatiochromatic grouping for saliency estimation. IEEE Trans. Pattern Anal. Mach. Intell. 35, 2810–2816. doi: 10.1109/TPAMI.2013.108

PubMed Abstract | CrossRef Full Text | Google Scholar

Ng, A. (2011). Sparse autoencoder. CS294A Lect. Notes 72, 1–19.

Google Scholar

Parisi, G. I., Kemker, R., Part, J. L., Kanan, C., and Wermter, S. (2019). Continual lifelong learning with neural networks: a review. Neural Netw. 113, 54–71. doi: 10.1016/j.neunet.2019.01.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Paul, G. (1993). Approaches to abductive reasoning: an overview. Artif. Intell. Rev. 7, 109–152. doi: 10.1007/BF00849080

PubMed Abstract | CrossRef Full Text | Google Scholar

Peirce, C. S. (1931). Collected Papers of Charles Sanders Peirce. Harvard University Press.

Google Scholar

Ponomarenko, N., Ieremeiev, O., Lukin, V., Egiazarian, K., and Carli, M. (2011). “Modified image visual quality metrics for contrast change and mean shift accounting,” in Proceedings of CADSM, 305–311.

Google Scholar

Ponomarenko, N., Jin, L., Ieremeiev, O., Lukin, V., Egiazarian, K., Astola, J., et al. (2015). Image database tid2013: Peculiarities, results and perspectives. Signal Process. 30, 57–77. doi: 10.1016/j.image.2014.10.009

CrossRef Full Text | Google Scholar

Prabhushankar, M., and AlRegib, G. (2021a). Contrastive reasoning in neural networks. arXiv [Preprint] arXiv:2103.12329. doi: 10.1109/ICIP40778.2020.9190927

CrossRef Full Text | Google Scholar

Prabhushankar, M., and AlRegib, G. (2021b). “Extracting causal visual features for limited label classification,” in 2021 IEEE International Conference on Image Processing (ICIP), 3697–3701. doi: 10.1109/ICIP42928.2021.9506393

CrossRef Full Text | Google Scholar

Prabhushankar, M., and AlRegib, G. (2022). “Introspective Learning: A two-stage approach for inference in neural networks,” in Advances in Neural Information Processing Systems.

Google Scholar

Prabhushankar, M., Kokilepersaud, K., Logan, Y.-y., Corona, S. T., AlRegib, G., and Wykoff, C. (2022). “OLIVES dataset: Ophthalmic labels for investigating visual eye semantics,” in Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track.

Google Scholar

Prabhushankar, M., Kwon, G., Temel, D., and AIRegib, G. (2018). “Semantically interpretable and controllable filter sets,” in 2018 25th IEEE International Conference on Image Processing (ICIP), 1053–1057. doi: 10.1109/ICIP.2018.8451220

PubMed Abstract | CrossRef Full Text | Google Scholar

Prabhushankar, M., Kwon, G., Temel, D., and AlRegib, G. (2020). “Contrastive explanations in neural networks,” in 2020 IEEE International Conference on Image Processing (ICIP), 3289–3293.

Google Scholar

Prabhushankar, M., Temel, D., and AlRegib, G. (2017a). “Generating adaptive and robust filter sets using an unsupervised learning framework,” in 2017 IEEE International Conference on Image Processing (ICIP), 3041–3045. doi: 10.1109/ICIP.2017.8296841

CrossRef Full Text | Google Scholar

Prabhushankar, M., Temel, D., and AlRegib, G. (2017b). Ms-unique: multi-model and sharpness-weighted unsupervised image quality estimation. Electron. Imaging 2017, 30–35. doi: 10.2352/ISSN.2470-1173.2017.12.IQSP-223

CrossRef Full Text | Google Scholar

Saad, M. A., Bovik, A. C., and Charrier, C. (2012). Blind image quality assessment: a natural scene statistics approach in the DCT domain. IEEE Trans. Image Proc. 21, 3339–3352. doi: 10.1109/TIP.2012.2191563

PubMed Abstract | CrossRef Full Text | Google Scholar

Sampat, M. P., Wang, Z., Gupta, S., Bovik, A. C., and Markey, M. K. (2009). Complex wavelet structural similarity: a new image similarity index. IEEE Trans. Image Proc. 18, 2385–2401. doi: 10.1109/TIP.2009.2025923

PubMed Abstract | CrossRef Full Text | Google Scholar

Sebastian, S., Abrams, J., and Geisler, W. S. (2017). Constrained sampling experiments reveal principles of detection in natural scenes. Proc. Natl. Acad. Sci. U.S.A. 114, E5731–E5740. doi: 10.1073/pnas.1619487114

PubMed Abstract | CrossRef Full Text | Google Scholar

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). “Grad-cam: visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE International Conference on Computer Vision, 618–626. doi: 10.1109/ICCV.2017.74

CrossRef Full Text | Google Scholar

Settles, B., Craven, M., and Ray, S. (2007). “Multiple-instance active learning,” in Advances in Neural Information Processing Systems, 20.

Google Scholar

Shafiq, M., Prabhushankar, M., and AlRegib, G. (2018). “Leveraging sparse features learned from natural images for seismic understanding,” in 80th EAGE Conference and Exhibition 2018 (European Association of Geoscientists & Engineers), 1–5. doi: 10.3997/2214-4609.201800737

CrossRef Full Text | Google Scholar

Shafiq, M. A., Prabhushankar, M., Di, H., and AlRegib, G. (2018). “Towards understanding common features between natural and seismic images,” in SEG Technical Program Expanded Abstracts 2018 (Society of Exploration Geophysicists), 2076–2080. doi: 10.1190/segam2018-2996501.1

CrossRef Full Text | Google Scholar

Summerfield, C., and Egner, T. (2009). Expectation (and attention) in visual cognition. Trends Cogn. Sci. 13, 403–409. doi: 10.1016/j.tics.2009.06.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, Y., Prabhushankar, M., and AlRegib, G. (2020). “Implicit saliency in deep neural networks,” in 2020 IEEE International Conference on Image Processing (ICIP), 2915–2919. doi: 10.1109/ICIP40778.2020.9191186

PubMed Abstract | CrossRef Full Text | Google Scholar

Temel, D., and AlRegib, G. (2015). “PerSIM: multi-resolution image quality assessment in the perceptually uniform color domain,” in IEEE International Conference on Image Processing, 1682–1686. doi: 10.1109/ICIP.2015.7351087

CrossRef Full Text | Google Scholar

Temel, D., and AlRegib, G. (2016a). “Bless: bio-inspired low-level spatiochromatic similarity assisted image quality assessment,” in 2016 IEEE International Conference on Multimedia and Expo (ICME), 1–6. doi: 10.1109/ICME.2016.7552874

CrossRef Full Text | Google Scholar

Temel, D., and AlRegib, G. (2016b). CSV: image quality assessment based on color, structure, and visual system. Signal Process. 48, 92–103. doi: 10.1016/j.image.2016.08.008

CrossRef Full Text | Google Scholar

Temel D. and AlRegib, G.. (2019). Perceptual image quality assessment through spectral analysis of error representations. Signal Process. 48, 92–103. doi: 10.1016/j.image.2018.09.005

CrossRef Full Text | Google Scholar

Temel, D., Kwon, G., Prabhushankar, M., and AlRegib, G. (2017). Cure-TSR: challenging unreal and real environments for traffic sign recognition. arXiv [Preprint] arXiv:1712.02463. doi: 10.1109/ICMLA.2018.00028

PubMed Abstract | CrossRef Full Text | Google Scholar

Temel, D., Lee, J., and AlRegib, G. (2018). “Cure-or: challenging unreal and real environments for object recognition,” in 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), 137–144.

Google Scholar

Temel, D., Prabhushankar, M., and AlRegib, G. (2016). UNIQUE: unsupervised image quality estimation. IEEE Signal Process. Lett. 23, 1414–1418. doi: 10.1109/LSP.2016.2601119

CrossRef Full Text | Google Scholar

Vasiljevic, I., Chakrabarti, A., and Shakhnarovich, G. (2016). Examining the impact of blur on recognition by convolutional networks. arXiv [Preprint] arXiv:1611.05760.

Google Scholar

Wang, Z., Bovik, A. C., Sheikh, H. R., and Simoncelli, E. P. (2004). Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13, 600–612. doi: 10.1109/TIP.2003.819861

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Z., and Li, Q. (2011). Information content weighting for perceptual image quality assessment. IEEE Trans. Image Process. 20, 1185–1198. doi: 10.1109/TIP.2010.2092435

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Z., Simoncelli, E. P., and Bovik, A. C. (2003). “Multiscale structural similarity for image quality assessment,” in Asilomar Conference on Signals, Systems, and Computers, Vol. 2, 1398–1402. doi: 10.1109/ACSSC.2003.1292216

CrossRef Full Text | Google Scholar

Zhai, G., Wu, X., Yang, X., Lin, W., and Zhang, W. (2011). A psychovisual quality metric in free-energy principle. IEEE Trans. Image Process. 21, 41–52. doi: 10.1109/TIP.2011.2161092

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, L., and Li, H. (2012). “SR-SIM: a fast and high performance IQA index based on spectral residual,” in 2012 19th IEEE International Conference on Image Processing, 1473–1476. doi: 10.1109/ICIP.2012.6467149

CrossRef Full Text | Google Scholar

Zhang, L., Zhang, L., Mou, X., Zhang, D., et al. (2011). FSIM: a feature similarity index for image quality assessment. IEEE Trans. Image Process. 20, 2378–2386. doi: 10.1109/TIP.2011.2109730

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: free-energy principle, neural networks, stochastic surprisal, image quality assessment, robust recognition, human visual saliency, abductive reasoning, active inference

Citation: Prabhushankar M and AlRegib G (2023) Stochastic surprisal: An inferential measurement of free energy in neural networks. Front. Neurosci. 17:926418. doi: 10.3389/fnins.2023.926418

Received: 22 April 2022; Accepted: 09 February 2023;
Published: 14 March 2023.

Edited by:

John Jarvis, University of Westminster, United Kingdom

Reviewed by:

Alexandra Psarrou, University of Westminster, United Kingdom
Yutao Liu, Tsinghua University, China

Copyright © 2023 Prabhushankar and AlRegib. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohit Prabhushankar, bW9oaXQucCYjeDAwMDQwO2dhdGVjaC5lZHU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.