Skip to main content

ORIGINAL RESEARCH article

Front. Control Eng., 13 December 2022
Sec. AI and Machine Learning Control

Inference of regulatory networks through temporally sparse data

  • Department of Electrical and Computer Engineering, Northeastern University, Boston, MA, United States

A major goal in genomics is to properly capture the complex dynamical behaviors of gene regulatory networks (GRNs). This includes inferring the complex interactions between genes, which can be used for a wide range of genomics analyses, including diagnosis or prognosis of diseases and finding effective treatments for chronic diseases such as cancer. Boolean networks have emerged as a successful class of models for capturing the behavior of GRNs. In most practical settings, inference of GRNs should be achieved through limited and temporally sparse genomics data. A large number of genes in GRNs leads to a large possible topology candidate space, which often cannot be exhaustively searched due to the limitation in computational resources. This paper develops a scalable and efficient topology inference for GRNs using Bayesian optimization and kernel-based methods. Rather than an exhaustive search over possible topologies, the proposed method constructs a Gaussian Process (GP) with a topology-inspired kernel function to account for correlation in the likelihood function. Then, using the posterior distribution of the GP model, the Bayesian optimization efficiently searches for the topology with the highest likelihood value by optimally balancing between exploration and exploitation. The performance of the proposed method is demonstrated through comprehensive numerical experiments using a well-known mammalian cell-cycle network.

1 Introduction

Gene regulatory networks (GRNs) play an important role in the molecular mechanism of underlying biological processes, such as stress response, DNA repair, and other mechanisms involved in complex diseases such as cancer. The topology inference of GRNs is critical in systems biology since it can generate valuable hypotheses to promote further biological research. Furthermore, a deep understanding of these biological processes is key in diagnosing and treating many chronic diseases. Advances in high-throughput genomic and proteomic profiling technologies have provided novel platforms for studying genomics. Meanwhile, single-cell gene-expression measurements allow capturing multiple snapshots of these complex biological processes. These advances offer an opportunity for seeking systematic approaches to understand the structure of GRNs.

In recent years, Boolean network models have been successfully employed for modeling different biological networks (Wynn et al., 2012; Saadatpour and Albert, 2013; Abou-Jaoudé et al., 2016). More specifically, these Boolean networks have been widely used for inferring GRNs from their state (i.e., gene) data (Pusnik et al., 2022). The state value of genes in the Boolean network is represented by 1 and 0, representing the activation and inactivation of genes, respectively. There are several Boolean network models, including the deterministic Boolean network models, Boolean network with perturbation, probabilistic Boolean network models, and Boolean control networks (Lähdesmäki et al., 2003; Shmulevich and Dougherty, 2010; Cheng and Zhao, 2011). Most of these models account for genes’ stochasticity and can effectively capture the dynamics of GRNs through relatively small times-series data.

Inference of Boolean network models consists of learning the parameters of their models given all the available data. Several advances have been made in the inference of Boolean network models in recent years. These techniques aim to find models that best fit the available time-series data. The fitness criteria are often likelihood or posterior, leading to well-known maximum likelihood and maximum aposteriori inference techniques (Shmulevich et al., 2002; Lähdesmäki et al., 2003). Despite the optimality of these inference techniques, lack of scalability has limited their applications to small GRNs. Several heuristic methods have been developed to scale the inference of Boolean network models; these include scale-free and cluster-based approaches (Hashimoto et al., 2004; Barman and Kwon, 2017), and methods built on evolutionary optimization techniques (Tan et al., 2020; Barman and Kwon, 2018). The former methods aim to build a topology from known seed nodes according to multiple heuristics, whereas the latter ones use evolutionary optimization techniques such as genetic algorithms and particle swarm algorithms for searching over the parameter space. Despite the scalability of these approaches, their incapability to effectively consider the temporal changes in data and efficiently search over possible network models leads to their unreliability in the inference process.

This paper focuses on developing a systematic approach for the inference of GRNs using Boolean network models. Two main challenges in the inference of GRNs are:

Large Topology Candidate Space: The modeling consists of estimating a large number of interacting parameters, which represent the connections between genes that govern their dynamics. This requires searching over a large number of topology candidates and picking the one with the highest likelihood value given the available data. Most existing inference methods for general nonlinear models are developed to deal with continuous parameter spaces, such as maximum likelihood (Johansen et al., 2008; Kantas et al., 2015; Imani and Braga-Neto, 2017; Imani et al., 2020), expectation maximization (Hürzeler and Künsch, 1998; Godsill et al., 2004; Schön et al., 2011; Wills et al., 2013), and multi-fidelity (Imani et al., 2019; Imani and Ghoreishi, 2021) methods. However, these methods cannot be applied for inference over large discrete parameter spaces, such as the large topology candidate space of GRNs. In this paper, we develop a method that is scalable with respect to the number of unknown interactions, and efficiently searches over the large topology candidate space. More specifically, our proposed method enables optimal inference in the presence of a large number of unknown regulations for GRNs with a relatively small number of genes.

Expensive Likelihood Evaluation: The likelihood function, which measures the probability that the available data come from each topology candidate, is often expensive to evaluate. The reasons for that are a large number of genes in GRNs, and the sparsity in the data, which require propagation of the system stochasticity across time and gene states. Given the limitation in the computational resources, evaluation of the likelihood functions for all of the topology candidates is impossible, and one needs to find the topology with the highest likelihood value with a few expensive likelihood evaluations.

This paper derives a scalable topology inference for GRNs observed through temporally sparse data. The proposed framework models the expensive-to-evaluate (log-)likelihood function using a Gaussian Process (GP) regression with a structurally-inspired kernel function. The proposed kernel function exploits the structure of GRNs to efficiently learn the correlation over the topologies, and enables Bayesian prediction of the log-likelihood function for all the topology candidates. Then, a sample-efficient search over topology space is achieved through a Bayesian optimization policy, which sequentially selects topologies for likelihood evaluation according to the posterior distribution of the GP model. The proposed method optimally balances exploration and exploitation, and searches for the global solution without getting trapped in the local solutions. The accuracy and robustness of the proposed framework are demonstrated through comprehensive numerical experiments using a well-known mammalian cell-cycle network.

The remainder of this paper is organized as follows. Section 2 provides a detailed description of the GRN model and the topology inference of GRNs. Further, the proposed topology optimization framework is introduced in Section 3. Section 4 presents various numerical results, and the main conclusions are discussed in Section 5.

2 Preliminaries

2.1 GRN model

This paper employs a Boolean network with perturbation (BNp) model for capturing the dynamics of GRNs (Shmulevich and Dougherty, 2010; Imani et al., 2018; Hajiramezanali et al., 2019). Previously, several works have successfully employed the BNp model for different purposes such as inference (Dougherty and Qian, 2013; Marshall et al., 2007) and classification (Karbalayghareh et al., 2018). This model properly captures the stochasticity in GRNs, coming from intrinsic uncertainty or unmodeled parts of the systems. Consider a GRN consisting of d genes. The state process can be expressed as {Xk; k = 0, 1, …}, where Xk ∈ {0,1}d represents the activation/inactivation state of the genes at time k. The gene state is updated at each discrete time through the following Boolean signal model:

Xk=fXk1nk,(1)

for k = 1, 2, …, where nk ∈ {0,1}d is Boolean transition noise at time k, “⊕” indicates component-wise modulo-2 addition, and f represents the network function.

The network function in Eq. 1 is expressed in component form as f = (f1, …, fd). Each component fi: {0,1}d → {0, 1} is a Boolean function given by:

fix=1,j=1dcijxj+bi>0,0,j=1dcijxj+bi0,(2)

for i = 1, …, d, where cij denotes the type of regulation from component j to component i; it takes +1 and −1 values if there is a positive and negative regulations from component j to component i respectively, and 0 if component j is not an input to component i. bi is a tie-breaking parameter for component i; it takes +12 if an equal number of positive and negative inputs lead to state value +1 and reverse for 12. The network function in Eq. 2 can also be expressed in matrix form as:

fXk1=CXk1+b̄,(3)

where the threshold operator v̄ maps the positive elements of vector v to 1 and negative elements to 0, C is the connectivity matrix with (C)ij = cij in the ith row and jth column, and b=[b1,,bd]T represents the bias vector. A schematic representation of the regulatory network model is shown in Figure 1.

FIGURE 1
www.frontiersin.org

FIGURE 1. The schematic representation of a regulatory network model. The step functions map outputs to 1 if the input is positive, and 0, otherwise.

In Eq. 1, the noise process nk indicates the amount of stochasticity in a Boolean state process. For example, nk(j) = 1, means that the jth gene’s state at time step k is flipped and does not follow the Boolean function. Whereas, nk(j) = 0 indicates that this state is governed by the network function. We assume that all the nk components are independent and have a Bernoulli distribution with parameter (p), which 0 ≤ p < 0.5 refers to the amount of stochasticity in each state variable (i.e., gene).

2.2 Topology inference of regulatory networks

In practice, the network function is unknown or partially known, and the unknown parameters need to be inferred through available data. The unknown information is often the elements of the connectivity matrix or bias units. We assume that L elements of the connectivity matrix {c1, …, cL} are unknown. Given that each element takes in values in space { + 1, 0, −1}, there will be 3L different possible models (i.e., connectivity matrices) denoted by parameter vectors: Θ={θ1,,θ3L}, where θj = [θj (1), …, θj(L)], and θj(i) denotes the type of the ith unknown interaction/parameter under the jth model. It is evident that each parameter vector corresponds to one specific topology/model; therefore, the phrases parameter vector and topology/model are used interchangeably throughout this paper. Further, Cθj represents the connectivity matrix associated with parameter vector θj, while only one parameter vector represents the true underlying system topology. Assuming that D1:T represents the available data, the inference process can be formulated as:

θ*=argmaxθΘPD1:Tθ,(4)

where P (D1:Tθ) is the likelihood function for the topology parameterized by θ. The solution to the optimization problem, θ* in Eq. 4, is known as the maximum likelihood solution. Note that without loss of generality, the proposed method, which will be described in the next section, can be applied to any arbitrary point-based estimator, such as maximum aposteriori.

It should be noted that the unknown parameters could include the bias units in the network model in Eq. 3. Depending on the regulatory network, the bias units are often 12 for the network in normal conditions, whereas a combination of +12 and 12 often represents the network in stress conditions. Therefore, if the network condition is not known, the topology inference could aim to simultaneously estimate the parameters of the connectivity matrix and bias units.

3 The proposed framework

3.1 Likelihood evaluation

Let {x1,,x2d} be an arbitrary enumeration of the possible Boolean state vectors (i.e., a GRN with d components). The available data in D1:T can be represented using the vector I1:T = {I1, …., IT}, where Ik specifies the index associated with the kth state (0 if the state at time step k is missing). For instance, D1:6={X̃2=x9,X̃3=x3,X̃6=x11} contains the information of time steps 2, 3 and 6, and denotes that data at time steps 1, 4 and 5 are missing. In this case, the indicator vector is defined as I1:6 = {0, 9, 3, 0, 0, 11}.

For any given model θΘ, we define the predictive posterior distribution (Πk|k1θ) and posterior distribution (Πk|kθ) of the states as:

Πk|k1θi=PXk=xiI1:k1,θ,Πk|kθi=PXk=xiI1:k,θ,(5)

for i = 1, …, 2d and k = 1, 2, ….

We define the transition matrix Mθ of size 2d × 2d associated with a GRN model parameterized by θ, through the following notation:

Mθij=PXr=xiXr1=xj,θ=Pnr=fθxjxi=pCθxj+b̄xi11pdCθxj+b̄xi1,(6)

for i, j = 1, …, 2d, where the second and third lines in Eq. 6 are obtained based on the GRN model in Eq. 1.

Let Π0|0θ(i)=P(X0=xiθ), for i = 1, …, 2d, be the initial state distribution. If no knowledge about this distribution is available, this can be represented by Π0|0θ(i)=1/2d, for i = 1, …, 2d, and θΘ. The predictive posterior distribution can be computed recursively as:

Πk|k1θ=MθΠk1|k1θ.(7)

The posterior probability of states at time step k can be computed according to predictive posterior and the available data at time step k. If the data at time step k is missing, i.e. Ik = 0, the predictive posterior becomes the posterior, as no data is available at time step k. This can be written as:

Πk|kθj=PXk=xjI1:k,θ=PXk=xjI1:k1,Ik=0,θ=Πk|k1θj,forj=1,,2d.(8)

However, if the ith state is observed at time step k, i.e., Ik = i, then the posterior probability of state at time step k becomes 1 for state i, as full knowledge about Xk = xi is available. The posterior probability in this case can be expressed as:

Πk|kθi=PXk=xiI1:k1,Ik=i,θ=1,Πk|kθj=PXk=xjI1:k1,Ik=i,θ=0,ji.(9)

To summarize, the posterior portability of any state at time k, i.e. Xk = xi can be derived through the following expression:

Πk|kθi=Πk|k1θiifIk=0,1ifIk=i,0otherwise(10)

for i = 1, …, 2d and k = 1, 2, ….

The likelihood value in optimization problem in Eq. 4 can be written in logarithmic format as:

LθlogPD1:Tθ=logPI1:Tθ=k=1TlogPIkI1:k1,θ,(11)

where

PIkI1:k1,θ=Πk|k1θIkif Ik01otherwise.(12)

The computation of the log-likelihood value for any given topology can be huge due to the large size of the transition matrices with 22d elements. The computational complexity of log-likelihood evaluation is of order O (22dT), where T is the time horizon. This substantial computational burden (especially in systems with a large number of components) is the motivation to come up with more efficient ways to solve the problem presented in Eq. 4.

3.2 Bayesian optimization for topology optimization

This article proposes a Bayesian optimization approach for scalable topology inference of regulatory networks observed through temporally sparse data. Bayesian optimization (BO) (Frazier, 2018) is a well-known approach that has been extensively used in recent years for optimization problems in domains with expensive to evaluate objective functions. BO has shown great promise in increasing the automation and the quality of the optimization tasks (Shahriari et al., 2016). In this paper, we are dealing with an expensive-to-evaluate likelihood function. A major issue in employing the conventional BO is its ability of dealing with continuous search spaces, whereas the search space in our problem is the topology of regulatory networks, which takes a large combinatorial space. Therefore, some key changes need to be applied to the original BO formulation so that it can be adapted to our problem. The main concepts of this approach are explained in detail in the following paragraphs.

3.2.1 GP Model over the Log-Likelihood Function

The transition matrix (Mθ) in Eq. 7 makes the log-likelihood function evaluation in Eqs. 4, 11 computationally expensive, especially when dealing with large scale regulatory networks. Therefore, it is vital to come up with an efficient way of searching over the topology space. In this article, the log-likelihood function L (.) is modeled using the Gaussian Process (GP) regression. The GP (Rasmussen and Williams, 2006) is mostly defined over continuous spaces, primarily due to the possibility of defining kernel functions that model the correlation over continuous spaces. In our case, the parameters are discrete interactions (i.e., parameters of the connectivity matrix that take +1, 0, or −1), which prevent constructing the GP model for representing the log-likelihood function over topology space.

This paper takes advantage of the topology structure of GRNs, encoded in connectivity matrix in (3), and defines the following GP model:

Lθ=GPμθ,kθ,θ,(13)

where μ(.) shows the mean function, and k (.,.) indicates the topology-inspired kernel function. The mean function, μ(.), in Eq. 13 represents the prior shape of the log-likelihood function over all the topologies. One possible choice for the mean function is the constant mean function. This mean function carries a single hyperparameter, which can be learned along with the kernel hyperparameters.

Knowing that each parameter vector θ corresponds to a connectivity matrix Cθ, the structurally-inspired kernel function is defined as:

kθ,θ=σf2expCθCθ2l,(14)

where ‖V2 is the sum of squares of elements of V, Cθ and Cθ′ represent the connectivity matrices related to topologies θ and θ′ respectively, l is the length-scale, and σf2 is the scale factor hyperparameters. These hyperparameters quantify how close the topologies are to each other. The more similar two topologies are (i.e., less difference in the connectivity matrices), the more they are correlated, and the kernel function value will be higher for them. While, for more distinct topologies, the kernel will have smaller values.

Figure 2 represents an example of a few possible topologies for a GRN with two genes. These four possible topologies differ in one or two interactions. If the log-likelihood value for topology θ1 is calculated, this information can be used for predicting log-likelihood values for other topologies. The connectivity matrices for these topologies can be expressed as:

Cθ1=0110,Cθ2=0010,Cθ3=0110,Cθ4=0110.(15)

FIGURE 2
www.frontiersin.org

FIGURE 2. An example of possible models (i.e, topologies) for a GRN with two genes.

The correlation between topology θ1 and all the aforementioned topologies, Θsub = {θ1, θ2, θ3, θ4} are calculated based on Eq. 14, and expressed through the following kernel vector:

Kθ1,Θsub=kθ1,θ1kθ1,θ2kθ1,θ3kθ1,θ4=σf2σf2exp1σf2exp4σf2exp16,(16)

where the length-scale hyperparameter is assumed to be 1. It can be seen that topology θ1 has the maximum correlation with itself, and the correlation rate decreases when we move from topology θ1 to θ4. This can also be understood in terms of the differences between the interacting parameters, expressed in the connectivity matrices in Eq. 15. Topology θ2 is different from θ1 in only missing interaction from gene 2 to gene 1. This results in a correlation of k(θ1,θ2)=σf2exp(1) between these two topologies. Further, in Eq. 15 we can see that the interaction from gene 2 to gene 1 in model θ1 is suppressive (−1), whereas the same interaction is activating (+1) in model θ3. This leads to smaller correlation between topologies θ1 and θ3, k(θ1,θ3)=σf2exp(4), in comparison to the correlation between topologies θ1 and θ2. Finally, in Eq. 15, it can be seen that θ1 and θ4 have two opposite types of interactions, leading to the k(θ1,θ4)=σf2exp(16), which is the smallest correlation between θ1 and all the other topologies.

The GP model has the capability of providing the Bayesian representation of the likelihood function across the topology space. Let θ1:t = (θ1, …, θt) be the first t samples from the parameter space (i.e., samples from the topology candidates) with the associated log-likelihood values L1:t=[L1,,Lt]T (i.e., L1 = L (θ1) in Eq. 11). The posterior distribution of L(θ) in Eq. 13 is derived as:

Lθθ1:t,L1:tNμθt,Σθt,(17)

where μθt and Σθt are the mean and variance for a specific model θΘ respectively. These values can be obtained as:

μθt=μθ+Kθ,θ1:tKθ1:t,θ1:t1L1:tμθ1:t,Σθt=kθ,θKθ,θ1:tKθ1:t,θ1:t1Kθ,θ1:tT,(18)

where μ(θ1:t)=[μ(θ1),,μ(θt)]T, and

KΘ,Θ=kθ1,θ1kθ1,θrkθl,θ1kθl,θr,(19)

for Θ = {θ1, …, θl}, Θ={θ1,,θr}. Using the aforementioned formulation, the GP constructs the log-likelihood function as a zero-mean Bayesian surrogate model with covariance k (.,.). Further, at iteration t, the log-likelihood function can be computed by employing the already chosen and evaluated log-likelihood values for topologies θ1:t, i.e., L1:t. The uncertainty of the surrogate model will be reduced as we evaluate the likelihood function for more topologies.

The GP hyperparameters, which consist of the hyperparameters of the topology-inspired kernel function and the mean function, can be learned by optimizing the marginal likelihood function of the GP model at each iteration through:

L1:tθ1:tNμθ1:t,Kθ1:t,θ1:t.(20)

3.2.2 Sequential Topology Optimization

The notion of efficient topology optimization is to come up with an efficient way of searching over all the topology space so that we utilize a minimum number of computationally expensive likelihood evaluations and eventually find the optimal topology, which yields the largest likelihood value.

As mentioned in Section 3.1, evaluation of the log-likelihood function for each topology is a computationally expensive task. Therefore, in here the sample-efficient and sequential topology selection is achieved as:

θt+1=argmaxθΘαtθ,(21)

where αt(θ) represents the acquisition function in the Bayesian optimization context, which is determined over the GP model posterior at iteration t. Multiple acquisition functions exist in the context of Bayesian optimization. For instance, probability improvement (Shahriari et al., 2016) is one of the most traditional acquisition functions, which makes selections to increase the likelihood of improvement in each iteration of BO. Other examples for acquisition functions include expected improvement (Mockus et al., 1978; Jones et al., 1998; Brochu et al., 2010), upper confidence bound (Auer, 2003), knowledge gradient (Wu et al., 2017; Frazier, 2009), and predictive entropy search (Henrández-Lobato et al., 2014). In this work, we use the expected improvement acquisition function, which is the most commonly used acquisition function. This acquisition function balances the exploration and exploitation trade-off, and furthermore has a closed form solution. The expected improvement acquisition function is defined as (Mockus et al., 1978; Jones et al., 1998):

αtθ=μθtLmaxtΦμθtLmaxt/Σθt+ΣθtϕμθtLmaxt/Σθt,(22)

Where ϕ(.) and Φ(.) refer to the probability density function and cumulative density function of standard normal distribution, Lmaxt=max{L1,Lt} is the maximum log-likelihood value until the latest turn, and μθt and Σθt are the mean and variance of the GP model at iteration t as defined in Eq. 18.

The acquisition function in Eq. 22 holds a closed-form solution and requires the mean and variance of the GP model for any given topology. To solve Eq. 21 for large regulatory networks with a large number of unknown interactions, we can implement some heuristic optimization methods including particle swarm optimization technique (Kennedy and Eberhart, 1995), genetic algorithm (Anderson and Ferris, 1994; Whitley, 1994), or the breadth-first local search (BFLS) (Atabakhsh, 1991) to obtain the model with the largest acquisition value. After the model with maximum acquisition value (θt+1) is chosen, the next log-likelihood evaluation is carried for topology θt+1 to derive the log-likelihood value Lt+1. The GP model is then updated based on all the new information, defined as θ1:t+1 = (θ1:t, θt+1) and L1:t+1=[L1:t,Lt+1]T.

The proposed Bayesian topology optimization continues its sequential process over all the topology space of the regulatory networks for a fixed number of turns, or until no significant change in the maximum log-likelihood value in consecutive iterations is spotted. When the optimization ends, the topology with the largest evaluated likelihood value is selected as the system topology, meaning that:

θ*θi*, where i*=argmaxi=1,,t+1Li.(23)

The inference process consists of three main components. Figure 3 represents the schematic diagram of the proposed method. The GP model predicts the log-likelihood values over the possible topology candidates, denoted by the black dots in Figure 3. The red dots denote the evaluated log-likelihood values for the selected topologies up to the current iteration. Using the posterior distribution of the GP model, the next topology with the highest acquisition function is selected, followed by the log-likelihood evaluation for the selected topology. The GP is then updated based on the selected topology and the evaluated log-likelihood, and this sequential process continues until a stopping criterion is met.

FIGURE 3
www.frontiersin.org

FIGURE 3. Schematic diagram of the proposed topology inference in GRNs.

The detailed steps of the proposed inference method are described in Algorithm 1. Θ denotes the topology space, and D1:T represents the available data. Line 3 to line 8 of the algorithm creates the state index associated with the data D1:T. The sequential topology optimization process is then carried out from line 10 to line 19, where in each loop, the log-likelihood value for a topology selected by the proposed Bayesian optimization technique is computed, followed by the GP posterior update and the next topology selection. Finally, upon the termination of the inference process, the topology with the maximum log-likelihood is chosen in line 20 as the inferred topology. The computation of the log-likelihood determines the complexity of the algorithm at each step of our proposed method, which is of order O (22dT). This means that the complexity at each step of the proposed method is the same as one log-likelihood evaluation. The log-likelihood evaluation at each iteration is used to update our knowledge (posterior), and to help choose the best candidate for future iterations.

www.frontiersin.org

Algorithm 1. The Proposed method for inference of regulatory networks through temporally-sparse data.

4 Numerical experiments

The code repository for replicating the numerical experiments of this paper is included in the data availability statement at the end of this paper. The well-known mammalian cell-cycle network (Fauré et al., 2006) is used to evaluate the performance of our proposed method. Figure 4 presents the pathway diagram of this network. The state vector for this network is assumed as the following x =(CycD, Rb, p27, E2F, CycE, CycA, Cdc20, Cdh1, UbcH10, CycB). The division of mammalian cells depends on the overall organism growth, controlled using signals that activate cyclin D (CycD) in the cell. As can be seen in the state vector, the mammalian cell-cycle network contains 10 genes (d = 10). The settings used for our experiments are as follows: data length of 100 (k = 100), process noise of 0.1 (p = 0.1), and missing data percentage (sparsity) of 50%. Furthermore, 10 regulations of the connectivity matrix are assumed to be unknown (L = 10), and a maximum of 100 likelihood evaluations are used for the inference process. All the parameters used throughout the numerical experiments are expressed in Table 1.

FIGURE 4
www.frontiersin.org

FIGURE 4. Pathway diagram for the cell-cycle network.

TABLE 1
www.frontiersin.org

TABLE 1. Parameter values of mammalian cell-cycle network experiments.

The connectivity matrix and bias vector in Eq. 3 for the mammalian cell-cycle network can be written as:

C=+100000000010+1011000110+1011000101+1001000101+1+1110000010+10+11110000000100+100+1001+100100000+1+11+1+10000001100,b=12121212121212121212T.(24)

In this section, we are assuming that the connectivity matrix is not fully known. This network has 10 genes, and there is a total of 210 = 1, 024 possible states for this network. Consequently, the transition matrix size is 210 × 210, which causes the likelihood evaluation to be computationally expensive for any possible topology. Using our proposed method, we show that the optimal topology with the largest log-likelihood value can be inferred with few likelihood evaluations; hence, we offer an efficient search over all possible topologies.

In all of the experiments, 10 unknown interactions (cij) were considered. Each of the unknown interactions can take their values in the set { + 1, 0, −1}, which leads to 310 = 59, 049 different possible system models, i.e., Θ={θ1,,θ310}. The 10 randomly chosen unknown regulations, which are elements of the connectivity matrix in Eq. 24, are:

c21=1,c35=1,c310=1,c42=1,c54=+1c67=1,c69=1,c83=+1,c96=+1,c98=1.(25)

We also considered a uniform prior distribution for the initial states, i.e., Π0|0θ(i)=1210 for all θΘ and i = 1, 2, …, 210. Furthermore, all the experiments are repeated for 10 independent runs, and the average results along the confidence bounds are reported in all the figures. Note that the randomness of early results come from the process noise (p), and the way the sequential topology optimization is being performed in each run.

For the first set of experiments, the performance of the proposed method is shown using two plots in Figure 5. The left plot represents the progress of the log-likelihood value of the inferred model with respect to the number of likelihood evaluations, meaning that it shows the maximum log-likelihood value obtained during the optimization process. Larger log-likelihood values mean that the chosen model can better represent the true model (i.e., the available data is more likely to come from models with larger likelihood values). As a comparison, we also repeated the same experiment using Genetic Algorithm (GA) (Anderson and Ferris, 1994; Whitley, 1994), which is a powerful and well-known solver for non-continuous problems. By looking at the left plot in Figure 5, we can see that the inference by the proposed method, indicated by the solid blue line, is better than the GA method (dashed red line). This superiority can be seen in terms of the mean and confidence intervals in Figure 5. As we evaluate more likelihoods for different models, the likelihood of the proposed method’s inferred model gets closer to the optimal log-likelihood value, indicated by the dotted red line. Hence, our proposed method is capable of reaching a better log-likelihood with less number of likelihood evaluations and has a more efficient way of searching over all the possible models. Furthermore, the 95% confidence interval is illustrated in the same plot for both methods during this experiment. We can observe that the proposed method’s confidence interval keeps getting smaller, and roughly after 70 evaluations, the confidence interval tends to go zero. This indicates the robustness of the proposed method, where after roughly 70 iterations, the log-likelihood gets to its optimal value at different independent runs. By contrast, the results from the GA still show a large confidence interval even after 100 evaluations, and its average is far less than the optimal log-likelihood value.

FIGURE 5
www.frontiersin.org

FIGURE 5. Results of the mammalian cell-cycle network with 10 unknown interactions.

The right plot of Figure 5 shows the progress of the connectivity error during the optimization process (i.e., number of likelihood evaluations) obtained by the proposed method. Let C* be the vectorized true connectivity matrix indicated in Eq. 24, and Ct be the vectorized inferred connectivity matrix at tth likelihood evaluation. The connectivity error at iteration t is defined as ‖C* − Ct1. Evidently, we will have a better estimate of the true model as this error gets closer to zero. In the right plot, we can see that the connectivity error decreases as we do more evaluations, and after about 75 likelihood evaluations, the error gets to zero, meaning that we successfully inferred the true connectivity matrix. Also, as expected, we can see that the 95% confidence interval gets smaller as we do more evaluations and eventually gets close to zero after about 75 evaluations.

In the second set of experiments, we aim to investigate the effect of missing data percentage on the performance of the proposed method. It is expected that with more missing data, it would be more difficult to infer the relationship between different components of the system; hence the connectivity error for the inferred model would be larger. For these experiments, we changed the missing data percentage from 0% to 90% and used Bernoulli noise value 0.2. Other parameters are fixed based on Table 1. The mean of the inferred models’ connectivity error obtained from these experiments, along with their 68% confidence interval are presented as bar plots in Figure 6. As expected, these results demonstrate that the mean of connectivity error increases as the missing data percentage gets larger.

FIGURE 6
www.frontiersin.org

FIGURE 6. Performance of the proposed method with respect to percentage of missing data.

The final set of experiments focuses on how the Bernoulli noise affects the performance of the proposed method. In all these experiments, we consider 50% missing data percentage, and we change the Bernoulli noise from 0.01 to 0.4. For performance comparison, the mean of the inferred models’ connectivity error derived from these experiments is shown using bar plots in Figure 7. In this bar plot, we can observe that the connectivity error is large for the Bernoulli noise of 0.01. As the noise increases to 0.05 and 0.1, the connectivity error keeps decreasing. However, increasing the noise to 0.2, 0.3, and finally 0.4 results in a continuous increase in the connectivity error. These results demonstrate the relationship between network stochasticity and data informativity needed for the inference process. For a small process noise (p = 0.01), the network is typically trapped in attractor states, which precludes the observation of the entire state space. This leads to the issue of statistical non-identifiability, which refers to the situation where multiple models are not clearly distinguishable using the available data. Once the noise value is slightly increased (p = 0.05, p = 0.1), the network gets out of its attractor states more often, which enhances the performance of the inference process. Finally, for too large process noise values (p = 0.2, p = 0.3, p = 0.4), the state transitions become more chaotic, making it more difficult to infer the true relations between the components.

FIGURE 7
www.frontiersin.org

FIGURE 7. Performance of the proposed method in presence of different Bernoulli noise.

5 Conclusion

This paper presents a highly scalable topology inference method for gene regulatory networks (GRNs) observed through temporally sparse data. The Boolean network model is used for capturing the dynamics of the GRNs. The inference process consists of inferring the interactions between genes or equivalently selecting a topology for the system among all the possible topologies that have the highest likelihood value. Evaluating the likelihood function for any given topology is expensive, preventing exhaustive search over the large possible topology space. The proposed method models the log-likelihood function by a Gaussian Process (GP) model with a structurally-inspired kernel function. This GP model captures the correlation between different possible topologies and provides the Bayesian representation of the log-likelihood function. Using the posterior distribution of the GP model, Bayesian optimization is used to efficiently search over the topology space.

The high performance of our proposed method is shown using multiple experiments on the well-known mammalian cell-cycle network. We have also repeated all the experiments multiple times to obtain a confidence interval and further demonstrate the accuracy and robustness of the solutions obtained by our method. In the first experiment, we considered the topology inference of the mammalian cell-cycle network with 10 unknown interactions and 50% missing data. From comparing the results of topology inference using our proposed method and genetic algorithm, we observed that our method is more efficient in searching over the topology space and reaches an optimal model with fewer likelihood evaluations. Meanwhile, the small confidence interval of our method justified the robustness of the solutions. The second experiment investigated the effect of missing data on the performance of the proposed inference method. From the results, we understand that as expected, with more missing data, the method’s accuracy reduces, and the inference error becomes larger. Finally, in the third experiment, we studied the performance of our method in the presence of different Bernoulli noise (i.e., stochasticity in the state process). The results show that for small stochasticity, the accuracy of the inference is low, as the system spends most of its time in a few states (i.e., attractors) and the interactions between different components of the system are not distinguishable. As the stochasticity increases, the accuracy of the proposed method increases (as the error decreases) until a certain point, and after that again, the accuracy starts decreasing. This is because too much stochasticity turns the system into a more chaotic form, making the inference of the true model more challenging.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://github.com/ImaniLab/Frontiers-2022.

Author contributions

MA and MI contributed to conceptualization, methodology, validation, formal analysis, and investigation of the study. MA was responsible for coding, resources, and data curation of the work. Supervision and project administration was done by MI. Both authors contributed to writing and editing of the manuscript, and they have read and agreed to the submitted version of the manuscript.

Funding

This work has been supported in part by the National Institutes of Health award 1R21EB032480-01, National Science Foundation award IIS-2202395, ARMY Research Office award W911NF2110299, and Oracle Cloud credits and related resources provided by the Oracle for Research program.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abou-Jaoudé, W., Traynard, P., Monteiro, P., Saez-Rodriguez, J., Helikar, T., Thieffry, D., et al. (2016). Logical modeling and dynamical analysis of cellular networks. Front. Genet. 7, 94. doi:10.3389/fgene.2016.00094

PubMed Abstract | CrossRef Full Text | Google Scholar

Anderson, E. J., and Ferris, M. C. (1994). Genetic algorithms for combinatorial optimization: The assemble line balancing problem. ORSA J. Comput. 6, 161–173. doi:10.1287/ijoc.6.2.161

CrossRef Full Text | Google Scholar

Atabakhsh, H. (1991). A survey of constraint based scheduling systems using an artificial intelligence approach. Artif. Intell. Eng. 6, 58–73. doi:10.1016/0954-1810(91)90001-5

CrossRef Full Text | Google Scholar

Auer, P. (2003). Using confidence bounds for exploitation-exploration trade-offs. J. Mach. Learn. Res. 3, 397–422.

Google Scholar

Barman, S., and Kwon, Y.-K. (2017). A novel mutual information-based boolean network inference method from time-series gene expression data. Plos one 12, e0171097. doi:10.1371/journal.pone.0171097

PubMed Abstract | CrossRef Full Text | Google Scholar

Barman, S., and Kwon, Y.-K. (2018). A boolean network inference from time-series gene expression data using a genetic algorithm. Bioinformatics 34, i927–i933. doi:10.1093/bioinformatics/bty584

PubMed Abstract | CrossRef Full Text | Google Scholar

Brochu, E., Cora, V. M., and de Freitas, N. (2010). A tutorial on bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. arxiv. doi:10.48550/ARXIV.1012.2599

CrossRef Full Text | Google Scholar

Cheng, D., and Zhao, Y. (2011). Identification of boolean control networks. Automatica 47, 702–710. doi:10.1016/j.automatica.2011.01.083

CrossRef Full Text | Google Scholar

Dougherty, E., and Qian, X. (2013). Validation of gene regulatory network inference based on controllability. Front. Genet. 4, 272. doi:10.3389/fgene.2013.00272

PubMed Abstract | CrossRef Full Text | Google Scholar

Fauré, A., Naldi, A., Chaouiya, C., and Thieffry, D. (2006). Dynamical analysis of a generic Boolean model for the control of the mammalian cell cycle. Bioinformatics 22, e124–e131. doi:10.1093/bioinformatics/btl210

PubMed Abstract | CrossRef Full Text | Google Scholar

Frazier, P. (2009). Knowledge-gradient methods for statistical learning. Princeton, NJ: Department of Operations Research and Financial Engineering, Princeton University. Dissertation.

Google Scholar

Frazier, P. I. (2018). A tutorial on bayesian optimization. arxiv. doi:10.48550/ARXIV.1807.02811

CrossRef Full Text | Google Scholar

Godsill, S. J., Doucet, A., and West, M. (2004). Monte Carlo smoothing for nonlinear time series. J. Am. Stat. Assoc. 99, 156–168. doi:10.1198/016214504000000151

CrossRef Full Text | Google Scholar

Hajiramezanali, E., Imani, M., Braga-Neto, U., Qian, X., and Dougherty, E. R. (2019). Scalable optimal Bayesian classification of single-cell trajectories under regulatory model uncertainty. BMC genomics 20, 435. doi:10.1186/s12864-019-5720-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Hashimoto, R. F., Kim, S., Shmulevich, I., Zhang, W., Bittner, M. L., and Dougherty, E. R. (2004). Growing genetic regulatory networks from seed genes. Bioinformatics 20, 1241–1247. doi:10.1093/bioinformatics/bth074

PubMed Abstract | CrossRef Full Text | Google Scholar

Henrández-Lobato, J. M., Hoffman, M. W., and Ghahramani, Z. (2014). “Predictive entropy search for efficient global optimization of black-box functions,” in Proceedings of the 27th International Conference on Neural Information Processing Systems, 918–926.

Google Scholar

Hürzeler, M., and Künsch, H. R. (1998). Monte Carlo approximations for general state-space models. J. Comput. Graph. Statistics 7, 175–193. doi:10.1080/10618600.1998.10474769

CrossRef Full Text | Google Scholar

Imani, M., and Braga-Neto, U. M. (2017). Maximum-likelihood adaptive filter for partially observed Boolean dynamical systems. IEEE Trans. Signal Process. 65, 359–371. doi:10.1109/tsp.2016.2614798

CrossRef Full Text | Google Scholar

Imani, M., and Ghoreishi, S. F. (2021). Two-stage Bayesian optimization for scalable inference in state space models. IEEE Trans. Neural Netw. Learn. Syst. 33, 5138–5149. doi:10.1109/tnnls.2021.3069172

CrossRef Full Text | Google Scholar

Imani, M., Dehghannasiri, R., Braga-Neto, U. M., and Dougherty, E. R. (2018). Sequential experimental design for optimal structural intervention in gene regulatory networks based on the mean objective cost of uncertainty. Cancer Inf. 17, 117693511879024. doi:10.1177/1176935118790247

PubMed Abstract | CrossRef Full Text | Google Scholar

Imani, M., Ghoreishi, S. F., Allaire, D., and Braga-Neto, U. (2019). MFBO-SSM: Multi-fidelity Bayesian optimization for fast inference in state-space models. Proc. AAAI Conf. Artif. Intell. 33, 7858–7865. doi:10.1609/aaai.v33i01.33017858

CrossRef Full Text | Google Scholar

Imani, M., Dougherty, E., and Braga-Neto, U. (2020). Boolean Kalman filter and smoother under model uncertainty. Automatica 111, 108609. doi:10.1016/j.automatica.2019.108609

CrossRef Full Text | Google Scholar

Johansen, A. M., Doucet, A., and Davy, M. (2008). Particle methods for maximum likelihood estimation in latent variable models. Stat. Comput. 18, 47–57. doi:10.1007/s11222-007-9037-8

CrossRef Full Text | Google Scholar

Jones, D. R., Schonlau, M., and Welch, W. J. (1998). Efficient global optimization of expensive black-box functions. J. Glob. Optim. 13, 455–492. doi:10.1023/a:1008306431147

CrossRef Full Text | Google Scholar

Kantas, N., Doucet, A., Singh, S. S., Maciejowski, J., and Chopin, N. (2015). On particle methods for parameter estimation in state-space models. Stat. Sci. 30, 328–351. doi:10.1214/14-sts511

CrossRef Full Text | Google Scholar

Karbalayghareh, A., Braga-neto, U., and Dougherty, E. R. (2018). Intrinsically bayesian robust classifier for single-cell gene expression trajectories in gene regulatory networks. BMC Syst. Biol. 12, 23–10. doi:10.1186/s12918-018-0549-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Kennedy, J., and Eberhart, R. (1995). “Particle swarm optimization,” in Proceedings of ICNN’95-International Conference on Neural Networks (IEEE), 1942–1948.

Google Scholar

Lähdesmäki, H., Shmulevich, I., and Yli-Harja, O. (2003). On learning gene regulatory networks under the boolean network model. Mach. Learn. 52, 147–167. doi:10.1023/A:1023905711304

CrossRef Full Text | Google Scholar

Marshall, S., Yu, L., Xiao, Y., and Dougherty, E. R. (2007). Inference of a probabilistic boolean network from a single observed temporal sequence. Eurasip J. Bioinforma. Syst. Biol. 2007, 1–15. doi:10.1155/2007/32454

PubMed Abstract | CrossRef Full Text | Google Scholar

Mockus, J., Tiesis, V., and Zilinskas, A. (1978). The application of Bayesian methods for seeking the extremum. Towards Glob. Optim. 2, 2.

Google Scholar

Pusnik, Z., Mraz, M., Zimic, N., and Moskon, M. (2022). Review and assessment of boolean approaches for inference of gene regulatory networks. Heliyon 8, e10222. doi:10.1016/j.heliyon.2022.e10222

PubMed Abstract | CrossRef Full Text | Google Scholar

Rasmussen, C. E., and Williams, C. (2006). Gaussian processes for machine learning. Cambridge, United States: MIT Press.

Google Scholar

Saadatpour, A., and Albert, R. (2013). Boolean modeling of biological regulatory networks: A methodology tutorial. Methods 62, 3–12. doi:10.1016/j.ymeth.2012.10.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Schön, T. B., Wills, A., and Ninness, B. (2011). System identification of nonlinear state-space models. Automatica 47, 39–49. doi:10.1016/j.automatica.2010.10.013

CrossRef Full Text | Google Scholar

Shahriari, B., Swersky, K., Wang, Z., Adams, R. P., and de Freitas, N. (2016). Taking the human out of the loop: A review of bayesian optimization. Proc. IEEE 104, 148–175. doi:10.1109/jproc.2015.2494218

CrossRef Full Text | Google Scholar

Shmulevich, I., and Dougherty, E. R. (2010). Probabilistic boolean networks: The modeling and control of gene regulatory networks. Philadelphia, United States: siam.

Google Scholar

Shmulevich, I., Dougherty, E. R., and Zhang, W. (2002). From Boolean to probabilistic Boolean networks as models of genetic regulatory networks. Proc. IEEE 90, 1778–1792. doi:10.1109/jproc.2002.804686

CrossRef Full Text | Google Scholar

Tan, Y., Neto, F. L., and Braga-Neto, U. (2020). Pallas: Penalized maximum likelihood and particle swarms for inference of gene regulatory networks from time series data. IEEE/ACM Trans. Comput. Biol. Bioinform. 19, 1807–1816. doi:10.1109/tcbb.2020.3037090

CrossRef Full Text | Google Scholar

Whitley, D. (1994). A genetic algorithm tutorial. Stat. Comput. 4, 65–85. doi:10.1007/bf00175354

CrossRef Full Text | Google Scholar

Wills, A., Schön, T. B., Ljung, L., and Ninness, B. (2013). Identification of hammerstein–wiener models. Automatica 49, 70–81. doi:10.1016/j.automatica.2012.09.018

CrossRef Full Text | Google Scholar

Wu, J., Poloczek, M., Wilson, A. G., and Frazier, P. (2017). “Bayesian optimization with gradients” in Advances in Neural Information Processing Systems (Curran Associates, Inc.) 30.

Google Scholar

Wynn, M. L., Consul, N., Merajver, S. D., and Schnell, S. (2012). Logic-based models in systems biology: A predictive and parameter-free network analysis method. Integr. Biol. 4, 1323. doi:10.1039/c2ib20193c

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: topology inference, maximum likelihood estimation, gene regulatory networks, Boolean dynamical systems, Bayesian optimization

Citation: Alali M and Imani M (2022) Inference of regulatory networks through temporally sparse data. Front. Control. Eng. 3:1017256. doi: 10.3389/fcteg.2022.1017256

Received: 11 August 2022; Accepted: 15 November 2022;
Published: 13 December 2022.

Edited by:

Mahyar Fazlyab, Johns Hopkins University, United States

Reviewed by:

Salar Fattahi, University of Michigan, United States
Valentina Breschi, Politecnico di Milano, Italy

Copyright © 2022 Alali and Imani. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohammad Alali, YWxhbGkubUBub3J0aGVhc3Rlcm4uZWR1

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.