Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 04 July 2024
Sec. Machine Learning and Artificial Intelligence
This article is part of the Research Topic Knowledge-guided Learning and Decision-Making View all 6 articles

Bayesian reinforcement learning for navigation planning in unknown environments

  • Department of Electrical and Computer Engineering, Northeastern University, Boston, MA, United States

This study focuses on a rescue mission problem, particularly enabling agents/robots to navigate efficiently in unknown environments. Technological advances, including manufacturing, sensing, and communication systems, have raised interest in using robots or drones for rescue operations. Effective rescue operations require quick identification of changes in the environment and/or locating the victims/injuries as soon as possible. Several techniques have been developed in recent years for autonomy in rescue missions, including motion planning, adaptive control, and more recently, reinforcement learning techniques. These techniques rely on full knowledge of the environment or the availability of simulators that can represent real environments during rescue operations. However, in practice, agents might have little or no information about the environment or the number or locations of injuries, preventing/limiting the application of most existing techniques. This study provides a probabilistic/Bayesian representation of the unknown environment, which jointly models the stochasticity in the agent's navigation and the environment uncertainty into a vector called the belief state. This belief state allows offline learning of the optimal Bayesian policy in an unknown environment without the need for any real data/interactions, which guarantees taking actions that are optimal given all available information. To address the large size of belief space, deep reinforcement learning is developed for computing an approximate Bayesian planning policy. The numerical experiments using different maze problems demonstrate the high performance of the proposed policy.

1 Introduction

Advances in robotics, sensing, manufacturing, and communication in recent years have raised interest in the use of autonomous agents instead of humans for rescue missions. Examples include using robots for time-sensitive and dangerous rescue missions, such as response to earthquakes, mass shootings, hurricanes, and warfare zones. The utmost factor in rescue operations is quick identification of changes in an environment or locating victims/injuries in need of critical care.

A maze environment containing a single robot and a victim is shown in Figure 1. Let the yellow cells represent the unknown parts of the environment after the disaster, which could possibly be blocked with debris. Three possible navigation paths are shown in the maze. Given the unknown parts of the environment, should the agent take the right path, which is the closest to the victim? Wouldn't it be better to take the longer black path without any potential blockages (no yellow cells)? How can the robot change its decision as more information about the victim's condition and the environment appears?

Figure 1
www.frontiersin.org

Figure 1. Illustrative example of the rescue operation problem in unknown environments. The rescue robot can take any of the black, green, and orange paths to get to the victim. Depending on the type of the unknown cells and whether each of the unknown cells is more probable to be wall, empty, or another victim/injury, the rescue robot should plan to take the best path.

Several techniques for achieving autonomy in rescue operations have been developed in recent years, with reinforcement learning (RL) being one of the prominent methods. In recent years, RL techniques have achieved remarkable success across various domains, including network security, biological applications, and robotics (Alali and Imani, 2023, 2024; Elguea-Aguinaco et al., 2023; Ravari et al., 2023; Alali et al., 2024; Asadi et al., 2024). For autonomy in rescue operations, various RL techniques have been developed for single-agent and multi-agent settings (Imanberdiyev et al., 2016; Zhang et al., 2018; Bøhn et al., 2019; Lin et al., 2019; Niroui et al., 2019; Sampedro et al., 2019; Ebrahimi et al., 2020; Hu et al., 2020; Wu et al., 2021). These RL techniques can be divided into model-based and simulation-based categories. The model-based RL approaches (Imanberdiyev et al., 2016; Pham et al., 2018; Ladosz et al., 2019; Sampedro et al., 2019; Xu et al., 2019) assume full knowledge (including potential changes/casualties) about the agent's environment during the rescue; the simulation-based RL techniques (Akcakoca et al., 2019; Wang et al., 2019; Blum et al., 2020; Hamid et al., 2021; Jagannath et al., 2021; Falcone and Putnam, 2022), on the other hand, rely on the availability of simulators to represent the environment during disaster response. However, the environment during the rescue operations is often unknown or partially known to the human and agent, and it is prudent for an agent to make decisions given the incomplete available information. For instance, in response to an earthquake, the number and location of victims/injuries and the extent of damage to the environment are often unknown at the early stages of rescue operations. This prevents the applicability of existing RL techniques in time-sensitive and unknown environments. It should also be noted that the RL techniques cannot be employed for learning a policy through real interactions with the unknown environment, as RL often requires thousands of interactions to learn to act in an unknown environment, which is impossible due to the time-sensitive nature of rescue operations.

Several techniques have been developed to combine Bayesian approaches with RL methods (Ghavamzadeh et al., 2015; Imani and Ghoreishi, 2022). Most of these approaches aim to address the sample efficiency of RL methods during the learning process (Ghavamzadeh et al., 2015; Imani et al., 2018; Kamthe and Deisenroth, 2018). Meanwhile, Bayesian approaches have been used to quantify the discrepancies between real-world environments and simulations, facilitating sim-to-real policy transfer (Feng et al., 2022; Rothfuss et al., 2024). Other Bayesian approaches have also been developed in multi-agent and human-AI teaming to learn the intentions and preferences of teammates using partial data (Lin et al., 2024; Ravari et al., 2024a,b; Zhang et al., 2024a,b). The most relevant class of approaches considering uncertainty in environments is Bayes-adaptive methods (Guez et al., 2012; Rigter et al., 2021; Zintgraf et al., 2021). These methods iteratively update the posterior distribution of the environment and simultaneously update the planning policy based on the latest interactions. However, these methods are applicable to domains with finite state spaces and fully unknown environments, where a huge number of interactions are needed to learn the distribution of the transition probabilities. These methods are also not applicable to partially known environments or domains with large and continuous state spaces and often perform poorly under limited available interactions.

Motion planning (Zhang et al., 2013; Perez-Imaz et al., 2016; Cabreira et al., 2019; de Almeida et al., 2019; Boulares and Barnawi, 2021) is another class of model-based approaches which aims to take advantage of the system model for offline planning. Examples of these methods are LQR/LQG methods and their non-linear variations (Richter and Roy, 2017; Kim et al., 2019, 2021; Bouman et al., 2020; Rosolia et al., 2022), which rely on the full or rich knowledge of the system model for planning or replanning, which prevents their applications in unknown environments. In this regard, active learning, model-predictive control, and online learning techniques have been developed for decision-making in unknown environments (Juang and Chang, 2011; Luo et al., 2014; Greatwood and Richards, 2019; Li et al., 2019; Chang et al., 2021). These methods mostly rely on a greedy and local view of the environment and perform poorly in complex realistic domains. Safety in navigation in unknown environments has also been studied extensively in the literature (Bajcsy et al., 2019; Krell et al., 2019; Tordesillas et al., 2019), which focuses on guaranteeing safety in domains with sensitive constraints.

This study develops a reinforcement learning Bayesian planning policy for rescue operations in unknown environments. We define a belief state, which keeps a joint probabilistic representation of all the possible models for the environment and agent movements. The belief state allows a Markov decision process (MDP) formulation of an agent in unknown and uncertain environments, allowing propagation of the entire uncertainty offline without the need for interaction with the real environment. We formulate the exact optimal Bayesian planning policy, which guarantees that an agent acts optimally given all available information. A Bayesian solution is introduced using a deep reinforcement learning technique, allowing offline learning of the policy over the whole belief space. We demonstrate that the proposed reinforcement learning Bayesian policy can be employed in real time for rescue missions in stationary and non-stationary environments as any additional information unfolds (i.e., without the need for learning or retraining). The effectiveness of the proposed method is demonstrated using comprehensive numerical experiments using different maze problems.

The article is organized as follows. In Section 2, the background of the Markov decision process is briefly described. In Section 3, the optimal Bayesian policy is formulated, and a solution based on deep reinforcement learning is introduced. Section 4 includes a discussion about the capabilities and complexity of the proposed method. Finally, Section 5 and Section 6 contain numerical examples and concluding remarks, respectively.

2 Background—A Markov decision process

A Markov decision process (MDP) can be defined by a 4-tuple S,A,Tθ,Rθ, where S is the state space, A is the action space, Tθ:S×A×S is the unknown or partially known state transition probability function such that Tθ(s,a,s)=P(ss,a,θ) with the set of unknown parameters θ ∈ Θ, and Rθ:S×A×S is a bounded reward function with a real value outcome such that Rθ(s,a,s) encodes the reward earned when action a is taken in state s and the agent moves to state s′ in model θ. The reward function could be model-dependent in general form, meaning that similar transitions in different models of the environment might lead to different rewards.

If a given MDP parameterized by θ is a true environment, we could define a deterministic policy π:SA as a mapping from states to actions. The expected discounted reward function at state sS after taking action aA and following policy π afterward is defined as follows:

qθπ(s,a)=𝔼[t=0hγtRθ(st,at,st+1)s0=s,a0=a,a1:~π,θ].    (1)

where γ is a discount factor. In the finite-horizon case, the discount factor is typically set to 1. In the infinite-horizon case (h = ∞), the discount factor γ ∈ [0, 1) is included to obtain a finite sum.

According to (1), the expected return under the optimal policy πθ* for the environment modeled by θ can be expressed as follows:

qθ*(s,a)=𝔼[t=0hγtRθ(st,at,st+1)s0=s,a0=a,a1:~πθ*,θ],    (2)

where qθ*(s,a) indicates the expected discounted reward after executing action a in state s and following optimal policy πθ* afterward. An optimal model-specific policy πθ* attains the maximum expected return as πθ*(s)=argmaxaAqθ*(s,a), for sS.

If the true environment model, that is, θ*, was fully known to the agent, the optimal policy πθ** could be obtained using (2). However, the environment is often unknown in rescue operations, and the agent needs to make decisions given partial knowledge about the environment and/or the number or location of injuries.

3 Proposed Bayesian planning policy for decision-making in unknown environments

3.1 Probabilistic/Bayesian formulation of unknown and uncertain environments

This section first describes the challenges of decision-making in an unknown environment, followed by the proposed framework to overcome these challenges effectively. Let us consider a maze problem as a simple example of navigation in rescue operations, where each cell in the maze could be one of the followings: wall “W”, empty “E”, or injury/victim “I”. Figure 2 represents an example of a maze problem, where the black and white colors indicate the wall and empty cells, respectively. The yellow cells are unknown parts of the environment, which each could potentially be a wall (i.e., blocked after a disaster), be empty, or contain a victim/injury. Let {c1, ..., cm} be m unknown cells in the environment, where ci ∈ {W, E, I}. These unknown cells lead to 3m different possible environment models (i.e., maze models) denoted by Θ = {θ1, ..., θ3m}, where θj = [θj(1), ⋯  , θj(m)] represents an MDP parametrized by θj, and θj(l) denotes the type of the lth unknown cell under the jth maze model. Note that the three unknown cells in Figure 2 lead to 33 = 27 possible maze models, where the environments could potentially contain 0, 1, 2, or 3 injuries. In practice, the true environment θ* is often hidden among all the possible models in Θ. For each θ ∈ Θ, the number and location of injuries and walls vary. Therefore, the optimal model-specific policies for these models [see Equation (2)] are different in general; thus, the policies obtained for different models θ ∈ Θ cannot be directly employed in unknown environments. It should be noted that the maze problem and its uncertain elements are examples of rescue missions, and the proposed method in this study could be applied to more general environments and environment uncertainties.

Figure 2
www.frontiersin.org

Figure 2. Visualization of the 6 × 4 maze problem. This maze has three unknown cells, where each could be either wall, empty, or victim/injury.

Let the initial knowledge about possible models for the true environment be represented as ϑ0=[P(θ*=θ1),...,P(θ*=θ3m)], where P(θ* = θj) shows the prior probability that the jth model is the true environment model and j=13mϑ0(j)=1. Let a0 : k−1 = {a0, ..., ak−1} be the agent's actions and s1 : k = {s1, ..., sk} be the agent states until time step k. The posterior distribution of models can be expressed as follows:

ϑk=[P(θ*=θ1s1:k,a0:k-1), ,P(θ*=θ3ms1:k,a0:k-1)],    (3)

where ϑk(j) indicates the posterior probability that model θj is the true model, and we also have j=13mϑk(j)=1. Note that if no prior information about the models is available, a uniform (i.e., non-informative) distribution can be employed. If the independency of distribution of m unknown cells is assumed, the posterior probability could be defined over the m unknown cells as (Equation 4)

pk=[P(c1=W|s1:k,a0:k1),P(c1=E|s1:k,a0:k1),             P(c1=I|s1:k,a0:k1),,P(cm=W|s1:k,a0:k1),             P(cm=E|s1:k,a0:k1),P(cm=I|s1:k,a0:k1)],    (4)

which leads to the posterior distribution over all the possible models (i.e., ϑk) as

ϑk(i)=l=1m[1θi(l)=Wpk(3l-2)+1θi(l)=Epk(3l-1)               +1θi(l)=Ipk(3l)],    (5)

for i = 1, ..., 3m and 1condition returns 1 if the condition is true, and 0 otherwise. Note that the rest of the study is derived for the general form of posterior in (3) without the independence assumption.

We define the belief state at time step k as the vector of joint agent's state (i.e., sk) and posterior probability of unknown models ϑk:

bk=[sk,ϑk]T,    (6)

where bk is a vector of size |sk|+3m, and b0=[s0,ϑ0]T is the initial belief state. sk is the agent state at time step k taking a value in S, and ϑk is a vector of size 3m, where each element takes continuous values between 0 and 1, and the sum of elements is 1. Thus, the space of belief state can be expressed as B={S×Δ3m}, where Δ3m represents a simplex of size 3m.

3.2 MDP representation in belief space

In Section 3.1, we described that navigation in an unknown environment could be represented by an unknown MDP. Here, we show that the belief state defined in (6) allows representation of the navigation task in an unknown environment through a known MDP. The rationale behind this mapping is that, unlike in unknown MDPs, reinforcement learning techniques can be employed to find the optimal policy for a known MDP. In the following paragraphs, we first define the MDP in the belief space, then we represent all its elements, and finally, we formulate the reinforcement learning policy in this known MDP.

The MDP in the belief space can be expressed through B,A,T~,R~, where B is the belief state space, T~:B×A×B is a known transition probability in the belief space such that T~(b,a,b)=P(bb,a) represents the transition from the belief state b to b′ if action a is taken. Note that [as proven in (7)] this transition is Markov and does not need the true model of the environment since the belief state contains the entire system uncertainty. Finally, the expected reward function in the belief state is represented by R~:B×A, where R~(b,a) represents the expected immediate reward if action a is taken at belief state b.

Here, we provide proof that the belief transition is a Markov process. Let b0, a0, ..., ak−1, bk be the sequence of actions and belief states up to time step k. If action ak is taken at time step k, the probability of the next belief state can be expressed as follows:

P(bk+1|ak,bk,...,b0,a0)=P(sk+1,ϑk+1|ak,sk,ϑk,...,s0,ϑ0,a0)             =P(sk+1|ak,sk,ϑk,...,s0,ϑ0,a0)             ×P(ϑk+1|sk+1,ak,sk,ϑk,...,s0,ϑ0,a0)             =P(sk+1|ak,sk,ϑk)P(ϑk+1|sk+1,ak,sk,ϑk)             =P(sk+1,ϑk+1|ak,sk,ϑk)=P(bk+1|ak,bk),    (7)

where the third line is written given the fact that ϑk includes the posterior distribution of models given all sequences of states and actions up to time step k. Therefore, the terms dropped in lines 2 and 3 of Equation (7) are already included in the posterior distribution ϑk.

Given that b = [s, ϑ]T is the current belief state, and a is the selected action at the current time, the next belief state can be one of the following |S| vectors:

bb,a~{b1=[s1,ϑ1]Tw.p.P(b1b,a)b2=[s2,ϑ2]Tw.p.P(b2b,a)b|S|=[s|S|,ϑ|S|]Tw.p.P(b|S|b,a).    (8)

The next belief state contains the next state and new posterior distribution of models. For instance, bi=[si,ϑi]T is one of S possible next belief states if the agent moves to state si after taking action a in state s. The posterior ϑi upon observing si can be computed as follows:

ϑi(j)=P(θ*=θjs=si,a,s,ϑ)=P(sis,a,θj)ϑ(j)l=13mP(sis,a,θl)ϑ(l),    (9)

for j=1,...,3m,i=1,...,|S|.

The probability for all |S| possible next belief state transitions denoted in (9) can be computed according to the Markovian properties of the belief transition in (7). In particular, the probability that the ith belief state bi=[si,ϑi]T is observed upon taking action a in belief state b = [s, ϑ]T can be expressed as follows:

P(bk+1=bibk=b,ak=a)    =P(sk+1=si,ϑk+1=ϑibk=b,ak=a)    =P(sk+1=sisk=s,ϑk=ϑ,ak=a)    ×P(ϑk+1=ϑisk+1=si,sk=s,ϑk=ϑ,ak=a)    =P(sk+1=sisk=s,ϑk=ϑ,ak=a).    (10)

The last line of (10) is obtained given that ϑk+1 can only take a single value ϑi [computed in (9)] with probability 1.

The expected reward function R~(b,a) in the belief space can be expressed in terms of the reward function of the true environment as follows:

R~(b=[s,ϑ]T,a)=b{b1,...,b|S|}P(b=[s,ϑ]Tb,a)Eθϑ[Rθ(s,a,s)]=i=1|S|P(bi=[si,ϑi]Tb,a)l=13mϑi(l)Rθl(s,a,si),    (11)

where Rθl(s,a,s) represents the improvement in the rescue operation after taking action a and moving from state s to s′ in model θl. It can be seen that the reward function in (11) depends on the posterior distribution of models and the uncertainty in the agent state transitions.

Aside from the ability to incorporate any arbitrary reward functions defined for the true environment, the proposed belief formulation allows actively exploring/learning the unknown parts of the environment. For instance, robots/drones might be deployed in rescue operations to quickly identify road closures or other environmental casualties. Depending on the application, the uncertainty at specific/targeted locations or the entire environment might be needed. The immediate gain in uncertainty reduction at belief state b = [s, ϑ]T after taking action a can be expressed as follows:

R~(b=[s,ϑ]T,a)=-𝔼b=[s,ϑ]Tb,a[H(ϑ)-H(ϑ)]    =-i=1|S|P(bib,a)[H(ϑi)-H(ϑ)]    =i=1|S|P(bib,a)l=13m[ϑi(l)logϑi(l)-ϑ(l)logϑ(l)],    (12)

where H(ϑ) denotes the remaining entropy (i.e., uncertainty) in the environment model represented by the posterior probability ϑ. Larger positive reward values correspond to more reduction of entropy/uncertainty upon moving to belief state b′. The entropy takes its lowest value 0 when representing the case where a single model has posterior probability 1 and others 0. Therefore, this reward function helps agents toward taking actions that provide the highest information about unknown parts of the environment, which is crucial in rescue operations. Note that depending on our application, a more general form of the reward function can be employed for learning the navigation policy.

3.3 Deep reinforcement learning Bayesian planning policy

The MDP defined in the belief space is fully known as it considers the posterior of all the possible environment models. We define μ:BA as a deterministic policy, which associates an action to each sample in the belief space. The optimal policy in the belief space can be formulated as follows:

μ*(b)=argmaxμ𝔼[t=0γtR~(bt,at)b0=b,a0:~μ],    (13)

for any bB; where the maximization is over all possible policies in the belief space. The expectation in (13) is with respect to the stochasticity in the belief space denoted in (8), which includes the uncertainty in the state transition and the posterior of environment models reflected in the belief states. The optimal Bayesian policy, μ*, yields optimality given all available information reflected in the belief state (i.e., the agent and model uncertainty). Finding the exact solution for the optimization problem in (13) is not possible due to the large size of the belief space. In the following paragraphs, we provide an approximate solution for finding the optimal Bayesian policy in (13) using a deep reinforcement learning approach.

This study employs the belief transition in (8) as a simulator to generate offline trajectories required for training a deep RL agent. These trajectories are belief transitions that propagate the agent states and the environment uncertainty, thus, do not require interaction with the real environment. Given the discrete nature of action space, we employ the deep Q-network (DQN) method (Mnih et al., 2015) for learning the Bayesian policy in (13). This approach aims to approximate the following expected discounted reward function defined over the belief space:

Q*(b,a)=𝔼[t=0γtR~(bt,at)b0=b,a0=a,a1:~μ*],    (14)

for any bB and aA.

DQN approximates the Q-function in (14) using two feed-forward deep neural networks, called Q-network and target-network, represented by Qw and Qw-, respectively. These two neural networks share the same structure; the Q-network's input is the belief state, and its outputs are Qw(b,a1),...,Qw(b,aA), each associated with an action. The initial weights for both Q-network and target-network are set randomly.

For training of the neural networks, a replay memory D of fixed size is considered. This memory is filled and replaced by repeated episodes of belief states governed by actions generated from the epsilon-greedy policy. Each episode starts from an initial belief state b0=[s0,ϑ0]T, which, if unknown, can be selected randomly from the belief space, that is, b0B. At step t of the episode, an action can be selected according to the epsilon-greedy policy defined using the Q-network Qw as follows:

at~{argmaxaAQw(bt,a)w.p. 1-ϵrandom{a1,...,aA}w.p.ϵ,    (15)

where 0 ≤ ϵ ≤ 1 is the epsilon-greedy policy rate, which controls the level of exploration during the learning process.

Upon generating a fixed number of steps, the Q-network Qw should be updated according to a minibatch set of experiences selected from the replay memory D. Letting

Z={(b~n,a~n,b~n+1,r~n+1)}n=1Nbatch~D,    (16)

be selected as a minibatch set, the target values for updating the Q-network can be computed as follows:

yn=r~n+1+γmaxaAQw-(b~n+1,a),    (17)

for n = 1, ..., Nbatch, where the target-network Qw- is used for computation of the target values. Using these target values, the Q-network weights, w, can be updated as follows:

w=w-αw[n=1Nbatch(yn-Qw(b~n,a~n))2],    (18)

where α is the learning rate, and the mean squared error is used for the loss function in the weights update. The optimization in (18) can be carried out using a stochastic gradient optimization approach such as Adam (Kingma and Ba, 2015). Upon updating w, the weights of the target-network, w, should also be updated using the soft update:

w-=(1-τ)w-+τw,    (19)

where τ is a hyperparameter.

It should be noted that the trajectories used in the DQN method are acquired offline through the belief transition in (8). The training can be stopped when performance improvement becomes negligible, or a pre-specified performance is achieved. Upon termination of the offline training, the Q-network approximates the optimal Bayesian policy as μ*(b)argmaxaAQw(b,a). This Bayesian policy prescribes an action for any given belief state bB; thus, it can be employed in real time during the execution as the new belief state is calculated according to the real interactions with the environment. The major computations of the proposed policy are during the offline process; during the execution, the belief state needs to be tracked/updated, and the learned policy reflected in the trained Q-network is applied according to the latest belief.

4 Proposed Bayesian policy's complexity analysis and capabilities

In this section, we briefly describe the key differences between the proposed method and some of the well-known techniques for rescue operations. Then, we discuss the advantage and capabilities of the proposed method as well as the computational complexity for real-time implementation.

Let qθ*(s,a) be the optimal expected return for model θ defined in (2). These values can be obtained offline using dynamic programming or reinforcement learning approaches tuned for any model θ ∈ Θ. It should be noted that the model-specific Q-values are defined over the original state space and not the belief space. Assuming sk is the current state of the agent and ϑk is the posterior distribution of the environment models at time step k, the proposed optimal Bayesian policy formulated in (13) and approximated using the Q-network through (15)-(19) can be expressed in the belief space as follows:

ak=argmaxaAQ*(bk,a)argmaxaAQw([sk,ϑk],a).    (20)

Another well-known approach commonly used for learning in unknown environments is the maximum aposteriori (MAP) navigation policy, which relies on the underlying model-specific policy with the highest posterior probability. This policy can be expressed as follows:

ak=argmaxaAqθkMAP*(sk,a),    (21)

where θkMAP=argmaxθi;i=1,...,3mϑk(i).

In addition, active learning approaches (and their variations) (Silver et al., 2012; Kaplan and Friston, 2018; Choudhury and Srinivasa, 2020; Taylor et al., 2021; Rckin et al., 2022, 2023) are widely used techniques for decision-making in unknown environments. As noted in their names, these methods aim to actively lookahead and evaluate the options. A one-step procedure can be expressed as follows:

ak=argmaxaA𝔼ϑk[qθ*(sk,a)]=argmaxaAi=13mϑk(i)qθi*(sk,a).    (22)

Comparing policies in Equations (2022), one can see that the primary computation of all the methods is done in an offline process. During the online execution, all methods require updating the belief state (or the posterior probability of models), followed by using the learned/computed Q-values obtained during the offline process. The MAP policy in Equation (21) relies on a single model (i.e., a model with the highest posterior probability). Thus, if several models yield the maximum posterior probability or the model with the highest posterior is not definitively distinguished from others, the decisions made by the MAP policy become unreliable. Unlike the MAP policy, the active learning approach in Equation (22) accounts for the posterior of all models for decision-making. This policy becomes more efficient in domains where the posterior is peaked over a single model. If the posterior distribution is uniform over the models, the active learning policy looks like averaging according to various models' Q-function. The main difference between active learning and the proposed method is the incapability of active learning methods to change the posterior of the models (i.e., acting to enhance modeling information that can lead to better rewards). The active learning decisions might keep the posterior distribution unchanged over time, meaning that all models (including the wrong models) contribute similarly in making decisions over time. By contrast, the proposed Bayesian policy in Equation (20) learns the policy over the state and posterior of models, meaning that the action selection optimally influences the agent state and the posterior of models in achieving the highest accumulated rewards. This can be seen as taking actions that lead to moving to belief states under which better navigation performance can be achieved.

Aside from the efficiency of the proposed Bayesian policy described above, another advantage of the proposed policy is the generality of learning. The generality of learning refers to the fact that the proposed policy could be employed for a wide range of objectives. As described in Equations (11, 12), the reward could be defined for locating victims in the environment, quick identification of the unknown parts of the environment (i.e., changing the posterior distribution of models) or any other reward functions that can be expressed using the belief state. However, the active learning and MAP policies in Equations (21, 22) can only consider the objectives (i.e., reward functions) that are defined according to the original state space (i.e., not the posterior of models). These capabilities are investigated and discussed in the next section through various numerical experiments.

Scalability could be a limitation of our proposed method. The size of the posterior distribution in Equation (3) grows exponentially as the number of unknown cells increases in the environment, and this leads to an increase in the size of the belief space. This increases the computational complexity of the proposed policy, making it intractable for domains with uncertainty represented in continuous (or infinite-dimensional) spaces. Note that the scalability of active learning and MAP policies also increases with the uncertainty in the environment models. Our future research will investigate approaches to scale the proposed Bayesian policy to large environments with possibly large and infinite number of environment models.

5 Numerical experiments

In this section, the performance of the proposed Bayesian policy is investigated using different maze problems with the following two objectives: (1) locating the victims/injuries in unknown environments as fast as possible; (2) exploring an unknown maze environment as quickly as possible, modeled through entropy reduction. Values of all the parameters used in our numerical experiments are presented as follows: number of hidden layers 3, number of neurons in hidden layers 128, α = 5 × 10−4, |D|=105, Nbatch = 64, γ = 0.95, ϵ = 0.1, τ = 10−3, and the update frequency for the Q-network is considered to be 4. Note that all the experiments in this section are repeated for 1,000 trials, and the average results along the 95% confidence bounds are displayed in all the figures.

5.1 Locating injuries

5.1.1 6 × 4 maze problem

For our first set of experiments, we consider the 6 × 4 maze shown in Figure 2. This maze consists of 5 walls and the agent can be in any of the other 19 cells at each step. Furthermore, the 3 unknown cells indicated by yellow lead to 33 = 27 possible maze models. For the actions, the agent at each step can select right, left, down, or up. We also consider some stochasticity in the environment as the agent moves to the anticipated direction with probability of 0.8, or it will move to either of the perpendicular directions with probability of 0.1. Each of the unknown cells could contain an injury, be empty, or blocked by a wall. To guide the agent to track three potential injuries, we define three new auxiliary variables as ηk = {η1, η2, η3}, where each variable turns to 0 if an agent moves to the corresponding unknown states and locates an injury. Note that these auxiliary variables are needed to track multiple objectives in the environment. Therefore, the state space in this case contains the location of the agent (i.e., one of 19 possible locations in the maze) and the auxiliary variables. The belief space size in this case is B={{1,...,19}×{0,1}3×Δ27}. The reward function in the belief state is modeled using (11), with Rθ(s,a,s)=1 if upon moving to s′ a new injury is located according to the maze model parameterized by θ, and Rθ(s,a,s)=0 otherwise.

Here, we assume that the agent can identify empty cells and cells with injury if it moves to those states. Therefore, the transition probability for model θ required for the belief state transition in Equations (8, 9) can be expressed as follows:

P(sis,a,θ)={pssa,θ1t(si)=θ(l)if si=sulpssa,θif siSu,    (23)

where Su contains the unknown cells, sul is the lth unknown cell, and pssa,θ is 0.8 if s′ is the neighboring cell to s at the direction indicated by action a in the environment model θ and 0.1 if it is in cross perpendicular neighborhood. Note that t(si) indicates the type of state si; that is, W, E or I. In addition, 1t(si)=θ(l) is 1 if the type of state si and unknown cell θ(l) is the same and zero otherwise. Using the defined P(sis, a, θ) and Equation (9), the posterior update in our problem can be explicitly formulated. According to Equation (5), we consider the independency assumption for the unknown cells, represented through the initial distribution p0=[p01,p02,p03], where p0i=[P(ci=W),P(ci=E),P(ci=I)] consists of the prior probability of the ith unknown cell.

ReLu is used as an activation function between each layer of our neural networks. Furthermore, a maximum of 50 steps is considered for testing our proposed planning policy; however, a larger horizon of 250 steps is used for training purposes to account for the discounted rewards in the final steps. In addition, the proposed Bayesian planning policy is trained over 5,000 episodes in each case.

To better show how different prior probabilities can affect the agent's decisions, we visualized the agent's movements in the 6 × 4 maze problem using two cases in Figure 3. The prior probabilities for all unknown cells are set to be equally distributed between wall, empty, and injury (i.e., p0i=[13,13,13] for i = 1, 2, 3), except for the first unknown cell in the left maze, which is set to p01=[23,16,16]. The true environment is assumed to include a wall in the first unknown cell and injuries in the other two unknown cells. The paths selected by the agent under the proposed Bayesian policy are indicated in each maze in Figure 3. One can see that the agent moves from left in Figure 3A since it has prior knowledge that the first unknown cell is likely to be a wall. However, in Figure 3B, the agent selects the right path since it predicts that this path leads to the quickest rescue operation given the equal prior probabilities for the unknown cells. Once the agent encounters the wall in the first unknown cell, the Bayesian policy guides it to go back and reach out to other potential injuries in the environment as quickly as possible. Note that the agent's movements are also uncertain, which can be seen as the difference between the two trajectories shown in each maze in Figure 3.

Figure 3
www.frontiersin.org

Figure 3. Agent movement trajectories under the proposed Bayesian policy in the 6 × 4 maze; each of the blue and red arrows corresponds to one independent movement trajectory, and the difference in the trajectories is due to the movement stochasticity in the environment. The trajectories are recorded using the proposed policy under two different initial distributions: (A) p01=[23,16,16],p02=p03=[13,13,13], (B) p0i=[13,13,13] for i = 1, 2, 3.

The average results for the proposed policy are compared with three approaches; baseline approach, which is the reinforcement learning (i.e., Q-learning) solution when the true model of the environment is known, and MAP and active learning policies formulated in Equations (21, 22). The baseline solution is the best achievable solution if no uncertainty in the environment exists; thus, it is used to assess how uncertainty in the environment model can deviate the solutions of various policies from the solution in the known environment. Figure 4A shows the average located injuries and their confidence bounds using different navigation policies for 1,000 trials and over the first 50 steps for the maze environment shown in Figure 3A. As can be seen in Figure 4A, the proposed policy has a superior performance compared to the active learning and MAP policies as it matches the performance of the baseline policy in all the steps, and after only 20 steps it can find the two injuries in the environment successfully. Active learning policy is the next best policy in this case, and it reaches maximum performance after 50 steps. Notice that although active learning gets to all the injuries after 50 steps, it still has a very low performance in the first 30 steps, which is not desirable especially in situations where we should get to the injuries in a timely manner. In addition, we can see that the MAP policy has a better performance than active learning in the first 32 steps; however, active learning performance improves significantly after that and the MAP policy becomes the worst policy since it shows poor performance even after 50 steps.

Figure 4
www.frontiersin.org

Figure 4. (A) Performance comparison in the 6 × 4 maze with the true environment θ* = [W, I, I] and initial probabilities of unknown cells as p01=[23,16,16],p02=p03=[13,13,13]. (B) Performance comparison in the 6 × 4 maze with the true environment θ* = [I, I, I] and initial probabilities of unknown cells as p01=[23,16,16],p02=p03=[13,13,13].

In the second test environment, similar prior probabilities are used for our experiments (i.e., p01=[23,16,16],p02=p03=[13,13,13]), while in the true environment, all the unknown cells are considered to be injuries. The obtained test results for this environment are shown in Figure 4B. In this figure, we can see that the active learning and baseline policies have the same performance, and our method has the second best performance. This is because in this test case all the unknown cells are injuries, whereas the prior in our case has been set to have a higher probability for wall for the first unknown cell. Regardless of setting a faulty prior for our method, after only 20 steps, our method shows a high performance in comparison with the baseline, and the performance keeps increasing to almost the same as the baseline as the steps increase. Moreover, we can see that the MAP policy again performs poorly as to other methods.

Figure 5 shows the average results for the maze shown in Figure 3B. All the unknown cells have the same initial probability of being a wall, empty, or injury (i.e., p0i=[13,13,13] for i = 1, 2, 3). One can see that our proposed policy has a better performance than active learning and MAP policy as it reaches a performance close to the baseline in only 30 steps. Active learning also reaches a good performance after about 40 steps; however, it has much lower performance than our approach specifically between steps 18 and 40. Finally, similar to the other previous cases, it can be observed that the MAP policy has the worst performance in comparison with others.

Figure 5
www.frontiersin.org

Figure 5. Performance comparison in the 6 × 4 maze with the true environment θ* = [W, I, I] and initial probabilities of unknown cells as p0i=[13,13,13] for i = 1, 2, 3.

In the next set of experiments, the impact of different values for the parameter pssa,θ in Equation (23) on the performance of our proposed method is investigated. As described before, pssa,θ refers to the movement stochasticity in the maze; a higher value of pssa,θ means that the movement is more deterministic and the agent moves in the desired direction, and a lower value denotes that the agent movement is more stochastic and it is more likely to end up in one of the perpendicular directions. Four values of 1, 0.8, 0.6, and 0.4 are considered for pssa,θ. These values are tested on the environment of Figure 3B, with all the unknown cells having the following prior probabilities: p0i=[13,13,13] for i = 1, 2, 3. The average performance of the baseline method, the proposed method, and the MAP policy in terms of average located injuries are shown in Figure 6. Figure 6A represents the performance of different policies at timestep 50. It can be seen that the best performance of the proposed policy is achieved at pssa,θ=1, where it has performed similarly to the baseline. The second best performance for the proposed policy occurs under the value pssa,θ=0.8, where our policy has slightly smaller performance than the baseline policy. The performance of our method under the values 0.6 and 0.4 is lower and almost similar in both cases since both values represent environments with high movement stochasticity. Moreover, it can be seen that the worst performance is achieved by the MAP in all the cases. Furthermore, Figure 6B shows the performance after 100 timesteps. In this case, our method's performance under the values of 0.8, 0.6, and 0.4 gets much closer to the baseline policy, whereas the MAP policy still performs poorly. This clearly demonstrates the robustness of our proposed method under different stochasticity levels in the environment.

Figure 6
www.frontiersin.org

Figure 6. Impact of movement stochasticity (pssa,θ) on the performance of different policies in the 6 × 4 maze with the true environment θ* = [W, I, I] and initial probabilities of unknown cells as p0i=[13,13,13] for i = 1, 2, 3. Higher values for pssa,θ correspond to more deterministic movements whereas lower values refer to more stochastic movements. The performance of different policies is shown in two timesteps: (A) Timestep 50 and (B) Timestep 100.

For further analysis, the impact of different unknown cells is studied in the next set of experiments. We consider three variations of the 6 × 4 maze problem. Figures 7AC represent the environments with two, three, and four unknown cells, respectively. A uniform prior probability is used for all the unknown cells during training, that is, [13,13,13]. Furthermore, for a fair comparison of these environments during the test, two of the unknown cells are assumed to contain victims/injuries as demonstrated in Figure 7, and the other unknown cells in environments (Figures 7B, C) can be either wall or empty. The average number of located injuries achieved in all the environments at timesteps 10, 20, and 30 using different policies is reported in Table 1. The performance of our proposed Bayesian policy is indicated by bold numbers. This table shows that at each timestep, the performance of our method decreases as the number of unknown cells (and the complexity of the maze) increases. Furthermore, one can see that the performance of our policy in all the environments and at all the timesteps surpasses other policies' performance, indicating the effectiveness of our approach. This indicates that our method is capable of reasoning about additional uncertainty in the environments to still make effective decisions.

Figure 7
www.frontiersin.org

Figure 7. Visualization of the 6 × 4 grid problem with different numbers of unknown cells: (A) Two unknown cells. (B) Three unknown cells. (C) Four unknown cells. For a more fair comparison, as shown in all the grids, two of the unknown cells are considered to be victims/injuries in the true underlying environment (θ*) when performing the tests, and the rest of the unknown cells in (B, C) could be wall or empty.

Table 1
www.frontiersin.org

Table 1. Average number of located injuries by different policies for locating two injuries in the 6 × 4 maze with different numbers of unknown cells.

5.1.2 4 × 4 maze problem

In this part, our policy is tested for locating injuries in a new 4 × 4 grid which is a smaller grid than the previous 6 × 4 maze problem. This grid, visualized in Figure 8A, has three unknown cells, leading to 27 possible environment models. The state space of this maze can be written as S={s1,...,s13}. Considering three auxiliary variables ηk = {η1, η2, η3} for tracking multiple targets, the belief space size in this problem can be represented as B={{1,...,13}×{0,1}3×Δ27}. For training our proposed policy, we consider a case where the priors for the unknown cells are as follows: p0i=[13,13,13] for i = 1, 2, 3. Furthermore, for testing different policies, a true underlying environment with two injuries and one wall is chosen as represented in Figure 8B. Figure 8C shows the average located injuries achieved by different policies. It is shown that our method has the best performance in all the steps and exhibits a more similar behavior to the baseline policy compared to active learning and MAP policies.

Figure 8
www.frontiersin.org

Figure 8. (A) Visualization of the 4 × 4 maze problem with three unknown cells. (B) The true test environment θ* = [I, W, I] with the initial probabilities of unknown cells as p0i=[13,13,13] for i = 1, 2, 3. (C) Performance comparison of different policies for locating the injuries in the test environment.

5.1.3 6 × 6 maze problem

In this subsection, we study the performance of the proposed method for locating injuries in a 6 × 6 maze with three unknown cells as depicted in Figure 9A, which is a larger maze than the previous two maze problems. This maze has seven walls; this leads to the following state and belief spaces for this maze: S={s1,...,s29}, B={{1,...,29}×{0,1}3×Δ27}. The prior probabilities of different unknown cells are assumed to be uniform during the training of our policy. Moreover, a maze with two injuries and one wall as shown in Figure 9B is considered for the tests. The performance of our policy, MAP policy, and active learning policy for this experiment is shown in Figure 9C. Our proposed policy reaches a higher value in all the steps in terms of average located injuries compared to the MAP and active learning policies. This figure also denotes that the performance of our method matches the performance of the baseline policy after 50 steps (almost 2), whereas the MAP and active learning policies show a low performance after the same 50 steps (less than 0.5).

Figure 9
www.frontiersin.org

Figure 9. (A) Visualization of the 6 × 6 maze problem with three unknown cells. (B) The true test environment θ* = [W, I, I] with the initial probabilities of unknown cells as p0i=[13,13,13] for i = 1, 2, 3. (C) Performance comparison of different policies for locating the injuries in the test environment.

5.2 Entropy reduction

5.2.1 6 × 4 maze problem

In this part of numerical experiments, we consider the same 6 × 4 maze problem with the objective of quick exploration of an unknown environment. Thus, the reward function in Equation (12) is considered for our experiments, which guides the agent to quickly navigate in the unknown environment and rapidly reduce the overall uncertainty/entropy in the environment models. The state part of the belief can be expressed using 19 potential states for the agent locations without the need for the previously defined auxiliary variables (i.e., used for tracking the injuries). The belief space, in this case, is B={{1,...,19}×Δ27}, which is still a large space. Note that in this case, the previously used approaches to compare with our proposed policy cannot be employed. This is due to the fact that the entropy reduction cannot be expressed in terms of the reward for underlying state space. In fact, tracking the posterior probabilities of the environment models is required in this case, which is not possible with the MAP and active learning approaches represented in Equations (21, 22). A one-step entropy reduction policy is employed instead for comparison purposes. This policy selects actions to maximally reduce the next step entropy as

ak=argmaxaA-𝔼b=[s,ϑ]Tb,a[H(ϑ)-H(ϑ)].    (24)

We consider the following initial probabilities for unknown cells: p0i=[12,14,14] for i = 1, 2, 3. Based on these priors, the starting value of entropy will be 3.12. Our proposed policy is then tested on two different environments. The first true environment is θ* = [E, E, E], where all the unknown cells are empty. The average negative entropy at each step of the test, along with the results of one-step entropy reduction, is shown in Figure 10A. The proposed policy obtains a much faster reduction in the entropy value than the one-step entropy reduction. Moreover, our approach reaches an entropy of 0 (least uncertainty in the environment) in only 20 steps, whereas the one-step entropy reduction shows poor performance until step 50. This demonstrates the importance of accounting for the possible future entropy in making decisions, as opposed to greedy one-step search, which is achieved by the proposed method with policy defined over the belief state. The second test environment consists of all walls in the unknown cells. Figure 10B presents the results of our method and the one-step entropy reduction policy. One can see that the entropy reduction is slower compared to the first test case in Figure 10A, which is due to the fact that all unknown cells are walls and the agent needs much larger time to visit them and reduce the overall uncertainty. However, our proposed approach, again, has achieved a much faster reduction in entropy compared to the one-step entropy reduction policy.

Figure 10
www.frontiersin.org

Figure 10. (A) Performance comparison in the 6 × 4 maze with the true environment θ* = [E, E, E] and initial probabilities of unknown cells as p0i=[12,14,14] for i = 1, 2, 3. (B) Performance comparison in the 6 × 4 maze with the true environment θ* = [W, W, W] and initial probabilities of unknown cells as p0i=[12,14,14] for i = 1, 2, 3.

5.2.2 10 × 10 maze problem

A larger 10 × 10 maze shown in Figure 11 is considered for this part of our numerical experiments. This maze contains four unknown cells, indicated by the color yellow in Figure 11, resulting in 34 = 81 possible maze models. In this part, we aim to tackle the problem of entropy reduction using this larger maze. This maze has 33 walls, so the state space can be represented by S={s1,...,s67}. The belief space is B={S×Δ81}, which is much larger than the previous maze problems. We consider hyperparameters and maze stochasticity similar to previous problems, except that since this is a larger maze, we consider a horizon of 500 steps for the training process. Finally, 20,000 episodes are used to train our policy in this larger maze.

Figure 11
www.frontiersin.org

Figure 11. Visualization of the 10 × 10 maze problem. This maze has four unknown cells, where each could be either wall, empty, or victim/injury.

The reward function in Equation (12) is used for this part of the experiments, meaning that it is desired for an agent to visit all the unknown cells as quickly as possible and reduce the overall uncertainty in the maze model to 0. Figure 12 represents two independent paths taken by the agent under the proposed Bayesian policy for two different initial probabilities: (a) first unknown cell is set to have a higher chance of being a wall, and the other three unknown cells have a higher chance of being empty or injury, that is, p01=[810,110,110],p0i=[112,612,512] for i = 2, 3, 4; (b) all the unknown cells have a higher chance of being empty or injury, that is, p0i=[112,612,512] for i = 1, 2, 3, 4. One can see that in case (a), the agent selects the left path to explore/visit all the unknown cells because unknown cell number 1 has more probability of being a wall and the Bayesian policy predicts a high likelihood that the agent might be stuck in the right side of the maze. However, in case (b), where all the unknown cells have more probability to be an injury or empty, the agent decides to choose the shortest path to get to all the unknown cells, which means that at first it goes from the right side of the maze. This again demonstrates the capability of the proposed method in accounting for potential uncertainty arising from agent movement and model uncertainty for making decisions.

Figure 12
www.frontiersin.org

Figure 12. Agent movement trajectories under the proposed Bayesian policy in the 10 × 10 maze; each of the blue and red arrows corresponds to one independent movement trajectory, and the difference in the trajectories is due to the movement stochasticity in the environment. The trajectories are recorded using the proposed policy under two different initial distributions: (A) p01=[810,110,110] and p0i=[112,612,512] for i = 2, 3, 4, (B) p0i=[112,612,512] for i = 1, 2, 3, 4.

Once again, similar to Figure 12B, consider that the prior probabilities are set equally for all the unknown cells as: p0i=[112,612,512] for i = 1, 2, 3, 4, which means that each cell has probability of 112 to be a wall. The starting entropy for this case is equal to 3.67. For this case, the average results of our proposed policy and one-step entropy reduction policy for the true environment θ* = [I, I, I, I] are shown in Figure 13A. Our proposed policy outperforms the one-step entropy reduction results by a far margin in all the steps, and it reaches an entropy of zero after only 40 steps. As a final scenario, only the prior probability of being a wall for the first unknown cell is set to be larger as opposed to the previous case, that is, p01=[810,110,110] (similar to Figure 12A). The starting entropy value, in this case, is 3.39, and the test results on θ* = [I, I, I, I] are presented in Figure 13B. As seen in this figure, our policy has a better performance in all the steps compared to one-step entropy reduction. Our navigation policy results in zero entropy after about 45 steps, where the one-step entropy reduction fails to achieve good performance over all 100 steps.

Figure 13
www.frontiersin.org

Figure 13. (A) Performance comparison in the 10 × 10 maze with the true environment θ* = [I, I, I, I] and initial probabilities of unknown cells as p0i=[112,612,512] for i = 1, 2, 3, 4. (B) Performance comparison in the 10 × 10 maze with the true environment θ* = [I, I, I, I], and initial probabilities of unknown cells as p01=[810,110,110] and p0i=[112,612,512] for i = 2, 3, 4.

6 Conclusion and future works

This study developed a reinforcement learning Bayesian planning policy for rescue operations in unknown or partially known environments. Unlike most existing approaches that rely on the availability of an exact model or simulator for representing the underlying environment, this study considers realistic problems in which no or limited prior information about the environment might be available. A new Bayesian formulation of navigation in unknown and uncertain environments is provided using the definition of belief state, which tracks the agent state and the uncertainty of the environment up to each step. This formulation is used to formulate the optimal Bayesian policy, which can be computed through the propagation of all the uncertainty in the agent state and environment models. A solution to the optimal Bayesian policy is introduced using a deep reinforcement learning method. Finally, the performance of the proposed method is demonstrated using different maze problems with various uncertainties. Note that the proposed method is versatile and can be applied not only to environments with discrete state spaces but also to those with continuous state spaces or large discrete state spaces.

The computational complexity of the proposed method increases exponentially with the number of unknown cells in the environments, potentially limiting its scalability. Our future work includes studying the scalability of the proposed policies in domains with extremely large belief spaces and different uncertainties in the model. We will also extend the idea of Bayesian planning to domains with partially observable states, as well as domains with multiple agents and continuous action spaces.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

MA: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing. MI: Formal analysis, Funding acquisition, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing, Conceptualization, Investigation.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The authors acknowledge the support of the National Science Foundation awards IIS-2311969 and IIS-2202395, ARMY Research Laboratory award W911NF2320179, ARMY Research Office award W911NF2110299, and Office of Naval Research award N00014-23-1-2850.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akcakoca, M., Atici, B. M., Gever, B., Oguz, S., Demirezen, U., Demir, M., et al. (2019). “A simulation-based development and verification architecture for micro UAV teams and swarms,” in AIAA Scitech 2019 Forum, 1979.

Google Scholar

Alali, M., and Imani, M. (2023). “Reinforcement learning data-acquiring for causal inference of regulatory networks,” in 2023 American Control Conference (ACC) (IEEE), 3957–3964.

PubMed Abstract | Google Scholar

Alali, M., and Imani, M. (2024). Bayesian lookahead perturbation policy for inference of regulatory networks. IEEE/ACM Trans. Comput. Biol. Bioinform. doi: 10.1109/TCBB.2024.3402220

PubMed Abstract | Crossref Full Text | Google Scholar

Alali, M., Kazeminajafabadi, A., and Imani, M. (2024). Deep reinforcement learning sensor scheduling for effective monitoring of dynamical systems. Syst. Sci. Cont. Eng. 12:2329260. doi: 10.1080/21642583.2024.2329260

PubMed Abstract | Crossref Full Text | Google Scholar

Asadi, N., Hosseini, S. H., Imani, M., Aldrich, D., and Ghoreishi, S. F. (2024). “Privacy-preserved federated reinforcement learning for autonomy in signalized intersections,” in ASCE International Conference on Transportation and Development (ICTD). Reston: American Society of Civil Engineers.

Google Scholar

Bajcsy, A., Bansal, S., Bronstein, E., Tolani, V., and Tomlin, C. J. (2019). “An efficient reachability-based framework for provably safe autonomous navigation in unknown environments,” in 2019 IEEE 58th Conference on Decision and Control (CDC) (Nice: IEEE), 1758–1765.

Google Scholar

Blum, T., Paillet, G., Laine, M., and Yoshida, K. (2020). Rl star platform: Reinforcement learning for simulation based training of robots. arXiv [Preprint]. arXiv:2009.09595. doi: 10.48550/arXiv.2009.09595

Crossref Full Text | Google Scholar

Bøhn, E., Coates, E. M., Moe, S., and Johansen, T. A. (2019). “Deep reinforcement learning attitude control of fixed-wing UAVs using proximal policy optimization,” in 2019 International Conference on Unmanned Aircraft Systems (ICUAS) (Atlanta, GA: IEEE), 523–533.

Google Scholar

Boulares, M., and Barnawi, A. (2021). A novel UAV path planning algorithm to search for floating objects on the ocean surface based on object's trajectory prediction by regression. Rob. Auton. Syst. 135:103673. doi: 10.1016/j.robot.2020.103673

Crossref Full Text | Google Scholar

Bouman, A., Ginting, M. F., Alatur, N., Palieri, M., Fan, D. D., Touma, T., et al. (2020). “Autonomous spot: Long-range autonomous exploration of extreme environments with legged locomotion,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Las Vegas, NV: IEEE), 2518–2525.

Google Scholar

Cabreira, T. M., Brisolara, L. B., and Paulo, R. F. J.. (2019). Survey on coverage path planning with unmanned aerial vehicles. Drones 3:4. doi: 10.3390/drones3010004

Crossref Full Text | Google Scholar

Chang, L., Shan, L., Jiang, C., and Dai, Y. (2021). Reinforcement based mobile robot path planning with improved dynamic window approach in unknown environment. Auton. Robots 45:51–76. doi: 10.1007/s10514-020-09947-4

Crossref Full Text | Google Scholar

Choudhury, S., and Srinivasa, S. S. (2020). “A bayesian active learning approach to adaptive motion planning,” in Robotics Research, N. M. Amato, G. Hager, S. Thomas, and M. Torres-Torriti (Cham: Springer International Publishing), 33–40.

Google Scholar

de Almeida, J. P. L. S., Nakashima, R. T., Neves-Jr, F., and de Arruda, L. V. R. (2019). Bio-inspired on-line path planner for cooperative exploration of unknown environment by a multi-robot system. Rob. Auton. Syst. 112, 32–48. doi: 10.1016/j.robot.2018.11.005

Crossref Full Text | Google Scholar

Ebrahimi, D., Sharafeddine, S., Ho, P.-H., and Assi, C. (2020). Autonomous UAV trajectory for localizing ground objects: A reinforcement learning approach. IEEE Trans. Mobile Comp. 20, 1312–1324. doi: 10.1109/TMC.2020.2966989

Crossref Full Text | Google Scholar

Elguea-Aguinaco, I., Serrano-Muoz, A., Chrysostomou, D., Inziarte-Hidalgo, I., Bh, S., and Arana-Arexolaleiba, N. (2023). A review on reinforcement learning for contact-rich robotic manipulation tasks. Robot. Comput. Integr. Manuf. 81:102517. doi: 10.1016/j.rcim.2022.102517

Crossref Full Text | Google Scholar

Falcone, G., and Putnam, Z. R. (2022). “Deep reinforcement learning for autonomous aerobraking maneuver planning,” in AIAA SCITECH 2022 Forum, 2497.

Google Scholar

Feng, J., Lee, J., Durner, M., and Triebel, R. (2022). “Bayesian active learning for sim-to-real robotic perception,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Kyoto: IEEE), 10820–10827.

Google Scholar

Ghavamzadeh, M., Mannor, S., Pineau, J., and Tamar, A. (2015). Bayesian reinforcement learning: a survey. Found. Trends Mach. Learn. 8, 359–483. doi: 10.1561/2200000049

Crossref Full Text | Google Scholar

Greatwood, C., and Richards, A. G. (2019). Reinforcement learning and model predictive control for robust embedded quadrotor guidance and control. Auton. Robots 43, 1681–1693. doi: 10.1007/s10514-019-09829-4

Crossref Full Text | Google Scholar

Guez, A., Silver, D., and Dayan, P. (2012). “Efficient Bayes-adaptive reinforcement learning using sample-based search,” in Advances in Neural Information Processing Systems 25 (Red Hook, NY: Curran Associates Inc.).

Google Scholar

Hamid, S., Nasir, A., and Saleem, Y. (2021). Reinforcement learning based hierarchical multi-agent robotic search team in uncertain environment. Mehran Univer. Res. J. Eng. Technol. 40, 645–662. doi: 10.22581/muet1982.2103.17

Crossref Full Text | Google Scholar

Hu, J., Niu, H., Carrasco, J., Lennox, B., and Arvin, F. (2020). Voronoi-based multi-robot autonomous exploration in unknown environments via deep reinforcement learning. IEEE Trans. Vehicul. Technol. 69, 14413–14423. doi: 10.1109/TVT.2020.3034800

PubMed Abstract | Crossref Full Text | Google Scholar

Imanberdiyev, N., Fu, C., Kayacan, E., and Chen, I.-M. (2016). “Autonomous navigation of UAV by using real-time model-based reinforcement learning,” in 2016 14th International Conference on Control, Automation, Robotics and Vision (ICARCV) (Phuket: IEEE), 1–6.

Google Scholar

Imani, M., and Ghoreishi, S. F. (2022). Scalable inverse reinforcement learning through multifidelity bayesian optimization. IEEE Trans. Neural Netw. Learn. Syst. 33, 4125–4132. doi: 10.1109/TNNLS.2021.3051012

PubMed Abstract | Crossref Full Text | Google Scholar

Imani, M., Ghoreishi, S. F., and Braga-Neto, U. M. (2018). “Bayesian control of large MDPs with unknown dynamics in data-poor environments,” in Advances in Neural Information Processing Systems (Curran Associates Inc.), 8156-8166.

Google Scholar

Jagannath, J., Jagannath, A., Furman, S., and Gwin, T. (2021). Deep Learning and Reinforcement Learning for Autonomous Unmanned Aerial Systems: Roadmap for Theory to Deployment. Cham: Springer International Publishing, 25–82.

Google Scholar

Juang, C.-F., and Chang, Y.-C. (2011). Evolutionary-group-based particle-swarm-optimized fuzzy controller with application to mobile-robot navigation in unknown environments. IEEE Transact. Fuzzy Syst. 19, 379–392. doi: 10.1109/TFUZZ.2011.2104364

Crossref Full Text | Google Scholar

Kamthe, S., and Deisenroth, M. (2018). “Data-efficient reinforcement learning with probabilistic model predictive control,” in International Conference on Artificial Intelligence and Statistics (New York: PMLR), 1701–1710.

Google Scholar

Kaplan, R., and Friston, K. J. (2018). Planning and navigation as active inference. Biol. Cybern. 112, 323–343. doi: 10.1007/s00422-018-0753-2

PubMed Abstract | Crossref Full Text | Google Scholar

Kim, S.-K., Bouman, A., Salhotra, G., Fan, D. D., Otsu, K., Burdick, J., et al. (2021). “Plgrim: Hierarchical value learning for large-scale exploration in unknown environments,” in Proceedings of the International Conference on Automated Planning and Scheduling, 652–662.

Google Scholar

Kim, S.-K., Thakker, R., and Agha-Mohammadi, A.-A. (2019). Bi-directional value learning for risk-aware planning under uncertainty. IEEE Robot. Automat. Lett. 4, 2493–2500. doi: 10.1109/LRA.2019.2903259

Crossref Full Text | Google Scholar

Kingma, D. P., and Ba, J. (2015). “Adam: a method for stochastic optimization,” in CoRR, abs/1412.6980.

Google Scholar

Krell, E., Sheta, A., Balasubramanian, A. P. R., and King, S. A. (2019). Collision-free autonomous robot navigation in unknown environments utilizing pso for path planning. J. Artif. Intell. Soft Comput. Res. 9:8. doi: 10.2478/jaiscr-2019-0008

Crossref Full Text | Google Scholar

Ladosz, P., Oh, H., Zheng, G., and Chen, W.-H. (2019). A hybrid approach of learning and model-based channel prediction for communication relay UAVs in dynamic urban environments. IEEE Robot. Automat. Lett. 4, 2370–2377. doi: 10.1109/LRA.2019.2903850

Crossref Full Text | Google Scholar

Li, H., Zhang, Q., and Zhao, D. (2019). Deep reinforcement learning-based automatic exploration for navigation in unknown environment. IEEE Trans. Neural Netw. Learn. Syst. 31, 2064–2076. doi: 10.1109/TNNLS.2019.2927869

PubMed Abstract | Crossref Full Text | Google Scholar

Lin, J., Yang, X., Zheng, P., and Cheng, H. (2019). “End-to-end decentralized multi-robot navigation in unknown complex environments via deep reinforcement learning,” in 2019 IEEE International Conference on Mechatronics and Automation (ICMA) (Tianjin: IEEE), 2493–2500.

Google Scholar

Lin, Y., Ghoreishi, S. F., Lan, T., and Imani, M. (2024). “High-level human intention learning for cooperative decision-making,” in IEEE Conference on Control Technology and Applications (CCTA) (Newcastle upon Tyne: IEEE).

PubMed Abstract | Google Scholar

Luo, C., Gao, J., Li, X., Mo, H., and Jiang, Q. (2014). “Sensor-based autonomous robot navigation under unknown environments with grid map representation,” in 2014 IEEE Symposium on Swarm Intelligence (Orlando, FL: IEEE), 1–7.

Google Scholar

Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G., et al. (2015). Human-level control through deep reinforcement learning. Nature 518:529. doi: 10.1038/nature14236

PubMed Abstract | Crossref Full Text | Google Scholar

Niroui, F., Zhang, K., Kashino, Z., and Nejat, G. (2019). Deep reinforcement learning robot for search and rescue applications: exploration in unknown cluttered environments. IEEE Robot. Automat. Lett. 4, 610–617. doi: 10.1109/LRA.2019.2891991

Crossref Full Text | Google Scholar

Perez-Imaz, H. I., Rezeck, P. A., Macharet, D. G., and Campos, M. F. (2016). “Multi-robot 3d coverage path planning for first responders teams,” in 2016 IEEE International Conference on Automation Science and Engineering (CASE) (Fort Worth, TX: IEEE), 1374–1379.

Google Scholar

Pham, H. X., La, H. M., Feil-Seifer, D., and Van Nguyen, L. (2018). “Reinforcement learning for autonomous UAV navigation using function approximation,” in 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR) (Philadelphia, PA: IEEE), 1–6.

Google Scholar

Ravari, A., Ghoreishi, S. F., and Imani, M. (2023). Optimal recursive expert-enabled inference in regulatory networks. IEEE Cont. Syst. Lett. 7, 1027–1032. doi: 10.1109/LCSYS.2022.3229054

PubMed Abstract | Crossref Full Text | Google Scholar

Ravari, A., Ghoreishi, S. F., and Imani, M. (2024a). “Implicit human perception learning in complex and unknown environments,” in American Control Conference (ACC) (San Diego, CA: IEEE).

Google Scholar

Ravari, A., Ghoreishi, S. F., and Imani, M. (2024b). “Optimal inference of hidden Markov models through expert-acquired data,” in IEEE Transactions on Artificial Intelligence (IEEE).

Google Scholar

Rckin, J., Jin, L., Magistri, F., Stachniss, C., and Popovi, M. (2022). “Informative path planning for active learning in aerial semantic mapping,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Kyoto: IEEE), 11932–11939.

Google Scholar

Rckin, J., Magistri, F., Stachniss, C., and Popovi, M. (2023). An informative path planning framework for active learning in uav-based semantic mapping. IEEE Trans. Robot. 39, 4279–4296. doi: 10.1109/TRO.2023.3313811

Crossref Full Text | Google Scholar

Richter, C., and Roy, N. (2017). “Learning to plan for visibility in navigation of unknown environments,” in International Symposium on Experimental Robotics (Cham: Springer), 387–398.

Google Scholar

Rigter, M., Lacerda, B., and Hawes, N. (2021). Risk-averse Bayes-adaptive reinforcement learning. Adv. Neural Inf. Process. Syst. 34, 1142–1154. doi: 10.48550/arXiv.2102.05762

Crossref Full Text | Google Scholar

Rosolia, U., Chen, Y., Daftry, S., Ono, M., Yue, Y., and Ames, A. D. (2022). “The mixed-observable constrained linear quadratic regulator problem: the exact solution and practical algorithms,” in IEEE Transactions on Automatic Control (IEEE).

Google Scholar

Rothfuss, J., Sukhija, B., Treven, L., Dörfler, F., Coros, S., and Krause, A. (2024). Bridging the sim-to-real gap with Bayesian inference. arXiv [Preprint]. arXiv:2403.16644. doi: 10.48550/arXiv.2403.16644

Crossref Full Text | Google Scholar

Sampedro, C., Rodriguez-Ramos, A., Bavle, H., Carrio, A., de la Puente, P., and Campoy, P. (2019). A fully-autonomous aerial robot for search and rescue applications in indoor environments using learning-based techniques. J. Intellig. Robot. Syst. 95, 601–627. doi: 10.1007/s10846-018-0898-1

Crossref Full Text | Google Scholar

Silver, D., Bagnell, J. A., and Stentz, A. (2012). “Active learning from demonstration for robust autonomous navigation,” in 2012 IEEE International Conference on Robotics and Automation (Saint Paul: IEEE), 200–207.

PubMed Abstract | Google Scholar

Taylor, A. T., Berrueta, T. A., and Murphey, T. D. (2021). Active learning in robotics: a review of control principles. Mechatronics 77:102576. doi: 10.1016/j.mechatronics.2021.102576

Crossref Full Text | Google Scholar

Tordesillas, J., Lopez, B. T., and How, J. P. (2019). “Faster: Fast and safe trajectory planner for flights in unknown environments,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Macau: IEEE), 1934–1940.

Google Scholar

Wang, C., Wang, J., Shen, Y., and Zhang, X. (2019). Autonomous navigation of UAVs in large-scale complex environments: a deep reinforcement learning approach. IEEE Trans. Vehic. Technol. 68, 2124–2136. doi: 10.1109/TVT.2018.2890773

Crossref Full Text | Google Scholar

Wu, J., Song, C., Ma, J., Wu, J., and Han, G. (2021). Reinforcement learning and particle swarm optimization supporting real-time rescue assignments for multiple autonomous underwater vehicles. IEEE Trans. Intellig. Transp. Syst. 23, 6807–6820. doi: 10.1109/TITS.2021.3062500

Crossref Full Text | Google Scholar

Xu, J., Guo, Q., Xiao, L., Li, Z., and Zhang, G. (2019). “Autonomous decision-making method for combat mission of UAV based on deep reinforcement learning,” in 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC) (Chengdu: IEEE), 538–544.

Google Scholar

Zhang, K., Niroui, F., Ficocelli, M., and Nejat, G. (2018). “Robot navigation of environments with unknown rough terrain using deep reinforcement learning,” in 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR) (Philadelphia, PA: IEEE), 1–7.

Google Scholar

Zhang, Y., Gong, D.-w., and Zhang, J.-h. (2013). Robot path planning in uncertain environment using multi-objective particle swarm optimization. Neurocomputing 103, 172–185. doi: 10.1016/j.neucom.2012.09.019

Crossref Full Text | Google Scholar

Zhang, Z., Imani, M., and Lan, T. (2024a). Modeling other players with Bayesian beliefs for games with incomplete information. arXiv [preprint] arXiv:2405.14122.

Google Scholar

Zhang, Z., Zhou, H., Imani, M., Lee, T., and Lan, T. (2024b). Collaborative AI teaming in unknown environments via active goal deduction. arXiv [preprint] arXiv:2403.15341.

Google Scholar

Zintgraf, L., Schulze, S., Lu, C., Feng, L., Igl, M., Shiarlis, K., et al. (2021). VariBAD: a very good method for Bayes-adaptive deep RL via meta-learning. J. Mach. Learn. Res. 22, 1–39. doi: 10.1145/3580305.3599254

Crossref Full Text | Google Scholar

Appendix

In this section, we provide the run times of our proposed method and the compared approaches for selected experiments to highlight the limitations and strengths of our method. Table A1 showcases the run times of the experiments using different methods. The reported times were recorded using a system with two 2.4 GHz Intel E5-2680 v4 CPUs and 64 GB RAM memory. As can be seen in Table A1, the run times of different approaches are divided into training times and test/execution times. The training time for our method refers to the time needed for training the weights of the neural network, and the training time for active learning and MAP refers to the time needed for training the policies for each of the possible environments in each experiment. Note that active learning and MAP use the same trained policies during the execution, therefore they have similar training times. The test/execution time is the time that was used to record the results over 1,000 trials for different methods. From Table A1, one can see that our method is much faster in terms of training compared to active learning and MAP. The execution time for different methods is however in the same range, with our method being faster overall throughout all the experiments. Further, it can be seen that the complexity of the problem increases as the number of unknown cells (Table 1, Four Unknown Cells) and the grid size (Figure 9) increases, and this naturally results in longer training and test/execution times as demonstrated in Table A1.

Table A1
www.frontiersin.org

Table A1. Training and execution times of different policies in selected experiments.

Keywords: rescue operations, Markov decision process, reinforcement learning, Bayesian decision-making, navigation planning

Citation: Alali M and Imani M (2024) Bayesian reinforcement learning for navigation planning in unknown environments. Front. Artif. Intell. 7:1308031. doi: 10.3389/frai.2024.1308031

Received: 05 October 2023; Accepted: 19 June 2024;
Published: 04 July 2024.

Edited by:

Mayukh Das, Microsoft Research, India

Reviewed by:

Anjaly Parayil, Microsoft Research, India
Xian Yeow Lee, Hitachi America Ltd, United States

Copyright © 2024 Alali and Imani. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohammad Alali, alali.m@northeastern.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.