Skip to main content

ORIGINAL RESEARCH article

Front. Neurorobot., 31 May 2022
This article is part of the Research Topic Human Inspired Robotic Intelligence and Structure in Demanding Environments View all 8 articles

A Supervised-Reinforced Successive Training Framework for a Fuzzy Inference System and Its Application in Robotic Odor Source Searching

  • 1Shenzhen Key Laboratory of Biomimetic Robotics and Intelligent Systems, Shenzhen, China
  • 2Guangdong Provincial Key Laboratory of Human-Augmentation and Rehabilitation Robotics in Universities, Southern University of Science and Technology, Shenzhen, China

Fuzzy inference systems have been widely applied in robotic control. Previous studies proposed various methods to tune the fuzzy rules and the parameters of the membership functions (MFs). Training the systems with only supervised learning requires a large amount of input-output data, and the performance of the trained system is confined by that of the target system. Training the systems with only reinforcement learning (RL) does not require prior knowledge but is time-consuming, and the initialization of the system remains a problem. In this paper, a supervised-reinforced successive training framework is proposed for a multi-continuous-output fuzzy inference system (MCOFIS). The parameters of the fuzzy inference system are first tuned by a limited number of input-output data from an existing controller with supervised training and then are utilized to initialize the system in the reinforcement training stage. The proposed framework is applied in a robotic odor source searching task and the evaluation results demonstrate that the performance of the fuzzy inference system trained by the successive framework is superior to the systems trained by only supervised learning or RL. The system trained by the proposed framework can achieve around a 10% higher success rate compared to the systems trained by only supervised learning or RL.

1. Introduction

Fuzzy inference systems have been applied in various classification and regression problems in machine learning (Nguyen et al., 2019; Wu et al., 2019; Cui et al., 2020) and have also been widely used in control and optimization in robotics (Chen C. et al., 2022; Su et al., 2022). Previous studies have proposed several methods to learn and tune the fuzzy rules and the parameters of the membership functions (MFs) to achieve the expected performance. Some widely applied methods design the fuzzy systems from (1) a manually-built fuzzy rule look-up table (Chen and Huang, 2020b); (2) learning from collected input-output data (Wang et al., 2020) through evolutionary algorithms (Wu and Tan, 2006) and gradient descent (Wang and Mendel, 1992).

Unfortunately, in unknown environments, prior knowledge may not be sufficient to build well-designed fuzzy rules, and the parameters of the system can hardly be tuned to an optimal solution (Dai et al., 2005). In terms of learning from collected data, a typical work is that Wang and Pang (2020) proposed to train adaptive neural fuzzy inference systems (ANFIS) to mimic existing bio-inspired controllers and probabilistic controllers for odor source searching utilizing collected input-output data and realize behavior patterns similar to the target controller. The performance of the trained fuzzy inference system-based controller can be further improved by fusing the input-output data of two different controllers.

The above learning process is in the scope of supervised training, in which a large amount of training data is required. Data collection can be time-consuming and the data collected from limited environmental settings may not include boundary conditions. In addition, for some complex environments, existing controllers may not be optimal. Learning from them cannot necessarily achieve the desired performance.

Reinforcement learning (RL) has attracted researchers' attention in the past decades because it provides an effective solution to robotic control and decision-making problems for which analytically optimal solutions are hard to obtain. RL is based on a human-inspired “trial-and-error” learning process that action will be reinforced if it is followed by a desired state of the robot. Since RL can tune the controllers in real-time, correct action or trajectory data is not required. Therefore, RL is especially suitable to operate in a knowledge-poor environment.

In previous studies, the fuzzy inference system has been integrated into RL in various application scenarios because of its high interpretability and flexibility. Kumar et al. (2020) used a fuzzy inference system to switch between three working modes for the traffic light control system, while a deep RL model was designed to switch the traffic lights. The fuzzy inference system and the RL model worked in a hierarchical framework. Wang et al. (2021) integrated a fuzzy inference system into the reward function of the RL model to balance the exploitation and exploration during odor source searching. Er and Deng (2004) proposed a fuzzy Q learning method to tune a fuzzy inference system-based actor model by RL, and similar methods have been applied in autonomous vehicle control (Dai et al., 2005) and robotic odor source searching (Chen and Huang, 2019; Chen X. et al., 2022).

Previous studies usually initialized the parameter of the fuzzy inference system with conventional clustering methods (Cui et al., 2020; Wang et al., 2020) or arbitrarily manual settings. Although RL can tune a fuzzy inference system to achieve a good performance, it remains an interesting problem to investigate whether the initial parameter setting of the fuzzy inference system will affect the performance of the system after numerous training epochs. To the best of our knowledge, no previous studies focused on this problem and provided a good solution to initialize the fuzzy inference system so that it can achieve better performance after training.

In this paper, a supervised-reinforced successive training framework for a multi-continuous-output fuzzy inference system (MCOFIS) was proposed. In this framework, the MCOFIS was first trained with input-output data from an existing state-action model. The input-output data was collected in multiple robotic tasks, in which the robot was running a pre-designed controller. The measured state of the environment and the resulting actions of the robot at each time step were recorded as the input-output data. After this supervised training stage, the trained MCOFIS model was utilized as the initial model in the process of reinforcement training and further trained with the deep deterministic policy gradient (DDPG) RL algorithm (Lillicrap et al., 2015). The proposed training framework was applied in a robotic odor source searching problem, which was usually solved by bio-inspired reactive algorithms (Shigaki et al., 2019), probabilistic algorithms (Vergassola et al., 2007; Chen and Huang, 2020a; Chen et al., 2020), and learning algorithms (Wang and Pang, 2020; Chen et al., 2021a) in previous studies. The performance of the trained MCOFIS-based odor source searching controller was compared with the MCOFIS-based controller trained with RL only. The results showed that the MCOFIS trained with the proposed successive framework can promote the success rate of odor source searching to around 95%, while the success rate of the model trained with only RL was around 85%.

The rest of the paper is organized as follows: Section 2 presents the structure of the MCOTSK model, how the successive training framework is utilized to tune the system, and the application of the proposed method in odor source searching. Section 3 compares the controller trained with the proposed method and the controller trained with only supervised training or reinforcement training and analyzes the results. Section 4 presents some discussions. Section 5 concludes the paper.

2. Methods

In this section, the proposed supervised-reinforced successive training framework for an MCOTSK is introduced. The MCOTSK serves as an “Actor” mapping the state s to the action a of the robot. The state means the observed state of the environment, which is measured by the sensing system of the robot. The action means estimated control commands for the robot. As illustrated in Figure 1, the proposed training framework consists of two parts: in the supervised training part, the MCOTSK is trained offline with numerous state-action pairs collected from robot-environment interactions when the robot is driven by a pre-designed controller; in the reinforced training part, the MCOTSK is trained online by maximizing the expected future cumulative reward when the robot's action is estimated by the MCOTSK Actor. The structure of the MCOTSK model and two successive training parts are introduced in the following subsections.

FIGURE 1
www.frontiersin.org

Figure 1. Illustration of the supervised-reinforced successive training framework for the multi-continuous-output TSK fuzzy inference system (MCOTSK).

2.1. The Structure of the MCOTSK Model

The MCOTSK model is a variation of the general TSK fuzzy inference system (Chen X. et al., 2022). As depicted in Figure 1, the MCOTSK model consists of five layers, in which the adjustable nodes are represented by rectangles, and the fixed nodes are represented by circles. An, m(n = 1, …, N; m = 1, …, M) are fuzzy sets.

Assuming the MCOTSK model has M inputs: x1, …, xMR, the inputs are fuzzified by N fuzzy rules in the first layer, which is called the fuzzification layer. The outputs of this layer are formulated as follows:

θn,m1=μAn,m(xm)=e-(xm-cn,m)22an,m2,    (1)

where μAn, m represents the membership function of the fuzzy set An, m(n = 1, …, N; m = 1, …, M) and is set to be a Gaussian membership function (MF) in this paper. ar, m and cr, m are hyper-parameters adjusting the distribution of the Gaussian MFs.

The section layer is a fixed layer, in which all the nodes are marked as π. The outputs are the firing level of the rules, and are formulated as follows:

θn2=m=1MμAn,m(xm) (n=1,,N).    (2)

The third layer is the normalization layer. It normalizes the outputs of the second layer to represent the contribution of the nth fuzzy rule to the sum of the firing level of all rules. The output of this layer can be expressed as follows:

θn3=θn2k=1Nθk2.    (3)

The fourth layer is an adaptive layer, of which the output is the product of the normalized firing level calculated by the third layer and a linear polynomial of the inputs of the MCOTSK model:

θn4=θn3yn(x1,,xM)=θn3(bn,0+m=1Mbn,mxm),    (4)

where yn is the linear polynomial of Rule n, and bn, 0 and bn, m are adjustable weight parameters.

The last layer is the output layer. It calculates the weighted sum of θn4. Assuming the MCOTSK model has P outputs, they can be expressed as follows:

θp5=n=1Nωp,nθn4,    (5)

where ωp, n are adaptive weight parameters (p = 1, …, P; n = 1, …N).

In order to make the MCOTSK model estimate optimal actions from the input states of the environment, the adaptive parameters an, m, bn, m, bn, 0, cn, m, and ωp, n (n = 1, …, N; m = 1, …, M; p = 1, …, P) need to be tuned.

2.2. The Supervised Training Part

In the supervised training part, the proposed MCOTSK model learns from an existing suboptimal Actor, which was designed with prior knowledge. By running a robotic task with the suboptimal Actor for multiple trails, the state of the environment and the action the robot takes can be recorded. Numerous collected state-action pairs are utilized as input-output samples to train the MCOTSK model.

The centers of the MFs cn, m are initialized using a conventional K-means clustering method, which is the same as Cui et al. (2020) and Wang et al. (2020). The SDs of the MFs an, m are initialized to be 1.

At each training epoch, a batch of state-action pairs (ai, si), (i = 1, …, BSs) (batch size BSs = 32 in this paper) are randomly selected from all the collected samples to tune the parameters of MCOTSK by minimizing the mean squared error between the estimated actions and the collected actions:

ϕ*=argminϕi=1BSs[MCOTSK(si|ϕ)-ai]2,    (6)

where ϕ*={an,m,bn,m,bn,0,cn,m,ωp,n}* is the optimal parameter set for the supervised-trained MCOTSK model.

The training process will terminate when the recorded minimum mean squared error on the evaluation set keeps unchanged for 40 training epochs. The optimal MCOTSK model is further used as the initial model in the reinforced training part.

2.3. The Reinforced Training Part

In the reinforced training part, the DDPG RL algorithm (Lillicrap et al., 2015) is applied to further train the MCOTSK model through the “trial-and-error” process.

A “Target actor” is initialized the same as the MCOTSK Actor optimized in the supervised training part. A “Critic” model and its twin “Target critic” model are two artificial neural networks initialized with the same structure and parameters and serve as the action-value functions q(s, a) and qtarg(s, a), which calculated the expected cumulative future reward of the current state-action pair.

At each step t during the robot's task, an action command at is estimated from the input state st with the MCOTSK model, and the robot takes the corresponding action. Then an updated state st + 1 of the environment is perceived by the robot and serves as the input of MCOTSK at the next step. The experience of the robot (st, at, rt, st + 1) is stored in an experience replay buffer D (buffer size = 5,000 in this paper). A batch of stored experience in D was randomly selected to tune the Actor and Critic model in each training epoch (batch size BSr = 32 in this paper).

In a reinforced training epoch, st + 1 is sent to the Target actor to estimate an action command atarg, t + 1 for the next state. The reward rt the robot obtains at step t and the action value qtarg calculated with the Target critic were used to calculate the target action value r + γqtarg(st + 1, atarg, t + 1). The Temporal-Difference error between the action value q(st, at) estimated by the Critic model and the target action value estimated by the Target critic model are used to optimize the Critic model by minimizing the following loss with stochastic gradient descent:

L(ϕc,D)=E(st,at,rt,st+1)~D[(q(st,at|ϕc)(r+γqtarg(st+1,atarg,t+1|ϕc,targ)))2],    (7)

where ϕc is the parameters of the Critic model, and ϕc, targ is the parameters of the Target critic model. The MCOTSK Actor is tuned by maximizing the estimated action value from the Critic model. Therefore, the loss function for gradient descent is set as follows:

  L(ϕa,D)=-E(st,at,rt,st+1)~D[q(st,at|ϕc)]=-E(st,at,rt,st+1)~D[q(st,MCOTSK(st|ϕa)|ϕc)],    (8)

where ϕa is the parameters of the Actor model, and ϕa, targ is the parameters of the Target actor model.

The parameters ϕa, targ and ϕc, targ are updated through a soft updating policy at each training epoch:

ϕc,targρϕc,targ+(1-ρ)ϕc,    (9)
ϕa,targρϕa,targ+(1-ρ)ϕa,    (10)

where ρ is 0.9 in this paper.

In order to reduce overfitting and increase generalization in training the MCOFIS, the DropRule technique (Wu et al., 2019) is applied in the training process. DropRule randomly drops some fuzzy rules (sets the firing level to zero) during the training process with probability P ∈ (0, 1) and remains the firing level unchanged with probability 1 − P. DropRule can promote the robustness of each individual rule. The Layer Normalization (LN) technique is used to normalize the firing level of the rules. The LN layer added in the MCOTSK model is expected to mitigate the gradient vanishing issues (Cui, 2022).

2.4. Application of the Training Framework in Odor Source Searching

In this paper, the proposed successive training framework is applied to an odor source searching problem to demonstrate its feasibility and superiority.

The odor source searching problem in this paper is defined as follows: in an outdoor environment in which the wind field is changing over time, the robot starts from a position away from the odor source and tracks dynamic odor plumes and reaches within 2m from the odor releasing source. The searching area is set to be 40m × 10m, and the coordinate system is shown in Figure 2. The odor leakage source can be regarded as a point and is located at (5, 0). The wind velocity is set as 1m/s in the searching space. The wind direction is aligned to X-axis at t = 0. The noise gain on the wind direction is 5. The odor plumes (illustrated as the red puffs in Figure 2) are released from the odor source and dispersed by the wind. The plumes are modeled by the filament-based odor plume dispersion model (Farrell et al., 2002) to simulate an intermittent 2D odor concentration distribution.

FIGURE 2
www.frontiersin.org

Figure 2. Illustration of a snapshot of the searching area in the odor source searching problem. The odor source is represented by the star. The yellow patches represent the area within 2 m from the odor source. The wind field is illustrated by the black arrows. The red puffs are the simulated odor plumes, which resemble the real-world plumes well as shown in Figure 1.

The robot runs a Lévy Taxis-based odor plume tracking algorithm, which is a variation of Fuzzy Lévy Taxis (Chen and Huang, 2020b), integrating the proposed MCOTSK Actor model. At each searching step, the robot turns its heading θa to an angle Ta and moves forward for a length Ml. Ta and Ml follow the probability distribution presented in Equations (11), (12), and (13):

Ta=[2·arctan(1-α1+αtan(π(rnd-0.5)))]+bias,    (11)
Ml=Lmin·rnd11-μ,    (12)
bias=βθu+(1-β)θa.    (13)

rnd is a random value uniformly distributed in [0, 1] and is resampled in each searching step. The key parameters α, β, and μ of the Fuzzy Lévy Taxis algorithm are determined by the proposed MCOTSK model. The inputs of MCOTSK are the states of the environment: the odor concentration Ct measured by the robot at its current position at time t and the concentration gradient ∇Ct = CtCt − 1. The outputs of MCOTSK go through a Tanh activation layer limiting the outputs in the range of [−1,1]. After a further rescale process, the range of the outputs can be adjusted suitable for the key parameters α, β, and μ. The rescaled outputs are the estimated action commands and are utilized to drive the robot.

To apply the proposed successive training framework, 50 trails of odor source searching tasks are conducted, during which the robot runs the Fuzzy Lévy Taxis algorithm, and the state-action pair {Ct, ∇Ct, α, β, μ} is recorded at each time step t. A total number of 1,816 state-action pairs are collected in this study to train the MCOTSK model firstly with supervised learning. The learning rate of the supervised learning part is 0.01. The number of rules is set to be 10. The DropRule rate is 0.2. The mean squared errors on the evaluation set between the collected actions and the outputs of MCOTSK at each epoch are recorded during the supervised training and are shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. The mean squared errors on the evaluation set during the supervised training stage.

The trained model is used as an initial model in the reinforced training stage. Every odor source searching task is a training episode. An episode will stop when the robots arrive within 2m from the odor source, exceeds the boundaries of the searching area, or the number of searching steps exceeds a limit, which is 60 steps in this paper. The learning rate of Actor is 0.0001 and that of Critic is 0.002. The reward of the robot obtained in step t is as follows:

rt={20if the robot arrives within 2 m  from the odor source,10the robot exceeds the boundaries  of the searching area,1+CcC0cos(θuθa)otherwise.    (14)

where C0 is a constant and set to 30 in this paper. This reward setting is designed to let the robot learn bio-inspired anemotaxis and chemotaxis behaviors. The models were trained for 360 episodes. During the process of training, we recorded the reward the robot obtained in each episode. Figure 4 presents the average reward for every 20 episodes during the reinforced training. It can be seen that the average reward started from around −22 because a large variation was added to the estimated action for exploration. With the added variation decayed, the average reward increased and converged to around 10. From the average reward curves, we can know that the robot can learn to track the dynamic plumes and find the odor source with the MCOTSK model trained by the proposed method.

FIGURE 4
www.frontiersin.org

Figure 4. The average reward in each episode in the reinforced training stage.

3. Performance Evaluation

In order to demonstrate the advantages of the proposed training framework, Monte Carlo tests were conducted in a testing environment that is different from the training environment. The robots started from random positions in the rectangle area shown in Figure 2 and searched the odor source with 21 different action models—1: the Fuzzy Lévy Taxis algorithm used in the supervised training; 2 ~ 20: the trained MCOTSK model after every 20 reinforced training episodes (from 0 to 360 episodes); 21: the MCOTSK model trained with RL only. For each model, 200 trials were conducted.

The controllers were evaluated with three metrics. The first metric was the success rate: the proportion of trials in which the robot reached <2m from the odor source. The second metric was the number of searching steps in all successful trials. The third metric was the distance overhead, which is the traveled distance from the starting position to the stopping position divided by the straight distance in all successful trails. The latter two metrics reflect the efficiency of the searching process. The results of the Monte Carlo tests were shown in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. Performance evaluation results of the three controllers in the Monte Carlo tests. (A) Success rate, (B) is the number of searching steps, and (C) is the distance overhead. Model 1: the Fuzzy Lévy Taxis algorithm used in the supervised training; Models 2–20: the trained MCOTSK model after every 20 reinforced training episodes (from 0 episodes to 360 episodes); Model 21: the MCOTSK model trained with RL only.

It can be seen that the Fuzzy Lévy Taxis algorithm and the model trained with only the supervised stage can achieve a similar success rate (around 85%) and efficiency. It demonstrated that the MCOTSK has been trained to a suboptimal action model. When the reinforced training stage started, the success rate first decreased and then increased fast and exceeded 95%. The decrease at the early reinforce training stage is because the Critic model was being tuned. Once the Critic model can estimate the action value accurately, the performance of the MCOTSK-based Actor returned to the desired condition. With the proposed framework, the robot can learn some bio-inspired searching behaviors in the supervised training stage and fine-tune the parameters of the Actor model in the reinforcement training stage, which can avoid too much random parameter exploration and accelerate the reinforced training process. The performance of the trained model can also benefit from the pre-designed controller because it can provide correct guidance for the robot at the early stage during RL and serve as a baseline behavior pattern. Therefore, the model trained by the proposed framework can achieve better results compared with the model trained by RL only. Compared with the MCOTSK model trained with RL only (the last model in Figure 5), the success rate of the model trained with the successive framework was 10% higher, and the median searching steps and distance overhead were similar. This result can demonstrate that the proposed training framework can initialize the action model to a suboptimal parameter setting, and a more robust model can be obtained through further RL training compared with the model trained by RL only.

A typical odor source searching trajectory generated by the MCOTSK model trained with the proposed framework was shown in Figure 6. It can be seen that when the robot was in the plumes, it went through an upwind surge path, which is a typical anemotaxis behavior learned in the reinforced training process. When it missed the plumes, it conducted a random walk, which is a behavior inherited from the Fuzzy Lévy Taxis algorithm.

FIGURE 6
www.frontiersin.org

Figure 6. A typical odor source searching trajectory with the MCOTSK model trained with the proposed successive training framework. The blue curves represent the trajectories.

4. Discussion

4.1. Limitations of the Proposed Framework

Intuitively, the proposed framework can be more time-consuming compared with tuning the controller by supervised training only. In a scenario where edge cases can be ignored and the manually-designed controllers can perform well enough to achieve the goal, the successive training framework may be redundant. Compared with training by RL only, the proposed framework requires a pre-designed controller or some prior knowledge for supervised training, which can be hard in some complex scenarios where existing controllers are not available.

4.2. Application Potentials

In this paper, the proposed framework was applied to search for a single odor leakage source. When applied in a scenario where there are multiple odor sources, the proposed training framework can be integrated with various multi-robot odor source searching algorithms (Feng et al., 2019; Wiedemann et al., 2019), that is, to train the robots with supervised learning using the state-action data collected from the existing multi-robot searching algorithm and then to further tune the Actor with RL to learn an optimal action policy.

It is also promising to apply the proposed framework to other robotic problems, e.g., controlling surgical robots (Zhou et al., 2020), industrial manipulators (Su et al., 2020, 2022), and robotic grasping (Deng et al., 2021). The controllers are first initialized with a manually-designed suboptimal controller, and then trained by RL to achieve better performance. Human-robot interactions can also benefit from the proposed framework. The monitored physiological signals (Qi and Su, 2022) and motion signals (Chen et al., 2021b) can serve as the input of the Actor model. The Actor model may be initialized by a generic parameter setting in the supervised training stage. After being trained by RL on each individual user, the robot is expected to cooperate with the user better.

5. Conclusion

In this paper, a supervised-reinforced successive training framework for a fuzzy inference system was proposed and applied to a robotic odor source searching problem. The performance evaluation results showed that the proposed method can train the FIS to a suboptimal model through supervised training, and the model trained with further RL can perform better than the model trained with RL only. The results of this paper can inspire researchers to initialize the fuzzy actor model through supervised training using some prior knowledge and then tune a better model with RL.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The name of the repository and accession number can be found at: GitHub, https://github.com/cxxacxx/MCOTSK.

Author Contributions

XC contributed to the conception and implementation of the study. YL and CF contributed to supervising the study, reviewing, and revising the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the National Key R&D Program of China [Grant 2018YFC2001601]; the National Natural Science Foundation of China [Grants U1913205, 62103180, 52175272, and 51805237]; the China Postdoctoral Science Foundation (2021M701577); Guangdong Basic and Applied Basic Research Foundation [Grant 2020B1515120098]; Guangdong Innovative and Entrepreneurial Research Team Program [Grant 2016ZT06G587]; the Science, Technology, and Innovation Commission of Shenzhen Municipality [Grants SGLH20180619172011638, ZDSYS20200811143601004, and KQTD20190929172505711]; the Stable Support Plan Program of Shenzhen Natural Science Fund [Grant 20200925174640002]; Joint Fund of Science & Technology Department of Liaoning Province and State Key Laboratory of Robotics, China [Grant No. 2020-KF-22-03]; Centers for Mechanical Engineering Research and Education at MIT and SUSTech.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Chen, C., Huang, J., Wu, D., and Tu, X. (2022). Interval type-2 fuzzy disturbance observer-based t-s fuzzy control for a pneumatic flexible joint. IEEE Trans. Indus. Electron. 69, 5962–5972. doi: 10.1109/TIE.2021.3090708

CrossRef Full Text | Google Scholar

Chen, X., Fu, C., and Huang, J. (2021a). A Deep Q-Network for robotic odor/gas source localization: modeling, measurement and comparative study. Measurement 183, 109725. doi: 10.1016/j.measurement.2021.109725

CrossRef Full Text | Google Scholar

Chen, X., and Huang, J. (2019). Odor source localization algorithms on mobile robots: A review and future outlook. Robot. Auton. Syst. 112, 123–136. doi: 10.1016/j.robot.2018.11.014

CrossRef Full Text | Google Scholar

Chen, X., and Huang, J. (2020a). Combining particle filter algorithm with bio-inspired anemotaxis behavior: a smoke plume tracking method and its robotic experiment validation. Measurement 154, 107482. doi: 10.1016/j.measurement.2020.107482

CrossRef Full Text | Google Scholar

Chen, X., and Huang, J. (2020b). “Towards environmentally adaptive odor source localization: fuzzy Lévy Taxis algorithm and its validation in dynamic odor plumes,” in 2020 5th International Conference on Advanced Robotics and Mechatronics (ICARM) (Shenzhen), 282–287. doi: 10.1109/ICARM49381.2020.9195363

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, X., Marjovi, A., Huang, J., and Martinoli, A. (2020). Particle source localization with a low-cost robotic sensor system: algorithmic design and performance evaluation. IEEE Sensors J. 20, 13074–13085. doi: 10.1109/JSEN.2020.3002273

CrossRef Full Text | Google Scholar

Chen, X., Yang, B., Huang, J., Leng, Y., and Fu, C. (2022). “A reinforcement learning fuzzy system for continuous control in robotic odor plume tracking,” in 2022 7th International Conference on Advanced Robotics and Mechatronics (ICARM) (Guilin).

Google Scholar

Chen, X., Zhang, K., Liu, H., Leng, Y., and Fu, C. (2021b). A probability distribution model-based approach for foot placement prediction in the early swing phase with a wearable IMU sensor. IEEE Trans. Neural Syst. Rehabil. Eng. 29, 2595–2604. doi: 10.1109/TNSRE.2021.3133656

PubMed Abstract | CrossRef Full Text | Google Scholar

Cui, Y. (2022). PyTSK. Available online at: https://github.com/YuqiCui/PyTSK

Google Scholar

Cui, Y., Wu, D., and Huang, J. (2020). Optimize tsk fuzzy systems for classification problems: minibatch gradient descent with uniform regularization and batch normalization. IEEE Trans. Fuzzy Syst. 28, 3065–3075. doi: 10.1109/TFUZZ.2020.2967282

CrossRef Full Text | Google Scholar

Dai, X., Li, C.-K., and Rad, A. B. (2005). An approach to tune fuzzy controllers based on reinforcement learning for autonomous vehicle control. IEEE Trans. Intell. Transp. Syst. 6, 285–293. doi: 10.1109/TITS.2005.853698

CrossRef Full Text | Google Scholar

Deng, Z., Fang, B., He, B., and Zhang, J. (2021). An adaptive planning framework for dexterous robotic grasping with grasp type detection. Robot. Auton. Syst. 140, 103727. doi: 10.1016/j.robot.2021.103727

CrossRef Full Text | Google Scholar

Er, M. J., and Deng, C. (2004). Online tuning of fuzzy inference systems using dynamic fuzzy q-learning. IEEE Trans. Syst. Man Cybern. Part B 34, 1478–1489. doi: 10.1109/TSMCB.2004.825938

PubMed Abstract | CrossRef Full Text | Google Scholar

Farrell, J. A., Murlis, J., Long, X., Li, W., and Cardé, R. T. (2002). Filament-based atmospheric dispersion model to achieve short time-scale structure of odor plumes. Environ. Fluid Mech. 2, 143–169. doi: 10.21236/ADA399832

CrossRef Full Text | Google Scholar

Feng, Q., Zhang, C., Lu, J., Cai, H., Chen, Z., Yang, Y., et al. (2019). Source localization in dynamic indoor environments with natural ventilation: an experimental study of a particle swarm optimization-based multi-robot olfaction method. Build. Environ. 161, 106228. doi: 10.1016/j.buildenv.2019.106228

CrossRef Full Text | Google Scholar

Kumar, N., Rahman, S. S., and Dhakad, N. (2020). Fuzzy inference enabled deep reinforcement learning-based traffic light control for intelligent transportation system. IEEE Trans. Intell. Transp. Syst. 22, 4919–4928. doi: 10.1109/TITS.2020.2984033

CrossRef Full Text | Google Scholar

Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., et al. (2015). Continuous control with deep reinforcement learning. arXiv[Preprint].arXiv:1509.02971. doi: 10.48550/arXiv.1509.02971

PubMed Abstract | CrossRef Full Text | Google Scholar

Nguyen, A.-T., Taniguchi, T., Eciolaza, L., Campos, V., Palhares, R., and Sugeno, M. (2019). Fuzzy control systems: past, present and future. IEEE Comput. Intell. Mag. 14, 56–68. doi: 10.1109/MCI.2018.2881644

PubMed Abstract | CrossRef Full Text | Google Scholar

Qi, W., and Su, H. (2022). A cybertwin based multimodal network for ECG patterns monitoring using deep learning. IEEE Trans. Indus. Informatics. 1–1. doi: 10.1109/TII.2022.3159583

CrossRef Full Text | Google Scholar

Shigaki, S., Shiota, Y., Kurabayashi, D., and Kanzaki, R. (2019). Modeling of the adaptive chemical plume tracing algorithm of an insect using fuzzy inference. IEEE Trans. Fuzzy Syst. 28, 72–84. doi: 10.1109/TFUZZ.2019.2915187

CrossRef Full Text | Google Scholar

Su, H., Hu, Y., Karimi, H. R., Knoll, A., Ferrigno, G., and De Momi, E. (2020). Improved recurrent neural network-based manipulator control with remote center of motion constraints: experimental results. Neural Netw. 131, 291–299. doi: 10.1016/j.neunet.2020.07.033

PubMed Abstract | CrossRef Full Text | Google Scholar

Su, H., Qi, W., Chen, J., and Zhang, D. (2022). Fuzzy approximation-based task-space control of robot manipulators with remote center of motion constraint. IEEE Trans. Fuzzy Syst. 1–1. doi: 10.1109/TFUZZ.2022.3157075

CrossRef Full Text | Google Scholar

Vergassola, M., Villermaux, E., and Shraiman, B. I. (2007). ‘infotaxis' as a strategy for searching without gradients. Nature 445, 406–409. doi: 10.1038/nature05464

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, L., Huang, J., Wu, D., Duan, T., Zong, R., and Jiang, S. (2020). “Hand gesture recognition based on multi-classification adaptive neuro-fuzzy inference system and PMMG,” in 2020 5th International Conference on Advanced Robotics and Mechatronics (Shenzhen), 460–465. doi: 10.1109/ICARM49381.2020.9195286

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, L., and Pang, S. (2020). “An implementation of the adaptive neuro-fuzzy inference system (ANFIS) for odor source localization,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (Las Vegas, NV: IEEE), 4551–4558. doi: 10.1109/IROS45743.2020.9341688

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, L., Pang, S., and Li, J. (2021). Olfactory-based navigation via model-based reinforcement learning and fuzzy inference methods. IEEE Trans. Fuzzy Syst. 29, 3014–3027. doi: 10.1109/TFUZZ.2020.3011741

CrossRef Full Text | Google Scholar

Wang, L.-X., and Mendel, J. M. (1992). “Back-propagation fuzzy system as nonlinear dynamic system identifiers,” in IEEE International Conference on Fuzzy Systems (San Diego, CA: IEEE), 1409–1418. doi: 10.1109/FUZZY.1992.258711

CrossRef Full Text | Google Scholar

Wiedemann, T., Shutin, D., and Lilienthal, A. J. (2019). Model-based gas source localization strategy for a cooperative multi-robot system-a probabilistic approach and experimental validation incorporating physical knowledge and model uncertainties. Robot. Auton. Syst. 118, 66–79. doi: 10.1016/j.robot.2019.03.014

CrossRef Full Text | Google Scholar

Wu, D., and Tan, W. W. (2006). Genetic learning and performance evaluation of interval type-2 fuzzy logic controllers. Eng. Appl. Artif. Intell. 19, 829–841. doi: 10.1016/j.engappai.2005.12.011

CrossRef Full Text | Google Scholar

Wu, D., Yuan, Y., Huang, J., and Tan, Y. (2019). Optimize TSK fuzzy systems for regression problems: minibatch gradient descent with regularization, droprule, and adabound (MBGD-RDA). IEEE Trans. Fuzzy Syst. 28, 1003–1015. doi: 10.1109/TFUZZ.2019.2958559

CrossRef Full Text | Google Scholar

Zhou, M., Yu, Q., Huang, K., Mahov, S., Eslami, A., Maier, M., et al. (2020). Towards robotic-assisted subretinal injection: a hybrid parallel-serial robot system design and preliminary evaluation. IEEE Trans. Indus. Electron. 67, 6617–6628. doi: 10.1109/TIE.2019.2937041

CrossRef Full Text | Google Scholar

Keywords: supervised learner, reinforcement learning, fuzzy inference system, robotic odor source searching, Monte Carlo test

Citation: Chen X, Leng Y and Fu C (2022) A Supervised-Reinforced Successive Training Framework for a Fuzzy Inference System and Its Application in Robotic Odor Source Searching. Front. Neurorobot. 16:914706. doi: 10.3389/fnbot.2022.914706

Received: 07 April 2022; Accepted: 27 April 2022;
Published: 31 May 2022.

Edited by:

Hang Su, Fondazione Politecnico di Milano, Italy

Reviewed by:

Bin Fang, Tsinghua University, China
Longbin Zhang, Royal Institute of Technology, Sweden
Mingchuan Zhou, Zhejiang University, China

Copyright © 2022 Chen, Leng and Fu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Chenglong Fu, ZnVjbCYjeDAwMDQwO3N1c3RlY2guZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.