- 1Department of Mechanical Engineering, National Cheng Kung University, Tainan, Taiwan
- 2Department of Biomedical Engineering, National Yang Ming University, Taipei, Taiwan
- 3The Ph.D. Program for Neural Regenerative Medicine, College of Medical Science and Technology, Taipei Medical University, Taipei, Taiwan
- 4Key Laboratory of Medical Neurobiology of Zhejiang Province, Department of Neurology of the Second Affiliated Hospital, Interdisciplinary Institute of Neuroscience and Technology, Zhejiang University School of Medicine, Hangzhou, China
- 5Key Laboratory of Biomedical Engineering of Ministry of Education, College of Biomedical Engineering and Instrument Science, Qiushi Academy for Advanced Studies, Zhejiang University, Hangzhou, China
- 6Research Center for Information Technology Innovation, Academia Sinica, Taipei, Taiwan
- 7Department of Regulatory & Quality Sciences, University of Southern California, Los Angeles, CA, United States
- 8Buddhist Tzu Chi Medical Foundation, Department of Neurology, Hualien Tzu Chi Hospital, Hualien, Taiwan
- 9Department of Neurology, School of Medicine, Tzu Chi University, Hualien, Taiwan
- 10Department of Neurosurgery, Neurological Institute, Taipei Veterans General Hospital, Taipei, Taiwan
- 11Department of Neurological Surgery, University of Washington, Seattle, WA, United States
Objective: In brain machine interfaces (BMIs), the functional mapping between neural activities and kinematic parameters varied over time owing to changes in neural recording conditions. The variability in neural recording conditions might result in unstable long-term decoding performance. Relevant studies trained decoders with several days of training data to make them inherently robust to changes in neural recording conditions. However, these decoders might not be robust to changes in neural recording conditions when only a few days of training data are available. In time-series prediction and feedback control system, an error feedback was commonly adopted to reduce the effects of model uncertainty. This motivated us to introduce an error feedback to a neural decoder for dealing with the variability in neural recording conditions.
Approach: We proposed an evolutionary constructive and pruning neural network with error feedback (ECPNN-EF) as a neural decoder. The ECPNN-EF with partially connected topology decoded the instantaneous firing rates of each sorted unit into forelimb movement of a rat. Furthermore, an error feedback was adopted as an additional input to provide kinematic information and thus compensate for changes in functional mapping. The proposed neural decoder was trained on data collected from a water reward-related lever-pressing task for a rat. The first 2 days of data were used to train the decoder, and the subsequent 10 days of data were used to test the decoder.
Main Results: The ECPNN-EF under different settings was evaluated to better understand the impact of the error feedback and partially connected topology. The experimental results demonstrated that the ECPNN-EF achieved significantly higher daily decoding performance with smaller daily variability when using the error feedback and partially connected topology.
Significance: These results suggested that the ECPNN-EF with partially connected topology could cope with both within- and across-day changes in neural recording conditions. The error feedback in the ECPNN-EF compensated for decreases in decoding performance when neural recording conditions changed. This mechanism made the ECPNN-EF robust against changes in functional mappings and thus improved the long-term decoding stability when only a few days of training data were available.
Introduction
Brain machine interface (BMI) technology converts the brain's neural activity into kinematic parameters of limb movements. This allows controlling a computer cursor or prosthetic devices (Kao et al., 2014; Slutzky, 2018), which can greatly improve the quality of life. Intracortical BMIs have used microelectrodes implanted in the cortex to decode neural signals. These signals have then been converted into motor commands to control an anthropomorphic prosthetic limb, thereby restoring natural function (Collinger et al., 2013; Roelfsema et al., 2018).
The decoder was the most crucial component of a BMI; it modeled the functional mapping between neural activities and kinematic parameters (e.g., movement velocity or position), and assumed that this functional mapping was time-invariant (i.e., a stationary statistical assumption) (Kim et al., 2006). However, under real neural recording conditions, there existed a high degree of within- and across-day variability (Simeral et al., 2011; Perge et al., 2013, 2014; Wodlinger et al., 2014; Downey et al., 2018) that prevented satisfaction of the stationary statistical assumption. This variability consisted of the relative position of the recording electrodes—and surrounding neurons, electrode properties, tissue reaction to electrodes, and neural plasticity—and might affect the functional mapping between neural activities and kinematic parameters (Jackson et al., 2006; Barrese et al., 2013; Fernández et al., 2014; Salatino et al., 2017; Michelson et al., 2018; Hong and Lieber, 2019). The variability in neural recording conditions resulted in unstable long-term decoding performance and led to frequent decoder retraining (Jarosiewicz et al., 2013, 2015).
Conventional decoder retraining required the subject to periodically perform a well-defined task to collect new training data for preventing model staleness (Jarosiewicz et al., 2015). This manner may lead to additional training time before the BMI can be used. Traditional linear neural decoders did not need frequent retraining but possessed limited computational complexity to deal with neural recording condition changes owning to linear properties (Collinger et al., 2013). It is known that the newly encountered neural recording conditions in chronic BMI systems have some commonality with past neural recording conditions (Chestek et al., 2011; Perge et al., 2013; Bishop et al., 2014; Nuyujukian et al., 2014; Orsborn et al., 2014). Therefore, computationally powerful non-linear decoders were proposed to learn a diverse set of neural-to-kinematic mappings corresponding to various neural recording conditions collected over many days before BMI use (Sussillo et al., 2016). This approach avoided BMI interruption by keeping model parameters fixed during BMI use and made BMI inherently robust to changes in neural recording conditions by exploiting the similarities between newly encountered and past neural recording conditions. Therefore, the BMIs were trained with several days of data in order to learn various neural recording conditions and achieve stable long-term decoding. However, they heavily relied on the huge training data where a large training set may not be available for both non-human primates and rodent models.
The limited training data have become an issue for BMI application in long-term performance. A chronic inflammatory reaction results in neural signal loss and decrease in quality over time (Chen et al., 2009). Also, the number of implanted electrodes is limited by the size of the neural nuclei in the rodent brain. Therefore, limited neurons and limited recording times lead to limited training data.
Rodent models with small numbers of implanted electrodes have been widely used to investigate state-of-the-art neural prostheses. Previous studies have demonstrated the decoding performances of various methods at the motor cortex (Zhou et al., 2010; Yang et al., 2016), somatosensory cortex (Pais-Vieira et al., 2013), and hippocampus (Tampuu et al., 2019) in rodent models. The results indicated that good decoding methods should be considerably more robust to small sample sizes caused by limited neurons or limited recording times. In general, a limited amount of training data made traditional decoding methods inaccurate, because they usually required a large number of neurons to achieve desirable levels of performance. Furthermore, small amounts of data have made modern decoding methods unreliable, because their increasing model complexities required a large amount of training data (Glaser et al., 2017). Whether the BMIs could deal with the scenario in which only a few days of training data were available is unknown. This motivated us to develop a neural decoder that could learn from limited training data based on rodent models.
In time-series prediction applications, neural networks (NNs) usually employed prediction error as an additional input of the networks. This has been proven to yield superior performance compared with that without error feedback (Connor et al., 1994; Mahmud and Meesad, 2016; Waheeb et al., 2016). The error feedback determined the difference between the network output and the target value. This information could provide the network with information concerning previous prediction performance and might thereby guide the network to accurate prediction. In a feedback control system, the output signal was fed back to form an error signal, which was the difference between the target and actual output, in order to drive the system. Using feedback could reduce the effects of model uncertainty (Løvaas et al., 2008). Furthermore, feedback control could cope with trial-to-trial variability caused by complex dynamics or noise in motor behavior (Todorov and Jordan, 2002). Based on the contemporary physiological studies in the human cortex (Miyamoto et al., 1988), a feedback motor command has been used as an error signal for training an NN (Kawato, 1990). One study hypothesized that the user intended to directly move toward the target when using BMI. This study fitted the neural decoder by estimating user's intended velocity which was determined from target position, cursor position, and decoded velocity (Gilja et al., 2012). A recent study took into account how the user modified the neural modulation to deal with the movement errors caused by neural variability in the feedback loop (Willett et al., 2019). Their framework simulated online/closed-loop dynamics of an intracortical BMI and calibrated its decoder by an encoded control signal, which was the difference between target position and cursor position. The encoded control signal using target position was first transformed into neural features which were then mapped to a decoded control vector for updating decoder output, i.e., cursor velocity. This motivated us to introduce an error feedback into a neural decoder for dealing with the variability in neural recording conditions because the error feedback might compensate for the changes in neural recording conditions. Then, the neural decoder did not need retraining and was expected to be robust to various neural recording conditions when only using a few days of training data.
Several characteristics make NNs computationally powerful decoders in BMIs. First, an NN with a sufficient number of hidden neurons can approximate any continuous function (Hornik et al., 1989). This makes an NN well-suited to learn the functional mapping between neural activity and kinematic parameters. Second, several types of NNs can successfully control motor movement in BMIs. These include recurrent NN (RNN) (Haykin, 1994; Shah et al., 2019), echo-state network (ESN) (Jaeger and Haas, 2004), and time-delay NN (TDNN) (Waibel et al., 1989). RNNs have feedback connections that are capable of processing neural signal sequences. Their feedback loop is applicable to system dynamics modeling and time-dependent functional mapping between neural activity and kinematic parameters (Haykin, 1994). ESN was developed as an RNN that only trains connections between the hidden neurons and the output neurons for a simple learning process (Jaeger and Haas, 2004). TDNNs are feedforward NNs with delayed versions of inputs that implement a short-term memory mechanism (Waibel et al., 1989). Of these NNs, RNNs are highly accurate in BMI applications (Sanchez et al., 2004, 2005; Sussillo et al., 2012; Kifouche et al., 2014; Shah et al., 2019). Therefore, the present work designed an RNN with error feedback as the neural decoder.
Because the performance of an NN relied heavily on its network structure, structure selection is a crucial concern. An NN with an excessively large architecture may overfit the training data and yield poor generalization. Furthermore, it often exhibited rigid timing constraints. By contrast, an NN with an excessively small architecture may underfit the data and fail to approximate the underlying function. The four most frequently used algorithms to determine a network's architecture are constructive, pruning, constructive-pruning, and evolutionary algorithms (EAs). Constructive algorithms (Kwok and Yeung, 1997) began with a simple NN and then increased the number of hidden neurons or connections to that network in each iteration. However, an oversized network may be constructed due to inappropriate stopping criterion. In other words, the matter of when to stop constructing networks lacked consensus. The pruning algorithm (Reed, 1993) began with an oversized NN and then removed insignificant hidden neurons or connections iteratively. However, it was difficult to initially determine an oversized network architecture for a given problem (Kwok and Yeung, 1997). The constructive-pruning algorithm (Islam et al., 2009; Yang and Chen, 2012) combined both a constructive algorithm and pruning algorithm to build an NN. Starting with the simplest possible structures, the NN was first constructed using a constructive algorithm and then removed trivial hidden neurons or connections by using a pruning algorithm to achieve optimal network architecture. Several works have designed NNs using EAs (Huang and Du, 2008; Kaylani et al., 2009; Masutti and de Castro, 2009). EAs were developed as a biologically plausible strategy to adapt various parameters of NNs, such as weights and architectures (Angeline et al., 1994). However, encoding an NN into a chromosome depended on the maximum structure of the network, which is problem-dependent and must be defined by user. This property limited the flexibility of problem representation and the efficiency of EAs. One study (Yang and Chen, 2012) proposed an evolutionary constructive and pruning algorithm (ECPA) without predefining the maximum structure of the network, which made the evolution of the network structure more efficient. Because the NNs used in BMIs were designed in a subject-dependent manner, the automatic optimization of the NN for a specific task is a desired feature. This study adopted the ECPA (Yang and Chen, 2012) to develop an NN with an appropriate structure as a neural decoder for each subject in BMI applications.
This work proposed an evolutionary constructive and pruning neural network with error feedback (ECPNN-EF) to decode neural activity into the forelimb movement of a rat by using only a few days of data to train the neural decoder. A lever-pressing task for the rat was designed to evaluate the effectiveness of the proposed neural decoder. The error feedback providing the difference between the decoded and actual kinematics might compensate for decreases in decoding performance when the neural recording conditions change. Thus, the ECPNN-EF might achieve stable and accurate long-term decoding performance. The rest of this paper is organized as follows. First, we describe the experimental setup and the proposed decoder. Second, we demonstrate the influence of several parameters, namely the probabilities of crossover and mutation. Furthermore, the effects of evolution progress, cluster-based pruning (CBP) and age-based survival selection (ABSS) on the performance of the proposed decoder are also shown and discussed. Finally, we describe how partially connected topology and error feedback improve the long-term decoding performance.
Materials and Methods
Animals
Four male adult Wistar rats were aged 8 weeks old, weighed between 250 and 350 g, and were kept in the animal facility with well-controlled laboratory conditions (12: 12 light/dark cycle with light at 7 AM; 20° ± 3°C) and fed on ad libitum. The care and experimental manipulation of the animals were reviewed and approved by the Institutional Animal Care and Use Committee of the National Yang Ming University.
Surgery for Neural Implantation
Animals were anesthetized with 40 mg/kg Zolazepam and Tiletamine (Zoletil 50, Virbac., Corros, France) and 8 μg/kg dexmedetomidine hydrochloride (Dexdomitor®, Pfizer Inc., New York, NY, USA) through intramuscular injections. Rats were positioned in a stereotaxic frame (Stoelting Co. Ltd., Wood Dale, IL, USA) and secured with the ear bars and tooth bar. An incision was made between the ears. The skin of the scalp was pulled back to expose the surface of the skull from the bregma to the lambdoid suture. Small burr holes were drilled into the skull for the microwire electrode array implanted and for the positioning of screws (Shoukin Industry Co., Ltd., New Taipei City, Taiwan).
For each rat, an 8-channel laboratory-made stainless microwire electrode array (product # M177390, diameter of 0.002 ft., California Fine Wire Co., Grover Beach, CA, USA; the electrodes were spaced 500 μm apart) was vertically implanted into the layer V of the forelimb territory of the primary motor (M1) cortex (anterior-posterior [AP]: +2.0 mm to −1.0 mm, medial-lateral [ML]: +2.7 mm, dorsal-ventral [DV]: 1.5 mm. For determining the location of the forelimb representation of M1 for the electrode implantation, the intracortical microstimulation was applied to confirm via forelimb muscle twitches observed (Yang et al., 2016). Following a 1-week post-surgery recovery period, the animals received the water reward-related lever-pressing training.
Behavioral Training
The rats were trained to press a lever with their right forelimb to obtain a water reward. Before reward training, the rats were single-housed and deprived of water for at least 8 h. During reward training, the rats were placed in a 30 × 30 × 60 cm3 laboratory-designed Plexiglas testing box, and a 14 × 14 × 37.5 cm3 barrier was placed to construct an L-shaped path for the behavioral task. A lever (height of 15 cm from bottom) was set at one end of the path, and an automatic feeder with a water tube that provided water on a plate was set at the other end of the path. The rats could obtain 0.25-ml water drop as a reward on the plate after pressing the lever. Thirsty rats were trained to press a lever in order to receive water reward without any cues because they learned to make an operant response for positive reinforcement (water reward). Rats were trained to press a lever on the left side of the box then freely move along the U-shaped path to the right side of the box. This had to be completed within 3 s to receive a reward. The experimental time course included the behavioral training and data collection phases. In the behavioral training phase, implanted rats learned the lever pressing and water reward association within 3–5 d without neural recordings. To meet criteria for successful learning of the behavioral task, the rats had to complete continuous repetition of five successive trials of associated lever pressing and water reward without missing any trial between successive trials (Lin et al., 2016). Once reaching the criteria, animals entered the data collection phase. During this 12-d phase, forelimb movement trajectories were simultaneously acquired with corresponding electrophysiological recordings of neural spikes as they performed the water reward task.
Data Recording
In this study, forelimb kinematics and neuronal activity were simultaneously recorded while the animal performed the water reward-related lever pressing as shown in Figure 1. During the behavioral task, a blue-colored marker made of nylon was mounted on the right wrist of the rat to track forelimb trajectory. The trajectory of the rat's forelimb movement was captured by a charge-coupled device camera (DFK21F04, Imaging Source, Bremen, Germany) that provided a 640 × 480 RGB image at 30 Hz and then analyzed by a video tracking system (CinePlex, Plexon Inc., Dallas, TX, USA). When the lever was pressed, it triggered the micro-switch of the pull position to generate a transistor–transistor logic pulse to the multichannel acquisition processor (Plexon Inc., Dallas, TX, USA) which allowed the neuronal data to be accurately synchronized to the lever pressing event and then water reward was delivered through a computer-controlled solenoid valve connected to the laboratory-designed pressurized water supply.
Figure 1. System architecture and experimental setting of the water-reward lever-pressing task. While the rat was pressing the lever to obtain the water reward, the neural recording system recorded and preprocessed the neuronal spiking activity from the electrode array implanted in the rat's cortex. The trajectory-tracking system acquired the corresponding forelimb trajectory from the camera.
Neuronal spiking activity of the rat was sampled at 40 kHz and analog filtered from 300 to 5,000 Hz. A spike-sorting algorithm was used to determine single-unit activity. First, an amplitude threshold with four standard deviations of filtered neuronal signals was set to identify spikes from the filtered neuronal signals. Then, spikes were sorted by a trained technician through principal component analysis using a commercial spike-sorting software (Sort Client, Plexon Inc., Dallas, TX, USA).
Neural Decoder: ECPNN-EF
The firing rates of each sorted unit from M1 cortex of rat were decoded into the instantaneous velocity of the forelimb trajectory. Both horizontal and vertical velocities were estimated from the position of the blue-colored marker by a two-point digital differentiation. The firing rates of each sorted neuron was determined by counting spikes in a given time bin whose length was 33 ms and was equal to the temporal resolution of the video tracking system. Figure 2 showed an example of the rat forelimb movement while pressing the lever and corresponding neural spike trains. A time-lag was known to exist between neuronal firing and the associated forelimb state because of their causal relationship (Paninski et al., 2004; Wu et al., 2004; Yang et al., 2016). Furthermore, the decoding accuracy was improved when the optimal time-lag is considered. Here, the water-restricted rats easily learn a lever pressing behavior within a few training sessions, allowing for recording neuronal activity during acquisition of a motor sequence as shown in Figures 2A,B showed some M1 neurons displayed increased activity for sequential motor behavior prior to the lever-pressing event, which presented the maximum firing rate at the third time-bin (with 99 ms lag). Therefore, we empirically choose 363 ms of spike train over 11 time-bins (8 bins before and 2 bins after the 3rd time-bin prior to the lever pressing) to predict a series of movement velocities.
Figure 2. Simultaneous forelimb movement trajectory and spike train recordings during the water-reward-related lever-pressing task. (A) Stop-motion animation representing forelimb movement from the video-tracked time-series data (see Supplemental Video). Six consecutive photographs showed a rat in the test cage successfully reaching and pressing the lever (marked with green) while the forelimb movement trajectories and neural activity were simultaneously recorded. (B) Neuronal activities recorded from eight neurons during one movement displayed as spike trains and the neuronal activity histogram (a bin size of 33 ms). The red line indicates the moment when the rat pressed down the lever with its right forelimb.
The spike train was discretized in 11 time-bins for each trial, corresponding to each entire trajectory of forelimb movements during the lever reaching task. With total k neurons sorted from all channels, we defined as a set of neuronal features, where t denoted time step which was time bin in this study, ni(t) represented as spike count of sorted neuron in current bin, and i denoted index of sorted neuron from 1 to k. In this study, we used both concurrent and preceding bin as neuronal features, which is N(t) and N(tpre), to predict current velocity , where t and tpre represented current time bin and preceding time bin, respectively. The step of data processing was shown in Figure 3.
Figure 3. The data structure of input for the ECPNN-EF decoder. The input of the decoder was consisted of (t), N(tpre) and verror, whose length was 2k + 1, including 2k spike counts of both concurrent and preceding bin, and an error feedback calculated by . To predict a whole forelimb movement in a trial, 11 bins were used and decoded to in the time series.
The ECPNN-EF is an RNN-based neural decoder designed using the ECPA as proposed in Yang and Chen (2012). The input–output function of the ECPNN-EF was denoted by:
where represented the predicted velocity and W represented the weights of the ECPNN-EF. The prediction error verror was adopted as the error feedback which was the absolute value of difference between the actual velocity and predicted velocity, and was calculated by where tpre denoted as preceding time bin.
The structure of ECPNN-EF was showed in Figure 4. Note that there was only one output neuron representing the predicted velocity in the neural decoder. The vertical and horizontal velocities were predicted in two separate neural decoders.
Figure 4. Structure of ECPNN-EF. Network included input, hidden, and output layer. ECPNN-EF took the neuronal feature from combination of concurrent bin N(t), preceding bin N(tpre) and prediction error (error feedback) verror to predict velocity .
Mean squared error was adopted as loss function due to its wide use in regression application and was defined as follows:
The optimal weights of the ECPNN-EF were obtained by minimizing the loss as follows:
This study applied backpropagation through time (BPTT) to find the optimal weights of the ECPNN-EF by iteratively determining the gradient of the loss with respect to the weights as follows:
where η is learning rate. Details of the BPTT (Werbos, 1988) are described in the Supplementary Note 1. The details of designing structure of the ECPNN-EF are described in the Supplementary Note 2.
The ECPNN-EF adopted a hyperbolic tangent sigmoid transfer function for all hidden neurons and the output neuron. Skip connections existed between discontinuous layers, such as from the input layer to the output layer. Furthermore, the hidden layer possessed self-recurrent connections. After both the structure and weights of the ECPNN-EF had been trained, the fixed model was adopted to predict the velocity of the rat's forelimb without the additional cost of training.
The pseudo code of ECPNN-EF training algorithm appears in Algorithm 1. The initial population started with a set of initial NNs, each NN of which had a single hidden neuron. A single connection was generated from one non-error-related input neuron to the hidden neuron. A skip connection was generated from one non-error-related input neuron to the output neuron. A single connection was generated from one error-related input neuron to the hidden neuron or to the output neuron. The detail description of population initialization was in the Supplementary Note 3. Furthermore, a self-recurrent connection was constructed in the hidden neuron with a probability of 0.5. Here, the error-related input neuron received the prediction error, as indicated in Figure 4, whereas non-error-related input neurons received the instantaneous firing rate of each unit. This mechanism of separately generating connections of error-related input neuron and non-error-related input neurons ensures that the initial NNs can immediately process error feedback. As a result, a set of initial NNs with partially connected topology was generated.
The purpose of the network crossover operator was to explore the structural search space and thus improve the processing capabilities of the ECPNN-EF. The network crossover operation randomly selected two parent NNs through tournament selection and then combined their structures to generate an offspring NN with a crossover probability, pc (see Supplementary Note 4). The network mutation operator exploited the structural search space to achieve a small perturbation of structure by randomly generating a new connection from the input to the hidden neuron. Furthermore, the network mutation operation randomly constructed a self-recurrent connection of a hidden neuron or a new skip connection from the output or the hidden neuron to its previous consecutive or non-consecutive layer with a mutation probability, pm (see Supplementary Note 5).
CBP mainly pruned insignificant hidden neurons to avoid an excessively complex ECPNN-EF with poor generalization performance owing to the use of network crossover operation. It first clustered the hidden neurons into two groups (i.e., better and worse groups, depending on their significance in the NN) and then removed the hidden neurons in the worse group in a stochastic manner. The detailed description of CBP was in the Supplementary Note 6.
ABSS prevented the ECPNN-EF from achieving a fully connected structure. It selected NNs for the next generation according to age, which indicated how many generations the NN had survived. Older NNs tended to have a fully connected structure because of the use of network mutation operation. ABSS replaced old age NNs with initial NNs (see Supplementary Note 3) in a stochastic manner and thus prevented the population from achieving a fully connected structure. ABSS mainly removed fully connected networks and made the rest of the NNs survive to the next generation through a stochastic mechanism. The detailed description of ABSS was in the Supplementary Note 7. The evolution process terminated when a generalization loss (GL) met an early stopping criterion or the maximum number of generations was reached. The early stopping criterion motivated from Islam et al. (2009) evaluated the evolution progress using training and validation errors in order to avoid overfitting. It first defined the GL at the τth generation as:
where Eva(τ) is a validation error of the NN with the best fitness at the τth generation and Elow(τ) is the lowest Eva(τ) up to the τth generation. The difference between the average training error and the minimum training error at the τth generation of a strip k was defined as:
where Etr(ω) is the training error of the NN with the best fitness at the ωth generation and k is the strip length. k was set to 5 in this work. Note that GL(τ) and Pk(τ) were determined using the validation and training sets, respectively. Eva(τ) and Etr(ω) were calculated by the loss function provided in (2). The ECPNN-EF training algorithm terminated when GL(τ) > Pk(τ). The optimal NN with a partially connected topology was selected as the neural decoder.
In summary, network crossover and mutation evolved NNs in a constructive manner to improve their processing ability, whereas CBP and ABSS evolved NNs in a destructive way that enhanced their generalization capabilities and reduced hardware costs (Yang and Chen, 2012). An early stopping criterion was adopted to terminate the evolution process by observing both training and validation errors to avoid overfitting during training phase, which reduced the training time and retained generalization capability (Islam et al., 2009). The ECPNN-EF was implemented and trained in MATLAB (MathWorks, Natick, MA, USA).
Data Sets and Optimizing the Structure of Neural Networks
Data collected in a recording session were divided into a training set for developing the neural decoder, a validation set for avoiding overfitting during training phase, and a testing set for evaluating the generalization ability of the neural decoder. For each rat, the experimental trials of the first 2 days were used as training and validation sets, and the remaining 10 days were used as testing set. The number of trials used for each rat was shown in Table 1.
The present study evaluated the prediction accuracy (decoding performance) of the proposed neural decoder using Pearson's correlation coefficient (r), which measured the strength of a linear relationship between the observed and predicted forelimb trajectories (Manohar et al., 2012; Shimoda et al., 2012). When evolving ECPNN-EF with good generalization ability and compact structure, a 5-fold cross validation was adopted to determine the optimal pc, pm and terminated generation. For each given pair of pc and pm, the experimental trials of the first 2 days were randomly partitioned into five equal-sized disjoint sets where four sets were used as the training set to evolve ECPNN-EF and one set was used as the validation set to evaluate the decoding performance of the evolved ECPNN-EF during training phase. Once the ECPNN-EF was evolved through the optimal pc and pm, the validation set also was used to determine the best terminated generation of the evolved ECPNN-EF.
To investigate the effects of CBP and ABSS on the ECPNN-EF evolution, two variants of ECPNN-EF that only adopted either CBP or ABSS were implemented. One variant of ECPNN-EF only with CBP was referred to as ECPNN-EFWC, and the other variant of ECPNN-EF only with ABSS was referred to as ECPNN-EFWA. The decoding performances of ECPNN-EF, ECPNN-EFWC, and ECPNN-EFWA were compared using the validation set in terms of r, number of hidden neurons (Nh), number of connections (Nc), connection ratio (Rc), and termination generation (GT). The Rc was defined as follows:
where Nf is the number of connections in a network with a fully connected topology. The network had fully connected topology when Rc = 1; however, the network had partially connected topology when Rc < 1.
Statistical Analysis
In this study, we investigated the decoding performance dependency on the parameters of pc and pmby employing the statistical method, two-way analysis of variance (ANOVA) followed by Tukey's post-hoc test and adjusted the P-value by multiple comparison using Bonferroni correction, on the validation set. We set pc at 7 levels (0.6, 0.65, 0.7, 0.75, 0.8, 0.85, and 0.9). In addition, 7 levels of pm (0.6, 0.65, 0.7, 0.75, 0.8, 0.85, and 0.9) were employed for each pc to determine whether the algorithm found the near-optimum solution. This evaluated whether there were any significant differences in decoding performance according to the parameters used. Additionally, we analyzed the effects of CBP and ABSS on the ECPNN-EF reconfiguration, and then assessed the decoding performance comparison of ECPNN-EF, ECPNN-EFWC, and ECPNN-EFWA on the validation set by one-way ANOVA with post-hoc Tukey's HSD test.
In order to investigate the decoding performance and stability of ECPNN-EF as well as impact of the prediction error feedback on enhancing the prediction accuracy of ECPNN-EF without error-correction learning (ECPNN), a mixed model ANOVA with three decoders [ECPNN-EF, a fully connected RNN with error feedback (RNN-EF), and ECPNN] as fixed factors and the repeated measure of daily testing set over 10 testing days followed by Tukey's post-hoc test and then adjusted the P-value by Bonferroni multiple comparison correction. The decoding performances of the four rats were presented as means ± standard deviation (SD). The data analysis was performed in SPSS version 20.0 (SPSS Inc., Chicago, IL, USA).
Results
On the Decoding Performance of Different Crossover and Mutation Probabilities
The decoding performances of the ECPNN-EF evolved under different pc and pm values were evaluated by the validation set as shown in Figure 5. The results demonstrated that the more increasing in pc and pm and the worse decoding accuracy (r). A simple main effects analysis which examined the effects of 7 levels of pm at the fixed level of pc = 0.75 and 0.8 was provided in the Supplementary Note 9. The best decoding performance (r = 0.912 ± 0.019) was achieved using pc = 0.75 and pm = 0.75 (compared to other combinations of pc and pm, P < 0.05 analyzed by ANOVA for multiple comparisons).
Figure 5. Decoding performance of the ECPNN-EF under various pc and pm. We performed the post-hoc analysis based on the estimated marginal means of correlation coefficient (r) and adjusted the P-value by multiple comparison using Bonferroni correction (see the Table S2 in the Supplementary Note 8). We found that the highest r was observed when the pm= 0.75 and the pc = 0.75 (r = 0.912 ± 0.019) and showed significant differences in r of pc = 0.8 and pm = 0.75. Therefore, the near-optimum solution of the algorithm was pc= 0.75 and pm = 0.75.
Evolution Progress of ECPNN-EF
Figure 6 presented the evolution progress of the ECPNN-EF using the optimal probabilities of crossover and mutation (pc = 0.75 and pm = 0.75) obtained in Figure 5. The results showed that the GL and the difference between the average training error and minimum training error (Pk) were almost zero in the early generations. Afterward, the GL slightly increased but the Pk varied. Notably, the GL was not consistently larger than the Pk. Most GLs dramatically increased and were larger than the Pk after the 33th generation marked by a black vertical dashed line. This potentially indicated the overfitting problem that might lead to worse evolutions. Therefore, the ECPNN-EF was suggested to terminate evolution in this generation according to an early stopping criterion in order to maintain stable decoding performance. The mean termination generation was 33 in this study (GT = 33.2 ± 1.1).
Figure 6. Evolution progress of the ECPNN-EF. The shaded regions represented SD. The vertical axis on the left represented Pk (blue line). The vertical axis on the right represented GL (red line). Most GLs met the early stopping criterion in the mean terminated generation (GT = 33.2 ± 1.1) indicated by a black vertical dashed line.
Effects of CBP and ABSS on Decoding Performance
Table 2 showed the decoding performance of the ECPNN-EF, ECPNN-EFWC, and ECPNN-EFWA. ECPNN-EF achieved significantly higher decoding performance (r = 0.912 ± 0.019) than did the ECPNN-EFWC (r = 0.602 ± 0.083) and ECPNN-EFWA (r = 0.708 ± 0.066) (P < 0.05 analyzed by one-way ANOVA with post-hoc Tukey's HSD test). The ECPNN-EF possessed a more compact structure (Nh = 4.2 ± 2.7 and Nc = 20.6 ± 7.2) than both ECPNN-EFWC (Nh = 4.6 ± 3.8 and Nc = 23.0 ± 10.1) and ECPNN-EFWA (Nh = 14.4 ± 14.4 and Nc = 57.9 ± 52.1). Moreover, the ECPNN-EFWA had a greater standard deviation than the other two methods. All three methods possessed almost the same Rc. The ECPNN-EF (Rc = 0.12 ± 0.02) and ECPNN-EFWA (Rc = 0.12 ± 0.01) produced slightly more sparse structures than ECPNN-EFWC (Rc = 0.13 ± 0.01). All three methods terminated the evolution before 38th generation. The ECPNN-EF terminated slightly earlier (GT = 33.2 ± 1.1) than did the other two decoders.
Decoding Performance Comparison
We applied a mixed model ANOVA with three decoders (ECPNN-EF, RNN-EF, and ECPNN) as fixed factors and the repeated measure of time, then adjusted the P-value by multiple comparison using Bonferroni correction. The three decoders reconstructed movement trajectories similar to the actual movement trajectories (Figure 7). However, the ECPNN-EF decoder showed the best reconstruction and stability and was significantly better than the ECPNN and RNN-EF decoders over 10 test days.
Figure 7. Data visualization of average predicted trajectories of the ECPNN-EF, ECPNN, and RNN-EF. Representative daily reconstructed trajectories of the test trials in Rat #16. The average reconstructed trajectories of the ECPNN-EF (red line) were more similar to the actual ones (black line) and exhibited less variance than did those of the ECPNN (green line) and RNN-EF (blue line) over 10 test days, where shadow regions represented their corresponding SDs of the predicted trajectories.
To investigate the effectiveness of the partially connected topology of ECPNN-EF, the ECPNN-EF was compared with RNN-EF using the post-hoc analysis. Here, the number of hidden neurons of the RNN-EF was the same as that in the ECPNN-EF for fair comparison [the weights were adjusted by the BPTT] (Werbos, 1988). Figure 8 statistically showed the daily r comparison between the decoders of ECPNN-EF and RNN-EF. The mean r of the decoder of RNN-EF monotonically decreased with gradually increasing in the variability of r over 10 test days. The result showed that the RNN-EF could not offer a stable long-term decoding performance. By contrast, the decoding performance of the ECPNN-EF decreased slightly in each day and achieved r = 0.740 ± 0.042 at Test Day 10. Moreover, the variation in neural decoding performance (SD) of the ECPNN-EF was smaller than that of the RNN-EF in each day. The decoding performance of the ECPNN-EF was significantly higher than that of RNN-EF (P < 0.05 analyzed by repeated measures analysis using mixed model ANOVA with post-hoc test, N = 4) in each day.
Figure 8. Comparison of daily r of the ECPNN-EF, ECPNN, and RNN-EF. The decoding performance of ECPNN-EF at the 1st, 2nd, 3rd, 4th, 5th, 6th, 7th, 8th, 9th, and 10th. Test Day was significantly higher than those of ECPNN and RNN-EF (also see the post-hoc analysis of the comparison of decoding performance in the Table S3 in the Supplementary Note 10), and the corresponding variation of r was smaller than that of the ECPNN and RNN-EF after Test Day 4. The symbols * and # indicate P < 0.05, as analyzed by the repeated measures analysis using mixed model ANOVA with Bonferroni correction for multiple testing.
To investigate the effect of the error-correction learning (error feedback) in the decoder, the ECPNN-EF was compared with ECPNN using the post-hoc test. As depicted in Figure 8, the ECPNN-EF decoder performed higher and more stable accuracy of predicted trajectories in comparison with those of the ECPNN decoder over 10 test days. The mean r of the decoder of ECPNN dropped noticeably and the corresponding variability of r became huge after Test Day 5. By contrast, the ECPNN-EF's daily r slowly decreased, and the daily variability of r did not considerably change over 10 days. The ECPNN-EF's r was significantly higher than that of the ECPNN in each day (P < 0.05 analyzed by repeated measures analysis using mixed model ANOVA with post-hoc test, N = 4), and the corresponding variation (SD) in daily r of the ECPNN-EF was smaller than that of the ECPNN. The lowest r of the ECPNN-EF (r = 0.740 ± 0.042) and ECPNN (r = 0.413 ± 0.158) was observed at Test Day 10.
Discussion
Best Performance With NN Structure Determined by Near Optimal Probabilities of Crossover and Mutation
The pc and pm affected the evolution of the network structure and thus involved neuronal contributions to forelimb movement. Both crossover and mutation operators increased the model complexity. The crossover provided a chance to add hidden neurons while the mutation achieved a small perturbation of model structure by adding connections. Previous paper Schwartz et al. (1988) has reported that individual neuron in the motor cortex discharges with movements in its preferred direction. A high pc led to large network structure which possessed sufficient information processing capability but might result in overfitting. On the contrary, a low pc led to simple network structure which might result in underfitting. A high pm allowed the hidden neurons to have more connections from the neuronal inputs which led to fully connected topology. Increasing pc and pm may not consistently improve the decoding performance of the evolved neural decoder. Increasing pc and pm from 0.6 to 0.75 enlarged the computational complexity of the neural decoder so that relevant neuronal inputs could be accurately decoded into forelimb movement. However, frequent crossover or mutation (high pc and pm, respectively) in the 0.8–0.9 range may introduce redundant connections from irrelevant neurons with firing rates that did not contribute to the kinematic parameters. Conversely, a low pm allowed the hidden neurons to have few connections, resulting in sparse topology. However, some neuronal inputs had a lower likelihood of being processed. Low pc and pm may reproduce a topology that is too sparse to build connections between kinematic parameters and relevant neurons, resulting in less accurate neural decoding. Our experimental results showed that pc = 0.75 and pm = 0.75 could achieve the best decoding performance. The evolved ECPNN-EF possessed not only sufficient hidden neurons to decode neuronal activities, but appropriate topology which selected forelimb movement related inputs to the hidden neurons.
Early Stopping to Counteract Overfitting in Evolutionary Progress of ECPNN-EF
Most evolutions of the ECPNN-EF terminated around the mean termination generation because the early stopping criterion was met. The early stopping criterion employed both training and validation errors. In the early generations, the GL was almost zero, which indicated that the validation error was almost the same with the lowest validation error among the recent generations. This indicates that the validation error did not increase. Although the functional mapping between neural activity and kinematic parameters varied across days due to variability in the neural recording conditions, the training set may have similar neural recording conditions as the validation set. The ECPNN-EF learned the common functional mapping of the training and validation sets in the early generations, allowing for its evolved sparse topology to gradually learn to decode common firing patterns into forelimb movements in the validation set. Furthermore, the Pk was almost zero, which indicated that the average training error was not larger than the minimum training error among the recent generations. This demonstrated that the training error gradually decreased. Both the GL and Pk indicated that the evolution improved the generalization ability of the ECPNN-EF in the early generations. Before the mean termination generation, the slight increase of the GL might indicate overfitting, but the GL was not always higher than the Pk. This implied that the generalization ability of the ECPNN-EF had a chance to be repaired by the evolution as illustrated in Prechelt (1998); Sussillo et al. (2016). Most GLs dramatically increased and were consistently higher than the Pk after the mean termination generation. This demonstrated that the validation error increased, indicating overfitting. Previous work (Kao et al., 2015) has suggested that a neural decoder with too many parameters may result in overfitting. The evolution tended to construct a more complex neural decoder with several weights, potentially contributing to overfitting in the later generations. Therefore, the evolution terminated to prevent decreased generalization ability from overfitting the training set and to save computational time.
Best Performance Based on NN With Good Generalization Ability and Compact Structure
The fact that the ECPNN-EF significantly outperformed the ECPNN-EFWC and ECPNN-EFWA in terms of r suggested that both CBP and ABSS were essential to evolve the neural decoder with generalization ability. The CBP pruned insignificant hidden neurons and led to lower Nh and Nc in ECPNN-EF and ECPNN-EFWC. This mechanism made the network more compact and prevented the network from excessively complex structure caused by the network crossover through many generations. On the other hand, the ECPNN-EFWC's Rc was expected to be considerably larger than those of ECPNN-EF and ECPNN-EFWA because the ABSS tended to select network with sparsely connected topology. However, the difference of Rc among the ECPNN-EF, ECPNN-EFWC, and ECPNN-EFWA was not significant because of the effect of early stopping. All the three approaches stopped evolution before 38 generations. The networks in the population underwent only few crossovers and mutations, and thus their network structures were less complex. Nevertheless, the poor r in the ECPNN-EFWC suggested that although early stopping led to lower Rc, the evolution without ABSS would evolve a neural decoder with poor generalization ability. ABSS selected networks without redundant connections into next generation and thus prevented the network from fully connected topology caused by the network mutation. Excessively complex neural decoders may include redundant hidden neurons that overfit the training set. This can disrupt accurate decoding of neural activity in the testing set, which may have different neural recording conditions from the training set. Furthermore, redundant weights may connect to neurons with preferred directions that are irrelevant to vertical or horizontal velocities.
The ECPNN-EF terminated and obtained near-optimum neural decoder earlier than the ECPNN-EFWC and ECPNN-EFWA. The validation error in these three models increased after their termination generation. This resulted in overfitting because their evolved structures were more complex than the near-optimum neural decoders. Evolution of the ECPNN-EF was more efficient than the ECPNN-EFWC and ECPNN-EFWA, indicating that it obtained near-optimum neural decoder faster than its reduced models. Thus, ECPNN-EF's termination generation was earlier than its reduced models. These results indicate that CBP and ABSS helped evolve a neural decoder with less complex structure and better generalization ability. Power efficiency and power management are extremely important concerns for fully implantable neural decoders in BMIs. Due to its sparse topology, ECPNN-EF offers a practical approach to computationally efficient neural decoding by reducing the number of hidden neurons and interlayer connections, resulting in less memory usage and power consumption (Chen et al., 2015). Less power consumption results in a longer battery lifetime, which could facilitate brain implantation of neural decoders (Sarpeshkar et al., 2008).
Best Performance Based on Appropriate Connected Topology-Based Network
Several studies have shown that a partially connected NN (PCNN) achieves better performance than does a fully connected neural network (FCNN) (Elizondo and Fiesler, 1997). The ECPNN-EF, a type of PCNN, achieved significantly higher daily mean r than does the RNN-EF, which is a type of FCNN; this suggests that an FCNN might consist of a large amount of redundant connections and lead to overfitting with poor generalization when compared to a PCNN (Elizondo and Fiesler, 1997; Wong et al., 2010; Guo et al., 2012). Some information, which was irrelevant to the forelimb movement and was processed by the redundant connections, may hamper the performance of the neural decoder and increase the likelihood of NNs being stuck in local minima. Furthermore, the variation in decoding performance of the RNN-EF from Test Day 3 to Test Day 10 was larger than that for Test Day 1 and Test Day 2, whereas the variation in decoding performance of the ECPNN-EF did not change dramatically. This suggested that the redundant weights in the RNN-EF could not deal with the variation of the neural recording conditions and thus led to unstable decoding performance. The ECPNN-EF outperformed the RNN-EF due to the use of the partially connected topology. The trends observed in the present study followed the suggestion that the number of connections is not the key aspect of an NN but rather of an appropriate connected topology (Yang and Chen, 2012).
Comparing Linear Neural Decoder-Based Error-Correction Learning
Our prior work demonstrated a linear decoding model of the relationship between neural firing and kinematic parameters (Yang et al., 2016). A sliced inverse regression (SIR) with error-feedback learning (SIR-EF) was implemented based on an SIR linear neural decoder to fairly compare to the ECPNN-EF algorithm (see Supplementary Note 11). Because the ECPNN-EF had to process changing neural recording conditions over time, it possessed more processing capabilities than the linear model. The SIR-EF could not deal with long-term variability in neural recording conditions because of linear properties and limited computational complexity. The SIR-EF assigned weights to the slices with neurons that had a similar contribution to the lever-pressing forelimb movement. However, variations in neural recording conditions due to the tissue's reaction to neural implants or micromotion of the electrodes across days resulted in firing pattern variations (Barrese et al., 2013; Sussillo et al., 2016). Thus, the decoding performance decreased because the weights calculated using the training data over the first 2 days could not predict velocity in the subsequent testing days with different neural conditions.
NN-Based Error-Correction Learning to Improve Long-Term Decoding Stability
It has been revealed that the functional mapping between instantaneous firing rate and kinematic parameters might vary in chronic recording due to the changes of neural recording conditions (Sussillo et al., 2016). A relative increase in ECPNN's mean r at Test Day 3 and Test Day 5 might exhibit that the recording conditions probably had some commonality with those in the training phase. Therefore, the non-linear model of ECPNN, which learned the time-dependent functional mapping from the training set, could accurately decode the instantaneous firing rate into kinematic parameters. The error feedback played a subsidiary role of the neural decoding in this situation. A considerable decrease in ECPNN's mean r after Test Day 5 might indicate that the neural recording conditions were different from those in the training phase. The learned functional mapping between instantaneous firing rate and kinematic parameters was of no use for making ECPNN's long-term decoding performance stable.
In contrast, when the neural recording conditions changed after Test Day 5, ECPNN-EF′ error feedback provided immediate kinematic information and thus compensated for across-day changes in functional mapping between instantaneous firing rate and kinematic parameters. The ECPNN-EF could achieve not only more robust within-day decoding (smaller SD) but also more robust across-day decoding (smaller fluctuations in daily mean r) than those of the ECPNN. This demonstrated that employing error feedback in the ECPNN-EF improved the long-term decoding stability when only a few days of training data were available.
Data Availability Statement
The datasets generated for this study are available on request to the corresponding author.
Ethics Statement
The animal study was reviewed and approved by the Institutional Animal Care and Use Committee of the Taipei Medical University.
Author Contributions
S-HY, C-HK, and Y-YC designed the project, organized the entire research. Y-CL, H-YL, and S-HL conceived the experiments. CC, J-WH, H-LW, C-FW, and C-CK conducted the experiments. CC, S-HY, Y-CL, H-LW, K-YC, and C-FW implemented code from software designed and planned. CC, S-HL, J-WH, H-YL, and Y-HL analyzed the results. S-HY, CC, and Y-YC wrote the manuscript. All authors discussed the results and reviewed on the manuscript.
Funding
This work was financially supported by Ministry of Science and Technology of Taiwan under Contract numbers of MOST 109-2636-E-006-010 (Young Scholar Fellowship Program), 108-2321-B-010-008-MY2, 108-2314-B-303-010, 108-2636-E-006-010, 107-2221-E-010-021-MY2, and 107-2221-E-010-011. We also are grateful for support from the Headquarters of University Advancement at the National Cheng Kung University, which is sponsored by the Ministry of Education, Taiwan.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fncom.2020.00022/full#supplementary-material
Supplemental Video. The supplemental video showed forelimb reaching for lever pressing and simultaneous neural recording in M1.
Abbreviations
MI, brain machine interface; ECPNN-EF, evolutionary constructive and pruning neural network with error feedback; NN, neural network; RNN, recurrent neural network; ESN, echo-state network; TDNN, time-delay neural network; EA, evolutionary algorithm; ECPA, evolutionary constructive and pruning algorithm; CBP, cluster-based pruning; ABSS, age-based survival selection; BPTT, backpropagation through time; ECPNN-EFWC, ECPNN-EF only with CBP; ECPNN-EFWA, ECPNN-EF only with ABSS; RNN-EF, fully-connected RNN with error feedback; ECPNN, ECPNN-EF without error-correction learning; PCNN, partially connected NN; FCNN, fully connected neural network.
References
Angeline, P. J., Saunders, G. M., and Pollack, J. B. (1994). An evolutionary algorithm that constructs recurrent neural networks. IEEE Transact. Neural Netw. 5, 54–65. doi: 10.1109/72.265960
Barrese, J. C., Rao, N., Paroo, K., Triebwasser, C., Vargas-Irwin, C., Franquemont, L., et al. (2013). Failure mode analysis of silicon-based intracortical microelectrode arrays in non-human primates. J. Neural Eng. 10:066014. doi: 10.1088/1741-2560/10/6/066014
Bishop, W., Chestek, C. C., Gilja, V., Nuyujukian, P., Foster, J. D., Ryu, S. I., et al. (2014). Self-recalibrating classifiers for intracortical brain–computer interfaces. J. Neural Eng. 11:026001. doi: 10.1088/1741-2560/11/2/026001
Chen, Y., Yao, E., and Basu, A. (2015). A 128-channel extreme learning machine-based neural decoder for brain machine interfaces. IEEE Transact. Biomed. Circuits Syst. 10, 679–692. doi: 10.1109/TBCAS.2015.2483618
Chen, Y.-Y., Lai, H.-Y., Lin, S.-H., Cho, C.-W., Chao, W.-H., Liao, C.-H., et al. (2009). Design and fabrication of a polyimide-based microelectrode array: Application in neural recording and repeatable electrolytic lesion in rat brain. J. Neurosci. Methods 182, 6–16. doi: 10.1016/j.jneumeth.2009.05.010
Chestek, C. A., Gilja, V., Nuyujukian, P., Foster, J. D., Fan, J. M., Kaufman, M. T., et al. (2011). Long-term stability of neural prosthetic control signals from silicon cortical arrays in rhesus macaque motor cortex. J. Neural Eng. 8:045005. doi: 10.1088/1741-2560/8/4/045005
Collinger, J. L., Wodlinger, B., Downey, J. E., Wang, W., Tyler-Kabara, E. C., Weber, D. J., et al. (2013). High-performance neuroprosthetic control by an individual with tetraplegia. Lancet 381, 557–564. doi: 10.1016/S0140-6736(12)61816-9
Connor, J. T., Martin, R. D., and Atlas, L. E. (1994). Recurrent neural networks and robust time series prediction. IEEE Transact. Neural Netw. 5, 240–254. doi: 10.1109/72.279188
Downey, J. E., Schwed, N., Chase, S. M., Schwartz, A. B., and Collinger, J. L. (2018). Intracortical recording stability in human brain–computer interface users. J. Neural Eng. 15:046016. doi: 10.1088/1741-2552/aab7a0
Elizondo, D., and Fiesler, E. (1997). A survey of partially connected neural networks. Int. J. Neural Syst. 8, 535–558. doi: 10.1142/S0129065797000513
Fernández, E., Greger, B., House, P. A., Aranda, I., Botella, C., Albisua, J., et al. (2014). Acute human brain responses to intracortical microelectrode arrays: challenges and future prospects. Front. Neuroeng. 7:24. doi: 10.3389/fneng.2014.00024
Gilja, V., Nuyujukian, P., Chestek, C. A., Cunningham, J. P., Byron, M. Y., Fan, J. M., et al. (2012). A high-performance neural prosthesis enabled by control algorithm design. Nat. Neurosci. 15, 1752–1757. doi: 10.1038/nn.3265
Glaser, J. I., Benjamin, A. S., Chowdhury, R. H., Perich, M. G., Miller, L. E., and Kording, K. P. (2017). Machine learning for neural decoding. arXiv:1708.00909 [q-bio.NC].
Guo, Z., Wong, W. K., and Li, M. (2012). Sparsely connected neural network-based time series forecasting. Inform. Sci. 193, 54–71. doi: 10.1016/j.ins.2012.01.011
Hong, G., and Lieber, C. M. (2019). Novel electrode technologies for neural recordings. Nat. Rev. Neurosci. 20, 330–345. doi: 10.1038/s41583-019-0140-6
Hornik, K., Stinchcombe, M., and White, H. (1989). Multilayer feedforward networks are universal approximators. Neural Netw. 2, 359–366. doi: 10.1016/0893-6080(89)90020-8
Huang, D.-S., and Du, J.-X. (2008). A constructive hybrid structure optimization methodology for radial basis probabilistic neural networks. IEEE Transact. Neural Netw. 19, 2099–2115. doi: 10.1109/TNN.2008.2004370
Islam, M. M., Sattar, M. A., Amin, M. F., Yao, X., and Murase, K. (2009). A new adaptive merging and growing algorithm for designing artificial neural networks. IEEE Transact. Syst. Man Cybernet. 39, 705–722. doi: 10.1109/TSMCB.2008.2008724
Jackson, A., Mavoori, J., and Fetz, E. E. (2006). Long-term motor cortex plasticity induced by an electronic neural implant. Nature 444, 56–60. doi: 10.1038/nature05226
Jaeger, H., and Haas, H. (2004). Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 304, 78–80. doi: 10.1126/science.1091277
Jarosiewicz, B., Masse, N. Y., Bacher, D., Cash, S. S., Eskandar, E., Friehs, G., et al. (2013). Advantages of closed-loop calibration in intracortical brain–computer interfaces for people with tetraplegia. J. Neural Eng. 10:046012. doi: 10.1088/1741-2560/10/4/046012
Jarosiewicz, B., Sarma, A. A., Bacher, D., Masse, N. Y., Simeral, J. D., Sorice, B., et al. (2015). Virtual typing by people with tetraplegia using a self-calibrating intracortical brain-computer interface. Sci. Transl. Med. 7:313ra179–313ra179. doi: 10.1126/scitranslmed.aac7328
Kao, J. C., Nuyujukian, P., Ryu, S. I., Churchland, M. M., Cunningham, J. P., and Shenoy, K. V. (2015). Single-trial dynamics of motor cortex and their applications to brain-machine interfaces. Nat. Commun. 6:7759. doi: 10.1038/ncomms8759
Kao, J. C., Stavisky, S. D., Sussillo, D., Nuyujukian, P., and Shenoy, K. V. (2014). Information systems opportunities in brain–machine interface decoders. Proc. IEEE 102, 666–682. doi: 10.1109/JPROC.2014.2307357
Kawato, M. (1990). “Feedback-error-learning neural network for supervised motor learning,” in Advanced Neural Computers (Elsevier), 365–372. doi: 10.1016/B978-0-444-88400-8.50047-9
Kaylani, A., Georgiopoulos, M., Mollaghasemi, M., and Anagnostopoulos, G. C. (2009). AG-ART: an adaptive approach to evolving ART architectures. Neurocomputing 72, 2079–2092. doi: 10.1016/j.neucom.2008.09.016
Kifouche, A., Vigneron, V., Shamsollahi, M. B., and Guessoum, A. (2014). “Decoding hand trajectory from primary motor cortex ECoG using time delay neural network,” in International Conference on Engineering Applications of Neural Networks eds V. Mladenov, C. Jayne, and L. Iliadis (Sofia: Springer), 237–247. doi: 10.1007/978-3-319-11071-4_23
Kim, S.-P., Wood, F., Fellows, M., Donoghue, J. P., and Black, M. J. (2006). “Statistical analysis of the non-stationarity of neural population codes,” in Biomedical Robotics and Biomechatronics, 2006. The First IEEE/RAS-EMBS International Conference on: IEEE (Pisa), 811–816.
Kwok, T.-Y., and Yeung, D.-Y. (1997). Constructive algorithms for structure learning in feedforward neural networks for regression problems. IEEE Transact. Neural Netw. 8, 630–645. doi: 10.1109/72.572102
Lin, H.-C., Pan, H.-C., Lin, S.-H., Lo, Y.-C., Shen, E. T.-H., Liao, L.-D., et al. (2016). Central thalamic deep-brain stimulation alters striatal-thalamic connectivity in cognitive neural behavior. Front. Neural Circuits 9:87. doi: 10.3389/fncir.2015.00087
Løvaas, C., Seron, M. M., and Goodwin, G. C. (2008). Robust output-feedback model predictive control for systems with unstructured uncertainty. Automatica 44, 1933–1943. doi: 10.1016/j.automatica.2007.10.003
Mahmud, M. S., and Meesad, P. (2016). An innovative recurrent error-based neuro-fuzzy system with momentum for stock price prediction. Soft Comput. 20, 4173–4191. doi: 10.1007/s00500-015-1752-z
Manohar, A., Flint, R. D., Knudsen, E., and Moxon, K. A. (2012). Decoding hindlimb movement for a brain machine interface after a complete spinal transection. PLoS ONE 7:e52173. doi: 10.1371/journal.pone.0052173
Masutti, T. A., and de Castro, L. N. (2009). Neuro-immune approach to solve routing problems. Neurocomputing 72, 2189–2197. doi: 10.1016/j.neucom.2008.07.015
Michelson, N. J., Vazquez, A. L., Eles, J. R., Salatino, J. W., Purcell, E. K., Williams, J. J., et al. (2018). Multi-scale, multi-modal analysis uncovers complex relationship at the brain tissue-implant neural interface: new emphasis on the biological interface. J. Neural Eng. 15:033001. doi: 10.1088/1741-2552/aa9dae
Miyamoto, H., Kawato, M., Setoyama, T., and Suzuki, R. (1988). Feedback-error-learning neural network for trajectory control of a robotic manipulator. Neural Netw. 1, 251–265. doi: 10.1016/0893-6080(88)90030-5
Nuyujukian, P., Kao, J. C., Fan, J. M., Stavisky, S. D., Ryu, S. I., and Shenoy, K. V. (2014). Performance sustaining intracortical neural prostheses. J. Neural Eng. 11:066003. doi: 10.1088/1741-2560/11/6/066003
Orsborn, A. L., Moorman, H. G., Overduin, S. A., Shanechi, M. M., Dimitrov, D. F., and Carmena, J. M. (2014). Closed-loop decoder adaptation shapes neural plasticity for skillful neuroprosthetic control. Neuron 82, 1380–1393. doi: 10.1016/j.neuron.2014.04.048
Pais-Vieira, M., Lebedev, M., Kunicki, C., Wang, J., and Nicolelis, M. A. L. (2013). A brain-to-brain interface for real-time sharing of sensorimotor information. Sci. Rep. 3, 1319–1319. doi: 10.1038/srep01319
Paninski, L., Fellows, M. R., Hatsopoulos, N. G., and Donoghue, J. P. (2004). Spatiotemporal tuning of motor cortical neurons for hand position and velocity. J. Neurophysiol. 91, 515–532. doi: 10.1152/jn.00587.2002
Perge, J. A., Homer, M. L., Malik, W. Q., Cash, S., Eskandar, E., Friehs, G., et al. (2013). Intra-day signal instabilities affect decoding performance in an intracortical neural interface system. J. Neural Eng. 10:036004. doi: 10.1088/1741-2560/10/3/036004
Perge, J. A., Zhang, S., Malik, W. Q., Homer, M. L., Cash, S., Friehs, G., et al. (2014). Reliability of directional information in unsorted spikes and local field potentials recorded in human motor cortex. J. Neural Eng. 11:046007. doi: 10.1088/1741-2560/11/4/046007
Prechelt, L. (1998). Automatic early stopping using cross validation: quantifying the criteria. Neural Netw. 11, 761–767. doi: 10.1016/S0893-6080(98)00010-0
Reed, R. (1993). Pruning algorithms-a survey. IEEE Transact. Neural Netw. 4, 740–747. doi: 10.1109/72.248452
Roelfsema, P. R., Denys, D., and Klink, P. C. (2018). Mind reading and writing: the future of neurotechnology. Trends Cognit. Sci. 22, 598–610. doi: 10.1016/j.tics.2018.04.001
Salatino, J. W., Ludwig, K. A., Kozai, T. D., and Purcell, E. K. (2017). Glial responses to implanted electrodes in the brain. Nat. Biomed. Eng. 1:862. doi: 10.1038/s41551-017-0154-1
Sanchez, J., Principe, J., Carmena, J., Lebedev, M. A., and Nicolelis, M. (2004). “Simultaneus prediction of four kinematic variables for a brain-machine interface using a single recurrent neural network.” in Engineering in Medicine and Biology Society, 2004. IEMBS'04, 26th Annual International Conference of the IEEE (San Francisco, CA: IEEE), 5321–5324.
Sanchez, J. C., Erdogmus, D., Nicolelis, M. A., Wessberg, J., and Principe, J. C. (2005). Interpreting spatial and temporal neural activity through a recurrent neural network brain-machine interface. IEEE Transact. Neural Syst. Rehabil. Eng. 13, 213–219. doi: 10.1109/TNSRE.2005.847382
Sarpeshkar, R., Wattanapanitch, W., Arfin, S. K., Rapoport, B. I., Mandal, S., Baker, M. W., et al. (2008). Low-power circuits for brain–machine interfaces. IEEE Transact. Biomed. Circuits Syst. 2, 173–183. doi: 10.1109/TBCAS.2008.2003198
Schwartz, A. B., Kettner, R. E., and Georgopoulos, A. P. (1988). Primate motor cortex and free arm movements to visual targets in three-dimensional space. I. Relations between single cell discharge and direction of movement. J. Neurosci. 8, 2913–2927. doi: 10.1523/JNEUROSCI.08-08-02913.1988
Shah, S., Haghi, B., Kellis, S., Bashford, L., Kramer, D., Lee, B., et al. (2019). “Decoding Kinematics from Human Parietal Cortex using Neural Networks,” in 2019 9th International IEEE/EMBS Conference on Neural Engineering (NER): IEEE, 1138–1141. doi: 10.1109/NER.2019.8717137
Shimoda, K., Nagasaka, Y., Chao, Z. C., and Fujii, N. (2012). Decoding continuous three-dimensional hand trajectories from epidural electrocorticographic signals in Japanese macaques. J. Neural Eng. 9:036015. doi: 10.1088/1741-2560/9/3/036015
Simeral, J., Kim, S.-P., Black, M., Donoghue, J., and Hochberg, L. (2011). Neural control of cursor trajectory and click by a human with tetraplegia 1000 days after implant of an intracortical microelectrode array. J. Neural Eng. 8:025027. doi: 10.1088/1741-2560/8/2/025027
Slutzky, M. W. (2018). Brain-machine interfaces: powerful tools for clinical treatment and neuroscientific investigations. Neuroscientist 25, 139–154. doi: 10.1177/1073858418775355
Sussillo, D., Nuyujukian, P., Fan, J. M., Kao, J. C., Stavisky, S. D., Ryu, S., et al. (2012). A recurrent neural network for closed-loop intracortical brain–machine interface decoders. J. Neural Eng. 9:026027. doi: 10.1088/1741-2560/9/2/026027
Sussillo, D., Stavisky, S. D., Kao, J. C., Ryu, S. I., and Shenoy, K. V. (2016). Making brain-machine interfaces robust to future neural variability. Nat. Commun. 7, 13749–13749. doi: 10.1038/ncomms13749
Tampuu, A., Matiisen, T., Ólafsdóttir, H. F., Barry, C., and Vicente, R. (2019). Efficient neural decoding of self-location with a deep recurrent network. PLoS Comput. Biol. 15:e1006822. doi: 10.1371/journal.pcbi.1006822
Todorov, E., and Jordan, M. I. (2002). Optimal feedback control as a theory of motor coordination. Nat. Neurosci. 5, 1226–1235. doi: 10.1038/nn963
Waheeb, W., Ghazali, R., and Herawan, T. (2016). Ridge polynomial neural network with error feedback for time series forecasting. PLoS ONE 11:e0167248. doi: 10.1371/journal.pone.0167248
Waibel, A., Hanazawa, T., Hinton, G., Shikano, K., and Lang, K. J. (1989). Phoneme recognition using time-delay neural networks. IEEE Transact. Acoustics Speech Signal Proce. 37, 328–339. doi: 10.1109/29.21701
Werbos, P. J. (1988). Generalization of backpropagation with application to a recurrent gas market model. Neural Netw. 1, 339–356. doi: 10.1016/0893-6080(88)90007-X
Willett, F. R., Young, D. R., Murphy, B. A., Memberg, W. D., Blabe, C. H., Pandarinath, C., et al. (2019). Principled BCI decoder design and parameter selection using a feedback control model. Sci. Rep. 9:8881. doi: 10.1038/s41598-019-44166-7
Wodlinger, B., Downey, J., Tyler-Kabara, E., Schwartz, A., Boninger, M., and Collinger, J. (2014). Ten-dimensional anthropomorphic arm control in a human brain– machine interface: difficulties, solutions, and limitations. J. Neural Eng. 12:016011. doi: 10.1088/1741-2560/12/1/016011
Wong, W.-K., Guo, Z., and Leung, S. (2010). Partially connected feedforward neural networks on Apollonian networks. Physica A 389, 5298–5307. doi: 10.1016/j.physa.2010.06.061
Wu, W., Black, M. J., Mumford, D., Gao, Y., Bienenstock, E., and Donoghue, J. P. (2004). Modeling and decoding motor cortical activity using a switching Kalman filter. IEEE Transact. Biomed. Eng. 51, 933–942. doi: 10.1109/TBME.2004.826666
Yang, S.-H., and Chen, Y.-P. (2012). An evolutionary constructive and pruning algorithm for artificial neural networks and its prediction applications. Neurocomputing 86, 140–149. doi: 10.1016/j.neucom.2012.01.024
Yang, S.-H., Chen, Y.-Y., Lin, S.-H., Liao, L.-D., Lu, H. H.-S., Wang, C.-F., et al. (2016). A sliced inverse regression (SIR) decoding the forelimb movement from neuronal spikes in the rat motor cortex. Front. Neurosci. 10, 556–556. doi: 10.3389/fnins.2016.00556
Keywords: brain machine interfaces, neural decoding, error feedback, evolutionary algorithm, recurrent neural network
Citation: Yang S-H, Wang H-L, Lo Y-C, Lai H-Y, Chen K-Y, Lan Y-H, Kao C-C, Chou C, Lin S-H, Huang J-W, Wang C-F, Kuo C-H and Chen Y-Y (2020) Inhibition of Long-Term Variability in Decoding Forelimb Trajectory Using Evolutionary Neural Networks With Error-Correction Learning. Front. Comput. Neurosci. 14:22. doi: 10.3389/fncom.2020.00022
Received: 20 November 2019; Accepted: 04 March 2020;
Published: 31 March 2020.
Edited by:
Chun-Shu Wei, National Chiao Tung University, TaiwanReviewed by:
Eric Bean Knudsen, University of California, Berkeley, United StatesTing Zhao, Howard Hughes Medical Institute (HHMI), United States
Copyright © 2020 Yang, Wang, Lo, Lai, Chen, Lan, Kao, Chou, Lin, Huang, Wang, Kuo and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: You-Yin Chen, aXJyYWRpYW5jZSYjeDAwMDQwO3NvLW5ldC5uZXQudHc=