- 1Conec Laboratory, School of Psychology, Ottawa University, Ottawa, ON, Canada
- 2Department of Computer Science, Cégep du Vieux Montréal, Montreal, QC, Canada
Visual motion detection is essential for the survival of many species. The phenomenon includes several spatial properties, not fully understood at the level of a neural circuit. This paper proposes a computational model of a visual motion detector that integrates direction and orientation selectivity features. A recent experiment in the Drosophila model highlights that stimulus orientation influences the neural response of direction cells. However, this interaction and the significance at the behavioral level are currently unknown. As such, another objective of this article is to study the effect of merging these two visual processes when contextualized in a neuro-robotic model and an operant conditioning procedure. In this work, the learning task was solved using an artificial spiking neural network, acting as the brain controller for virtual and physical robots, showing a behavior modulation from the integration of both visual processes.
1. Introduction
Visual motion detection (MD), direction selectivity (DS) and orientation selectivity (OS) are essential basic mechanisms for processing visual input from the environment (Borst and Euler, 2011; Clark and Demb, 2016; Nath and Schwartz, 2016). However, the neural correlates at the level of cellular circuitry are not fully understood (Takemura et al., 2013). The study of elementary MD and DS models under the umbrella of computational vision is based on a few theories (Hassenstein and Reichardt, 1956; Hubel and Wiesel, 1959; Barlow and Levick, 1965). The basic algorithm of a MD relies on the integration across space and time of a moving light or dark stimuli (Yonehara and Roska, 2013; Behnia et al., 2014), while DS property is mainly achieved from facilitating the response to preferred motion and/or inhibiting the response to the null motion (Clifford and Ibbotson, 2002; Fried et al., 2002; Li et al., 2014; Mauss et al., 2015; Salay and Huberman, 2015).
Several studies have used the well-known Drosophila model in vision science (Paulk et al., 2013), validating underlying mechanisms of DS (Eichner et al., 2011; Gilbert, 2013; Maisak et al., 2013; Shinomiya et al., 2014; Leong et al., 2016; Haag et al., 2017). Recently, insights from the Drosophila brain have shown that few visual neurons display both directional tuning and orientation selectivity (Fisher et al., 2015). Notably, when the axis of motion is orthogonal to the orientation of the moving stimulus, directional tuning is sharpened. As the orientation of the moving stimulus aligns in parallel to the direction of the axis of motion, neuronal responses are reduced.
Orientation and direction selectivity from retina to cortex were experimentally shown in various mammalian species and rigorously quantified from statistics methods (Borg-Graham, 2001; Mazurek et al., 2014). The emergence of DS and the influence of visual experience are extensively studied (Li et al., 2008; Haag et al., 2016; Leong et al., 2016; Strother et al., 2017) as well as the elaboration of computational models (Mu and Poo, 2006; Elstrott and Feller, 2009; Berberian et al., 2017). One possibility may consist of a bias architecture early in the neural development toward specific preferences (Adams and Harris, 2015). There is also computational work suggesting that spontaneous activity appearing during the early stages may give rise to the emergence selectivity features (Van Hooser et al., 2012). The foundation of this may find echoes in the genetics and from the primary units in the retina that already compute and provide the information at that level.
This research presents a spiking neural network (SNN) model to study the interaction between visual orientation and direction selectivity features in a MD model that responds to basic visual motion stimuli. SNN is a relevant computational method to use given the temporal property that helps capture dynamic and coincidental events (Maass, 1997) using spike-timing-dependent plasticity (STDP) (Bi and Poo, 1998; Feldman, 2012). However, SNN remains poorly used regarding the MD and DS visual topics (Shon et al., 2004; Wenisch et al., 2005) especially in complete embodied models. A key advantage of using physical robots in neural modeling is to validate models under real world constraints (Webb, 2000).
Bio-inspired neural models in vision (Kerr et al., 2015) and motion detector models (Franceschini et al., 1992) are not new topic in neuro-robotics. Simulation of a stabilization and fixation robotic behavior from a motion stimulus reproduce mainly the visuomotor process of the fly (Huber and Bülthoff, 1998). But merging OS and DS in SNN paradigm is still unexplored. Furthermore, given that the relationship between orientation and direction selectivity remains to be investigated at the behavioral level, this research aims to embody these two related processes in virtual and physical robots as a proof of concept (Pezzulo et al., 2011; Krichmar, 2018). In this perspective, the present model was evaluated under an operant conditioning context, modulating its behavioral response when shown basic orientated stimuli in motion. More precisely, a detailed framework to trace dynamical visual stimuli from sensors to motors is proposed, which could be used in future robotic implementations in the computational vision domain.
In this experiment, the operant conditioning learning process (Cyr et al., 2014) is used as behavioral context. As such, a reward mechanism reinforces connections amongst units coding for the preferred direction of motion in relation to its neutral actions. The application of a positive reward provides the advantage of starting off with no initial assumption about the underlying behavior of units exposed to stimuli in motion. From this learning procedure and with the knowledge of the orthogonal (orientation/motion) aspect of a stimulus, a fasten motor response is proposed. The contribution of this paper is to introduce a bio-inspired model of motion detector integrating direction and orientation selectivity features, implementing these processes at a behavioral robotic level.
The next sections detail the SNN architecture and the simulation environments used, followed by an analysis of the obtained results. It concludes with a discussion on the model and its future perspectives.
2. Methodology
The goal is to simulate an enhanced behavioral response of a virtual and physical robot, when the orientation of a visual stimulus is orthogonal to its motion. From an operant conditioning procedure, the robot learns to link a positive reward with actions of lighting up LEDs and choosing the desired solution. The additive effect of the orientation and motion features of a visual stimulus was demonstrated in the Drosophila, a challenge to model in a neurorobotic paradigm.
2.1. Protocol
The virtual experimentation consists in displaying black lines (horizontal and vertical) that move horizontally and vertically in front of a robot (see Figures 1, 2). The SNN architecture, as well as the 3D world experiment were elaborated with the SIMCOG software (Cyr et al., 2009). Four scenarios were evaluated in the two opposite motion directions, for a total of 8 different trials. In each of them, a line passes from one end of the retina to the other. In this study, the retina is composed of a 3 × 3 sensory neurons matrix.
Figure 1. Four trial examples made in the experiment. The first two images show a line moving orthogonal to its orientation. The last two images show lines moving in their same orientation. Those four scenarios are evaluated for both directions, for example a horizontal line moving from left to right and right to left. The circles, representing the sensory receptors of the visual field, overlap to reduce unseen areas.
Figure 2. Virtual and real world environments, showing the robots in front of an image. The bottom left section represents what the virtual robot currently sees with its camera.
Each combination of orientation/direction trial is repeated several times in order to reach the learning criteria. The sequence of trials is pre-determined. At the beginning of the simulation, the robot randomly chooses an action, by blinking a light on one of its four possible LEDs, following the capture of a moving visual stimulus. The robot learns from a supervised positive reinforcement to correctly associate the desired output with its own previous actions. Finally, the simulation ends with the evaluation of the behavioral response combined with the orthogonal/non-orthogonal feature of stimuli (motion/direction). In this study, foreign patterns were not tested, since the other stimulus features were not used (only vertical/horizontal and motion).
The virtual simulation lasted 24,000 cycles (3,000 cycles for each trial). A particularity of the software is that it works using cycles instead of milliseconds. This allows computers of various power to have the same output at the end. The approximation conversion is 10 ms/cycle for the virtual experiment, running on an i7 desktop computer.
2.2. Architecture
The general topology of the SNN consists of several neural layers, as shown in Figure 3. A sensory layer captures the visual stimuli, then the orientation and motion features are extracted and forwarded to an integrative neural layer. A Decision layer then proposes a random action to the motor layer. Once the learning is completed, A Force neural layer overrides the initial random decisions. Following a desired output, an external positive reinforcement is applied to the robot and caught from arbitrary dedicated Reward neuron.
Figure 3. Flow diagram of the main components of the SNN, following the perception of a moving stimulus in front of the robot.
For illustrative reasons, a simplified neural circuit based on three sensory neurons instead of the full 3x3 matrix is shown in Figure 4 (see complete SNN architecture and the table values of neural parameters in Supplementary Material at http://aifuture.com/res/2018-dir). Though it is sufficient to highlight the visual mechanism of an enhanced motion/orientation neural response.
Figure 4. Simplified SNN architecture. Following the reception of a moving visual stimulus perceived from Vision neurons, the linked Direction and Orientation neurons forward the information to Integration neurons. At this point, the features of the stimulus are obtained and used for the learning task. This involves Predictor and the Reward neurons to enhance the synaptic links from the STDP function (i.e., synapses between Integration and Predictor neurons). In the virtual environment, vibrations act as rewards, hence the gray square at the center of the robot that represents the vibration transducer. When the rule is learnt, Force neurons bypass the random Decision neurons to trigger the appropriate action (i.e., LED action neurons).
On the left part of Figure 4, one can see that the robot has four different binary output responses. These consist of LEDs located at the four cardinal points on top of the robot. Each are attached to their respective motor-neurons (LEDGoRight, LEDGoDown, LEDGoLeft,LEDGoUp). This visual computational model includes a camera at the front of the robot (next to the yellow LED on the figure) and a sensor to capture the reward. In this study, a vibration sensor was chosen and is automatically triggered by the virtual environment when a proper decision is made. The visual black bars stimuli are caught by the camera and their linked sensory neurons (blue circles), see right part of the figure. At this point, Sensory neurons fall under a cooldown period using a refractory period parameter, to prevent the constant capture of stimuli. Then, these neurons forward the signal to both motion and orientation neurons. In case of motion, the neurons receive inputs with temporal fixed synaptic delays (see Figure 5) to achieve the integration.
Figure 5. Visual orientation and direction selectivity features extraction process. The orientation as well as the motion information (green neurons in this figure) are obtained by merging two adjacent sensory neurons (blue neurons). For the motion feature, delayed synapses were used.
For simplicity, instead of introducing inhibition in the neuronal responses for motion in the non-preferred direction as in the natural model, this SNN uses identical dedicated excitatory synapses for both opposite directions. Thus, direction and orientation neurons connect with excitatory synapses on integration neurons, providing the orthogonal feature (motion with the inverse orientation). Those integrative neurons are linked to Predictor neurons with a weak synaptic link. Therefore, the Integrative neurons cannot trigger Predictor neurons alone. Predictor neurons also receive a positive reinforcement signal from a Reward neuron (vibration sensory neuron), in order to meet the pre/post timing criteria relation of the learning rule. When a Predictor neuron spikes, it forwards signal to the Force neuron that will trigger the output response. Since there are multiple synapses between the Integrative neurons and the Predictor neurons, the behavioral plasticity is summed between the motion neurons and the actions of the robot. Finally, Motor neurons (gray circles in Figure 4) output to the attached LEDs. The result of the learning process is that after a certain time, rewards will not be necessary anymore to forward the signal through the circuit. Once an action is selected, all other actions are inhibited from a dedicated neuron (not shown in Figure 4, for visibility).
2.3. Neural Dynamic
The spiking neuron model includes standard features such as a time varying membrane potential variation, a refractory period, a threshold and a spike emission state (Equations 1, 2, 3, and 4). These neurons are connected through dynamical excitatory and inhibitory synapses, and some are modulated from a spike-timing-dependent plasticity (STDP) learning rule (Equation 5). Equation (2) represents a nonlinear potential variation simulating an excitatory or inhibitory postsynaptic potential (PSP). This architecture also uses some fixed PSPs with different lengths to regulate control neural inputs.
In this study, the learning rule from STDP needs a third factor (the reward) to be activated (Frémaux et al., 2010; Kuśmierz et al., 2017) (STDP-R). A negative reinforcement (punishment) could have been used to modulate the learning curve, but it was not implemented. Mainly, the objective is to reach a specific synaptic weight value in order to force the proper action upon an associative event. This value depends on the initial synaptic weight and the increase step of the STDP function (100%, as specified in Equation 5). Also, to prevent overshooting this threshold, a capping value was specified to 300%.
Equation 1: Leaky integrator neural dynamic.
vm(k) = membrane potential at cycle k
vi = synaptic input as calculated in equation 2
f = membrane potential curve as calculated in equation 3
Equation 2: General function describing the postsynaptic potential curve.
a = maximum amplitude (set to 20)
τ = tau (i.e. 8)
t = time since spike (in cycles)
tMax = maximum duration of a PSP (set to 15 cycles)
Equation 3: Membrane potential function
g(vm, 0) = see equation 4
vmRest = membrane potential rest value (set as 43)
vmThreshold = threshold value (set as 65)
Equation 4: Membrane potential output
vec = [4, 11, 18, 23, 28, 32, 36, 42, 43, 44, 45, 47, 50, 53, 58,
65, 100]
Ascending phase to reach threshold = exp(0.8 +
0.3*t) + 40 for eacht from 0 to 8
Ascending phase from post action potential to rest =
log10(0.9 + 0.2*t)*100 for each t from 1 to 7
Ation potential = 100
Equation 5: General STDP function.
Δw = synaptic weight change
αtpost−tpre = 1 or −1, depending on the sign of tpost−tpre
π = time constant
b = bias factor (1.0 for + timing, 1.0 for - timing)
STDP coefficients for Δ w:
Effect duration = 24,000 cycles
Max. synaptic change in one paired spike = 100%
Max. synaptic change = 300%
Max. STDP time window = 100 cycles
2.4. Physical Environment
A physical simulation was done to better evaluate the ability of the SNN to operate under suboptimal timings and conditions. In this environment, the SNN model was embedded in a Khepera IV robot (https://www.k-team.com/khepera-iv), with two modifications. First, instead of using a vibration sensor, the reward was given to the robot using the back infrared. Also, since the robot contains only three programmable LEDs, different colors were used to explicitly referred the four possible directions. These minor changes do not affect the functional aspect of the SNN.
3. Results
Figure 6 shows the dynamic of few main neural components that reflects the learning process of motion direction from an operant conditioning procedure. At the top, small images represent the displayed lines including their orientation and direction. Each of them is repeated several times. At the beginning, when the robot detects a visual stimulus motion, it randomly activates a LED. This output is represented in graphics A, C, E, G from triggering one of the four possible Decision neurons (GoDown, GoUp, GoLeft, GoRight). If the decision corresponds to the good motion direction, then an external supervised reward (graphic I) is sent to the robot. This reinforces the associated Predictor neuron. The pairing of the pre/post spikes and the STDP learning process (graphics J–M) results in an increase of synaptic weights along the operant conditioning procedure. In this simulation, STDP parameters are tuned to trigger Force neurons (Graphics B, D, F, H) with only three correct associations. In the graphics J–M, three steps are shown indicating the learning process. Learning curves are determined from two factors, the preset synaptic weight and the learning incremental step. This rapid learning was done to reduce the number of trials in the experiment considering the four possible directions as well as four possible output responses. The result of the learned association consists in overriding the random decision with Force neurons to trigger the proper action.
Figure 6. Results obtained from the virtual experiments. It represents the operant conditioning learning process that associates a motion direction of a visual stimulus with an external supervised reward. Specifically, after a randomized decision (graphics A,C,E,G), if a reward is given (graphic I), the associated predictor is allowed to spike, followed by a forced action (graphics B,D,F,H) from the STDP learning rule (graphics J–M).
In the experiments, the effect of orthogonal orientation of a visual stimulus in relation to its motion can be observed. In Figure 7, two scenarios are shown: a vertical bar moving horizontally, and a vertical bar moving vertically. In the upper right part, the graphic highlights the detection of the stimulus at a precise moment in the experiment and the Force GoRight neuron spikes accordingly after the learning process. The absolute timing difference is 14 cycles in this orthogonal orientation/motion trial. In the bottom part of the figure, the graphic caught the vertical stimulus moving in the vertical axis from precise spiking and the Force GoUp trigger also spikes after but with a longer period of onset with 23 cycles of difference. This motor response gained is identical in the opposite scenario, a horizontal bar moving vertically giving faster onset response than a horizontal bar moving horizontally.
Figure 7. Effect of merging the orientation feature of a visual stimulus and its motion direction feature. The onset of the motor response arrives earlier when the orientation is orthogonal to its motion. Values in the figure refer to algorithm cycles.
Under the Khepera IV robot, similar results were obtained, even with less precise timing of events. In Figure 8, graphics H and J show that the absolute timing difference between detecting the movement and opening a LED is around 28 cycles when an horizontal line was moving horizontally. When showing a vertical line moving horizontally, the timing difference is reduced to 20 cycles, hence having a response time 30% faster. This behavioral change and its concrete outcome is dependant of the robotic scenario.
Figure 8. Results from the real experiment (Khepera IV robot). As seen at cycles 858 and 1,253, the robot is able to react after the capture of the stimulus. This result shows an additive behavioral response from integrating the motion and the orthogonal orientation of the visual stimulus. In this case, a fasten responses of approximately 30% can be obtained.
4. Discussion
The proposed SNN architecture sustains basic visual orientation and direction selective processes. Integration of these two stimuli features in dedicated neurons was shown in the Drosophila model, sharpening the direction neuronal responses (Fisher et al., 2015). Moreover, a preference association was found when the orientation of the stimulus was orthogonal to its motion. These phenomena were successfully simulated in the proposed SNN model using a precise design of synaptic connections to reproduce the functional outcome at a robotic behavioral level. This, in the neural-robotic domain, suggests that merging two or more stimulus features could potentially modulate the behavioral response, sharpening or reducing it, and it is not restricted to vision only.
As a possible alternative to this current SNN model, the accumulation of sensory inputs across spatio-temporal changes in the movement of the visual stimulus could boost visual signal. In that sense, a vertical stimulus that moves from left to right direction scans a larger portion of the retina instead of an horizontal stimulus. If more sensory neurons are triggered, these extra inputs may also enhance or sharpen the response of the direction cells.
In this experiment, the robot displayed a faster motor output from simultaneously integrating the orientation and direction information of a visual stimulus. Other form of behavioral enhancements could certainly be drawn. For example, benefits could be anticipated from better accuracy, stronger intensity or a faster motor response of organisms; from a barely noticeable gain to a major survivability impact. As such, this paper represents a first step model tested in a static robotic context, but more realistic and dynamical scenarios still remain to be studied. Thus, the present study is limited in terms of motor behavioral complexity, though we believe that the core of the SNN would not change by any addition in the output, since the supervised reward is based on any manifestation of an appropriate response. Also, using visual stimuli with a dynamical robot often requires scaling and focus strategies that were beyond the scope of this article, but may be considered in future work.
Currently, the proposed SNN model contains highly designed connections which reflects the complexity and diversity of biological models (Briggman et al., 2011; Masland, 2012; Kim et al., 2014; Wernet et al., 2014; Demb and Singer, 2015; Fitzgerald and Clark, 2015; Ding et al., 2016; Serbe et al., 2016; Vlasits et al., 2016), but other computational SNN model could be elaborated to obtain more similarity of biological models. In this perspective, instead of using built-in synaptic connections that respond to pure black or white tones, artificial ganglion's cells could be integrated to mimic on-off-center receptive fields and dark/light edges motion (Joesch et al., 2010; Borst and Euler, 2011; Meier et al., 2014; Takemura et al., 2017). Another bio-inspired approach to integrate orientation could be to introduce a suppressive mechanism using inhibitory connections to enhance the direction neural response. Thus, comparative experiments between the biological, computational and robotic model still need to be explored.
In the present SNN model, only two different black bar orientations were used to simplify the process. Also, the simulations were done with a defined constant speed of moving stimuli. Expanding the current model to cope with all orientations is a matter of scaling units and synapses, but would not alter outcome since only two sensory neurons on two axis are needed to obtain the orientation feature. Affording all motion dynamics of stimuli (Li et al., 2017) is perhaps more complex. This remains to be studied, given that in the MD model, the SNN computational method and the STDP learning rules are intrinsically sensible to temporal aspects. In this perspective, variation in the timing of the reinforcement and its schedule as well as extending the STDP period limitation remain to be studied. Another interesting alternative would be to use the same amount of units present in the current architecture, but allow them to respond with a differential firing rate to changes in stimulus orientation, similarly observed in biological networks.
5. Conclusion
Following the recent evidence in vision neuroscience, this work focused on the effect of merging visual orientation and direction processes in a MD computational robotic model. The model was simulated with an SNN method and implemented in a robotic learning context to validate the results at the behavioral level. Specifically, the SNN learned the association between a particular action and a motion visual stimulus from rewards. Both the virtual and physical world experiments succeeded in showing an acceleration of the motor response onset when the visual stimulus orientation is orthogonal to its motion.
Author Contributions
AC did the most of review of literature. AC and FT elaborated the design of the study, the neural architecture and made the virtual simulation. They did most of the data analysis and article redaction. MR contributed in the physical simulation and he reviewed the article as well as providing help in the development of the neural architecture. NB and SC critically enhanced the article (in its structure and with ideas and directions).
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnbot.2018.00075/full#supplementary-material
Table S1. Neural network parameter values.
Table S2. Synaptic parameter values.
References
Adams, S. V., and Harris, C. M. (2015). A computational model of innate directional selectivity refined by visual experience. Sci. Reports 5:12553. doi: 10.1038/srep12553
Barlow, H. B., and Levick, W. R. (1965). The mechanism of directionally selective units in rabbit's retina. J. Physiol. 178, 477–504. doi: 10.1113/jphysiol.1965.sp007638
Behnia, R., Clark, D. A., Carter, A. G., Clandinin, T. R., and Desplan, C. (2014). Processing properties of on and off pathways for drosophila motion detection. Nature 512:427. doi: 10.1038/nature13427
Berberian, N., Ross, M., Chartier, S., and Thivierge, J.-P. (2017). “Synergy between short-term and long-term plasticity explains direction-selectivity in visual cortex,” in 2017 IEEE Symposium Series on Computational Intelligence (SSCI) (Honolulu, HI: IEEE), 1–8.
Bi, G., and Poo, M. (1998). Activity-induced synaptic modifications in hippocampal culture: dependence on spike timing, synaptic strength and cell type. J. Neurosci. 18, 10464–10472. doi: 10.1523/JNEUROSCI.18-24-10464.1998
Borg-Graham, L. J. (2001). The computation of directional selectivity in the retina occurs presynaptic to the ganglion cell. Nature Neurosci. 4:176. doi: 10.1038/84007
Borst, A., and Euler, T. (2011). Seeing things in motion: models, circuits, and mechanisms. Neuron 71, 974–994. doi: 10.1016/j.neuron.2011.08.031
Briggman, K. L., Helmstaedter, M., and Denk, W. (2011). Wiring specificity in the direction-selectivity circuit of the retina. Nature 471:183. doi: 10.1038/nature09818
Clark, D. A., and Demb, J. B. (2016). Parallel computations in insect and mammalian visual motion processing. Curr. Biol. 26, R1062–R1072. doi: 10.1016/j.cub.2016.08.003
Clifford, C. W., and Ibbotson, M. (2002). Fundamental mechanisms of visual motion detection: models, cells and functions. Progr. Neurobiol. 68, 409–437. doi: 10.1016/S0301-0082(02)00154-5
Cyr, A., Boukadoum, M., and Poirier, P. (2009). Ai-simcog: a simulator for spiking neurons and multiple animats' behaviours. Neural Comput. Applic. 18, 431–446. doi: 10.1007/s00521-009-0254-2
Cyr, A., Boukadoum, M., and Thériault, F. (2014). Operant conditioning: a minimal components requirement in artificial spiking neurons designed for bio-inspired robot's controller. Front. Neurorobot. 8:21. doi: 10.3389/fnbot.2014.00021
Demb, J. B., and Singer, J. H. (2015). Functional circuitry of the retina. Ann. Rev. Vis. Sci. 1, 263–289. doi: 10.1146/annurev-vision-082114-035334
Ding, H., Smith, R. G., Poleg-Polsky, A., Diamond, J. S., and Briggman, K. L. (2016). Species-specific wiring for direction selectivity in the mammalian retina. Nature 535, 105–110. doi: 10.1038/nature18609
Eichner, H., Joesch, M., Schnell, B., Reiff, D. F., and Borst, A. (2011). Internal structure of the fly elementary motion detector. Neuron 70, 1155–1164. doi: 10.1016/j.neuron.2011.03.028
Elstrott, J., and Feller, M. B. (2009). Vision and the establishment of direction-selectivity: a tale of two circuits. Curr. Opin. Neurobiol. 19, 293–297. doi: 10.1016/j.conb.2009.03.004
Feldman, D. E. (2012). The spike-timing dependence of plasticity. Neuron 75, 556–571. doi: 10.1016/j.neuron.2012.08.001
Fisher, Y. E., Silies, M., and Clandinin, T. R. (2015). Orientation selectivity sharpens motion detection in drosophila. Neuron 88, 390–402. doi: 10.1016/j.neuron.2015.09.033
Fitzgerald, J. E., and Clark, D. A. (2015). Nonlinear circuits for naturalistic visual motion estimation. Elife, 4:e09123. doi: 10.7554/eLife.09123
Franceschini, N., Pichon, J.-M., and Blanes, C. (1992). From insect vision to robot vision. Phil. Trans. R. Soc. Lond. B 337, 283–294. doi: 10.1098/rstb.1992.0106
Frémaux, N., Sprekeler, H., and Gerstner, W. (2010). Functional requirements for reward-modulated spike-timing-dependent plasticity. J. Neurosci. 30, 13326–13337. doi: 10.1523/JNEUROSCI.6249-09.2010
Fried, S. I., Münch, T. A., and Werblin, F. S. (2002). Mechanisms and circuitry underlying directional selectivity in the retina. Nature 420:411. doi: 10.1038/nature01179
Gilbert, C. (2013). Brain connectivity: revealing the fly visual motion circuit. Curr. Biol. 23, R851–R853. doi: 10.1016/j.cub.2013.08.018
Haag, J., Arenz, A., Serbe, E., Gabbiani, F., and Borst, A. (2016). Complementary mechanisms create direction selectivity in the fly. Elife 5:e17421. doi: 10.7554/eLife.17421
Haag, J., Mishra, A., and Borst, A. (2017). A common directional tuning mechanism of drosophila motion-sensing neurons in the on and in the off pathway. Elife 6:e29044. doi: 10.7554/eLife.29044.
Hassenstein, V., and Reichardt, W. (1956). System theoretical analysis of time, sequence and sign analysis of the motion perception of the snout-beetle chlorophanus. Z Naturforsch B 11, 513–524. doi: 10.1515/znb-1956-9-1004
Hubel, D. H., and Wiesel, T. N. (1959). Receptive fields of single neurones in the cat's striate cortex. J. Physiol. 148, 574–591. doi: 10.1113/jphysiol.1959.sp006308
Huber, S. A., and Bülthoff, H. H. (1998). Simulation and robot implementation of visual orientation behaviors of flies. From Anim. Animats 5, 77–85.
Joesch, M., Schnell, B., Raghu, S. V., Reiff, D. F., and Borst, A. (2010). On and off pathways in drosophila motion vision. Nature 468:300. doi: 10.1038/nature09545
Kerr, D., McGinnity, T. M., Coleman, S., and Clogenson, M. (2015). A biologically inspired spiking model of visual processing for image feature detection. Neurocomputing 158, 268–280. doi: 10.1016/j.neucom.2015.01.011
Kim, J. S., Greene, M. J., Zlateski, A., Lee, K., Richardson, M., Turaga, S. C., et al. (2014). Space–time wiring specificity supports direction selectivity in the retina. Nature 509:331. doi: 10.1038/nature13240
Krichmar, J. L. (2018). Neurorobotics'a thriving community and a promising pathway toward intelligent cognitive robots. Front. Neurorobot. 12:42. doi: 10.3389/fnbot.2018.00042
Kuśmierz, Ł., Isomura, T., and Toyoizumi, T. (2017). Learning with three factors: modulating hebbian plasticity with errors. Curr. Opin. Neurobiol. 46, 170–177. doi: 10.1016/j.conb.2017.08.020
Leong, J. C., Esch, J. J., Poole, B., Ganguli, S., and Clandinin, T. R. (2016). Direction selectivity in drosophila emerges from preferred-direction enhancement and null-direction suppression. J. Neurosci. 36, 8078–8092. doi: 10.1523/JNEUROSCI.1272-16.2016
Li, J., Lindemann, J. P., and Egelhaaf, M. (2017). Local motion adaptation enhances the representation of spatial structure at emd arrays. PLoS Comput. Biol. 13:e1005919. doi: 10.1371/journal.pcbi.1005919
Li, Y., Van Hooser, S. D., Mazurek, M., White, L. E., and Fitzpatrick, D. (2008). Experience with moving visual stimuli drives the early development of cortical direction selectivity. Nature 456:952. doi: 10.1038/nature07417
Li, Y. T., Liu, B. H., Chou, X. L., Zhang, L. I., and Tao, H. W. (2014). Strengthening of direction selectivity by broadly tuned and spatiotemporally slightly offset inhibition in mouse visual cortex. Cereb. Cortex 25, 2466–2477. doi: 10.1093/cercor/bhu049
Maass, W. (1997). Networks of spiking neurons: the third generation of neural network models. Neural Netw. 10, 1659–1671. doi: 10.1016/S0893-6080(97)00011-7
Maisak, M. S., Haag, J., Ammer, G., Serbe, E., Meier, M., Leonhardt, A., et al. (2013). A directional tuning map of drosophila elementary motion detectors. Nature 500:212. doi: 10.1038/nature12320
Masland, R. H. (2012). The tasks of amacrine cells. Vis. Neurosci. 29, 3–9. doi: 10.1017/S0952523811000344
Mauss, A. S., Pankova, K., Arenz, A., Nern, A., Rubin, G. M., and Borst, A. (2015). Neural circuit to integrate opposing motions in the visual field. Cell 162, 351–362. doi: 10.1016/j.cell.2015.06.035
Mazurek, M., Kager, M., and Van Hooser, S. D. (2014). Robust quantification of orientation selectivity and direction selectivity. Front. Neural Circ. 8:92. doi: 10.3389/fncir.2014.00092
Meier, M., Serbe, E., Maisak, M. S., Haag, J., Dickson, B. J., and Borst, A. (2014). Neural circuit components of the drosophila off motion vision pathway. Curr. Biol. 24, 385–392. doi: 10.1016/j.cub.2014.01.006
Mu, Y., and Poo, M. M. (2006). Spike timing-dependent ltp/ltd mediates visual experience-dependent plasticity in a developing retinotectal system. Neuron 50, 115–125. doi: 10.1016/j.neuron.2006.03.009
Nath, A., and Schwartz, G. W. (2016). Cardinal orientation selectivity is represented by two distinct ganglion cell types in mouse retina. J. Neurosci. 36, 3208–3221. doi: 10.1523/JNEUROSCI.4554-15.2016
Paulk, A., Millard, S. S., and van Swinderen, B. (2013). Vision in drosophila: seeing the world through a model's eyes. Ann. Rev. Entomol. 58, 313–332. doi: 10.1146/annurev-ento-120811-153715
Pezzulo, G., Barsalou, L. W., Cangelosi, A., Fischer, M. H., McRae, K., and Spivey, M. J. (2011). The mechanics of embodiment: a dialog on embodiment and computational modeling. Front. Psychol. 2:5. doi: 10.3389/fpsyg.2011.00005
Salay, L. D., and Huberman, A. D. (2015). When visual circuits collide: motion processing in the brain. Cell 162, 241–243. doi: 10.1016/j.cell.2015.06.051
Serbe, E., Meier, M., Leonhardt, A., and Borst, A. (2016). Comprehensive characterization of the major presynaptic elements to the drosophila off motion detector. Neuron 89, 829–841. doi: 10.1016/j.neuron.2016.01.006
Shinomiya, K., Karuppudurai, T., Lin, T. Y., Lu, Z., Lee, C. H., and Meinertzhagen, I. A. (2014). Candidate neural substrates for off-edge motion detection in drosophila. Curr. Biol. 24, 1062–1070. doi: 10.1016/j.cub.2014.03.051
Shon, A. P., Rao, R. P., and Sejnowski, T. J. (2004). Motion detection and prediction through spike-timing dependent plasticity. Netw. Comput. Neural Syst. 15, 179–198. doi: 10.1088/0954-898X_15_3_002
Strother, J. A., Wu, S. T., Wong, A. M., Nern, A., Rogers, E. M., Le, J. Q., et al. (2017). The emergence of directional selectivity in the visual motion pathway of drosophila. Neuron 94, 168–182. doi: 10.1016/j.neuron.2017.03.010
Takemura, S.-Y., Nern, A., Chklovskii, D. B., Scheffer, L. K., Rubin, G. M., and Meinertzhagen, I. A. (2017). The comprehensive connectome of a neural substrate for ‘on’motion detection in drosophila. Elife 6:e24394. doi: 10.7554/eLife.24394
Takemura, S. Y., Bharioke, A., Lu, Z., Nern, A., Vitaladevuni, S., Rivlin, P. K., et al. (2013). A visual motion detection circuit suggested by drosophila connectomics. Nature 500:175. doi: 10.1038/nature12450
Van Hooser, S. D., Li, Y., Christensson, M., Smith, G. B., White, L. E., and Fitzpatrick, D. (2012). Initial neighborhood biases and the quality of motion stimulation jointly influence the rapid emergence of direction preference in visual cortex. J. Neurosci. 32, 7258–7266. doi: 10.1523/JNEUROSCI.0230-12.2012
Vlasits, A. L., Morrie, R. D., Tran-Van-Minh, A., Bleckert, A., Gainer, C. F., DiGregorio, D. A., et al. (2016). A role for synaptic input distribution in a dendritic computation of motion direction in the retina. Neuron 89, 1317–1330. doi: 10.1016/j.neuron.2016.02.020
Webb, B. (2000). What does robotics offer animal behaviour? Animal Behav. 60, 545–558. doi: 10.1006/anbe.2000.1514
Wenisch, O. G., Noll, J., and Hemmen, J. L. (2005). Spontaneously emerging direction selectivity maps in visual cortex through stdp. Biol. Cybernet. 93, 239–247. doi: 10.1007/s00422-005-0006-z
Wernet, M. F., Huberman, A. D., and Desplan, C. (2014). So many pieces, one puzzle: cell type specification and visual circuitry in flies and mice. Genes Develop. 28, 2565–2584. doi: 10.1101/gad.248245.114
Keywords: spiking neurons, vision, direction selectivity, orientation selectivity, motion detection, artificial intelligence, robot
Citation: Cyr A, Thériault F, Ross M, Berberian N and Chartier S (2018) Spiking Neurons Integrating Visual Stimuli Orientation and Direction Selectivity in a Robotic Context. Front. Neurorobot. 12:75. doi: 10.3389/fnbot.2018.00075
Received: 20 June 2018; Accepted: 31 October 2018;
Published: 20 November 2018.
Edited by:
Alex Pitti, Université de Cergy-Pontoise, FranceReviewed by:
Zhenshan Bing, Technische Universität München, GermanyCristian Jimenez Romero, The Open University, United Kingdom
Copyright © 2018 Cyr, Thériault, Ross, Berberian and Chartier. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: André Cyr, acyr2@uottawa.ca