- 1Center for Large-Scale Intelligent Optimization and Networks, Department of Mathematics, University of Memphis, Memphis, TN, United States
- 2Center for the Future Mind, Florida Atlantic University, Boca Raton, FL, United States
- 3Society for MindBrain Sciences, San Diego, CA, United States
- 4MedNeuro, Inc., New York, NY, United States
Spatio-temporal brain activity monitored by EEG recordings in humans and other mammals has identified beta/gamma oscillations (20–80 Hz), which are self-organized into spatio-temporal structures recurring at theta/alpha rates (4–12 Hz). These structures have statistically significant correlations with sensory stimuli and reinforcement contingencies perceived by the subject. The repeated collapse of self-organized structures at theta/alpha rates generates laterally propagating phase gradients (phase cones), ignited at some specific location of the cortical sheet. Phase cones have been interpreted as neural signatures of transient perceptual experiences according to the cinematic theory of brain dynamics. The rapid expansion of essentially isotropic phase cones is consistent with the propagation of perceptual broadcasts postulated by Global Workspace Theory (GWT). What is the evolutionary advantage of brains operating with repeatedly collapsing dynamics? This question is answered using thermodynamic concepts. According to neuropercolation theory, waking brains are described as non-equilibrium thermodynamic systems operating at the edge of criticality, undergoing repeated phase transitions. This work analyzes the role of long-range axonal connections and metabolic processes in the regulation of critical brain dynamics. Historically, the near 10 Hz domain has been associated with conscious sensory integration, cortical “ignitions” linked to conscious visual perception, and conscious experiences. We can therefore combine a very large body of experimental evidence and theory, including graph theory, neuropercolation, and GWT. This cortical operating style may optimize a tradeoff between rapid adaptation to novelty vs. stable and widespread self-organization, therefore resulting in significant Darwinian benefits.
1. Introduction
1.1. Computers, Brains, and Energy
We tend to think of the field of computers and informatics as a major event in the history of ideas, and that is broadly correct. But the mathematics of computation can be traced back to ideas propounded by philosophers and linguists at least a thousand years ago. Western and Asian traditions are often traced to the first millennium BCE; certainly the readable scripts of that time seem to reveal ideas and observations that are remarkably “modern.” History is itself a massively parallel distributed network of events over many centuries. It was not until the invention of digital computers about 80 years ago that systematic studies became feasible to explore the possibility of developing man-made intelligent machines (Turing and Haugeland, 1950; Von Neumann, 1958), which have the potential of demonstrating problem-solving performance comparable to humans. Computer technology demonstrated exponential growth for over half a century. Computers support all aspects of our life. Indispensable and pervasive, they lift billions of people out of poverty worldwide and help them to benefit from technological progress in a modern, interconnected society. The dominant approaches in these applications use Neural Networks (NNs) (Barto et al., 1983; Bishop, 1995; Miller et al., 1995) and Deep Learning (DL), and produce cutting-edge AI with often super-human performance (LeCun et al., 2015; Mnih et al., 2015; Schmidhuber, 2015). The present development trend of intelligent technologies is unsustainable. DL has very high demand for computational power and it requires huge data resources, raising many questions from engineering, societal, and ethical perspectives (Jordan and Mitchell, 2015; Marcus, 2018; Kozma et al., 2019a). Computer chips reach hard limits, marked by the approaching end of Moore's law, which dominated computer development for over half a century (Waldrop, 2016). Energy considerations are an important part of the challenges. High-performance computers require increasing proportions of the available electrical energy to operate (Amodei et al., 2018). Moreover, it is increasingly complicated to remove the heat dissipated in the densely packed microchip circuitries.
Brains provide us valuable clues regarding efficient use of resources, including energy. The operation of brains is naturally constrained by the available metabolic resources following fundamental laws of thermodynamics. According to the free energy principle, brains optimize metabolic and computational efficiency by reconfiguring themselves while they interact with the environment in the action and perception cycle (Friston et al., 2006; Sengupta et al., 2013). Brains continuously optimize their energy resource allocation, while advanced computing algorithms are mostly agnostic when it comes to power consumption. Arguably, brains are several orders of magnitude more energy-efficient than cutting-edge AI when solving specific machine learning tasks (Amodei et al., 2018; Kozma et al., 2019b; Marković et al., 2020). The efforts to achieve human-level intelligence and machine understanding by scaling up computing using million-core chips are impressive, but alternative approaches may become useful as well. Energy-awareness is a basic manifestation of embodiment, which is crucial for the emergence of intelligence in brains and machines (Dreyfus, 2007), and it provides the key for progress in machine understanding as well (Yufik, 2013, 2019). Neuromorphic technologies have great potential in large-scale computing systems, including spiking neural networks (Furber, 2016; Hazan et al., 2018; Roy et al., 2019), and memristive hardware (Di Ventra et al., 2009; Chua, 2012; Kozma et al., 2012; Stieg et al., 2019). Combining neuromorphic technologies with brain-inspired thermodynamic models of computing has the potential of providing the required breakthrough in machine understanding (Yufik and Friston, 2016; Friston et al., 2020).
1.2. Cognitive Dynamics and Consciousness
It is often thought that the question of consciousness in the waking brain is so difficult and poorly understood that empirical science has nothing to say about it. However, beginning some decades ago, empirical scientists in psychology and neuroscience have published literally thousands of scientific papers, mostly on very specific aspects of conscious perception and cognition.1 Global Workspace Theory (GWT) is one of the prominent modeling approaches (Baars, 1997; Baars and Geld, 2019; Baars et al., 2021). GWT fundamentally proposes that the striking capacity limits of conscious percepts implies very widespread unconscious access to processing resources in the brain. This convergence of two very different theoretical traditions suggests that they are two sides of the same coin.
GWT first emerged around 1980, based on the cognitive architecture tradition in cognitive science, including global workspace architecture (Newell et al., 1972). The cognitive architecture program goes back many decades, when Herbert A. Simon and the Netherlands chess master Adrian De Groot began to carefully study the move-by-move “consciousness reports” of advanced chess players (Simon, 1967; De Groot, 2014). Since the middle of the last century, a number of cognitive architectures have been proposed and partially tested. The book by Newell (1994) can be considered to be a summary of this empirical modeling tradition. At least a dozen cognitive architectures have been proposed in this research practice. They proposed different computer implementations with two shared features: All cognitive architectures had a serial perception and problem-solving component, and in all cases the serial flow of immediately accessible events interacted with a very large long-term memory capacity, which appears to be a non-serial set of knowledge sources. Cognitive architectures also merged with a separate experimental cognitive research tradition, until, by the 1970s and 80s, it began to seem that both lines of research could be understood in a single framework (John and Newell, 1990). The work of Tversky and Kahneman (2011) is another example of this pattern of discoveries, focusing on the empirical phenomenon of automaticity. Newell (1994) discussed this striking convergence of a serial “stream of consciousness” reported by subjects, and a very large, non-serial set of memory domains, which are not in reportable consciousness at any given time; but the massively parallel memory domain is unconscious most of the time during chess playing.
Baars was one of the first cognitive scientists to explicitly use the word “conscious” for the serial component of chess-playing protocols, and “unconscious” for the large set of knowledge sources that players demonstrably use, but which may not become explicit in any single chess move. What Baars called Global Workspace Theory (GWT) in the 1980s combined two streams of scientific study, the cognitive architecture tradition and the field of cognitive psychology (Baars, 1997). That convergence seemed to be surprisingly easy to describe. By 1980 the field of cognitive science began to emerge, and the computational, mathematical, and cognitive-behavioral streams of development turned into a single, extensive field of study. Baars' GWT linked a vast empirical literature to the theoretical concept of consciousness, which could be inferred from the mass of evidence, and which also seemed to reflect the reported experiences of subjects in many tasks.
The distinctive feature of all cognitive architectures, including GWT, can be found in Newell's pioneering formulation. Rather than a passive unconscious long-term memory, with more powerful computers the idea emerged that the parallel component reflects a “society” of specialized knowledge sources that were not conscious by themselves, but which interacted to “post messages” on some shared knowledge domain, called a global workspace. Since that time, computational GWT has seen very widespread use in cognitive and computer science. The mathematics of parallel-interactive computation led to both fundamental and practical insights into human cognition. What seemed puzzling and scattered before 1980 gradually emerged with a greater degree of clarity (Franklin et al., 2012).2
Cognitive Science is now Cognitive Neuroscience, leading to another large set of converging ideas, with more and more brain and behavioral evidence interacting in fruitful ways. In fields like language studies, for example, it became routine to consider the perceptual aspects of a stream of words (like this one) as conscious, in fast-cycling interaction with multiple unconscious knowledge domains. “Society models” gradually merged with the brain sciences, giving rise to contemporary cognitive neuroscience theory. We prefer to think of a “family” of GWT architectures, where Baars' version is perhaps the best known today, but the family has many members that continue to evolve. Essentially empirical, this set of theories may be considered similar enough to be treated as a “family” of global workspace-like approaches, including (Dehaene et al., 1998; Fingelkurts et al., 2010; Edelman et al., 2011; Tononi and Koch, 2015; Kozma and Freeman, 2016; Mashour et al., 2020; Deco et al., 2021). Each approach is distinctive and each is based on a strong body of evidence; but they converge well. Much to our surprise, a very large scientific literature in neurobiology has also converged with all the fields in a remarkable way.
The current paper presents yet another region of convergence between multiple empirical and theoretical streams of development. With direct brain recordings of the electromagnetic activity of single neurons and massive neuronal networks, we may be seeing a convergence between many intellectual traditions. We view brains as large-scale complex networks, and brain dynamics as percolation processes evolving over these networks, with potentially adaptive structures. We introduce several key analysis methods, such as the thermodynamics of wave packets, statistical physics of criticality and phase transitions, cinematic theory of neurodynamics and metastability, and a hypothesis concerning the interpretation of the experimentally observed neurodynamics using the GWT framework. Two main computational results are introduced to illustrate the findings. The first describes the essential role of non-local axonal connections in maintaining a near-critical state of brain oscillations. The second result concerns the role of astrocyte-neural coupling in maintaining neural fields with rapid transitions between states with high and low synchrony, respectively. We conclude the work with discussing the potential implications of these results to lay down the principles of machine understanding.
The rest of the essay addresses the fundamental question: What could be the evolutionary advantage of brains utilizing phase transitions, as compared to possible alternatives with smooth dynamics?
2. Methods
Describing brains as open thermodynamic systems converting noisy sensory inputs and metabolic energy into conscious sensory percepts to explicit understanding of the world.
2.1. Thermodynamics of Wave Packets3
There is a vast literature on experimental investigations of thermodynamics of brains, see, e.g., Abeles and Gerstein (1988); Fuchs et al. (1992); Freeman (2000), and Friston et al. (2006). Freeman K sets provide a theoretical framework for brain models with a hierarchy of increasingly complex structure, dynamics, and function (Freeman, 1975, 1991, 2000; Kozma and Freeman, 2009). Several key aspects are summarized here, using the concept of metastability,4 as described in Kozma and Freeman (2016, 2017).
PROPOSITION 1 (Characterization of wave packets (WPs); Kozma and Freeman, 2016). The action-perception cycle is manifested through the self-organized sequence of metastable, highly synchronized patterns of spatio-temporal amplitude modulated (AM) activity at the beta/gamma carrier frequency (20-80 Hz). These AM patterns emerge and collapse, and as such they form spatio-temporal Wave Packets (WPs). The WPs evolve as follows:
(i) WPs exist for a time window of ~100 ms, corresponding to approx. 10Hz frequency band. They have spatially-localized evolving patterns, therefore they are sometimes called wave packets.
(ii) WPs have statistically significant correlations with sensory stimuli and reinforcement contingencies perceived by the subject.
(iii) WPs are separated in time by brief transitionary periods (10-20ms). During these transitionary periods, the AM patterns collapse and large-scale synchrony diminishes.
(iv) The repeated collapse of WPs points to recurring singularities in mammalian cortical dynamics ignited at a given location of the cortex. Following the selection and activation of a Hebbian cell assembly corresponding to the stimulus, the synchronized activity of neural populations rapidly propagates across the cortical sheet in the form of a phase cone.
(v) The rapid transitions and propagation of phase cones following their ignition cannot be explained by synaptic transmissions only, and it requires the emergence of collective dynamics.
The repeated collapse and emergence of the metastable wave packets defines a quasi-periodic oscillatory energy cycle with the following steps:
PROPOSITION 2 (Energy cycle of wave packets; Kozma and Freeman, 2017). The temporal evolution of Wave Packets is sustained by the corresponding energy cycle, described by thermodynamic processes involving energy and entropy transfer between highly-ordered (liquid) states and disordered (gaseous) states:
(i) The cycle starts with a disordered background state with low amplitude waves. This state has high entropy and in the thermodynamic sense it is analogous to a gaseous state.
(ii) At a certain space-time point, synchrony is ignited in the neural populations in response to a meaningful stimulus and a phase cone starts to grow from an incipient state. The phase cone develops into a highly structured, metastable WP with low entropy oscillating at a narrow beta/gamma frequency band. The emergence of the WP leads to the dissipation of energy in the form of heat, which is removed by the blood stream through the capillaries. This can be viewed as a condensation process to a liquid state.
(iii) The metastable WP continuously erodes with decreasing synchrony between the neuron components, due to the impact of input stimuli and random perturbations. The entropy increases, which corresponds to the thermodynamic process of evaporation.
(iv) At the end of the cycle, the intensity of the neural firing activity drops to a level when the activity patterns are dissolved and the thermodynamics returns to the high-entropy gaseous state.
This section summarized key aspects of experimental findings on EEG recordings in terms of thermodynamic processes. The next sections introduce methods of statistical physics and mathematical theory of graphs and networks to quantitatively characterize these findings.
2.2. Criticality in Brains and Neuropercolation Model
The thermodynamic interpretation of the action-perception cycle outlined above implies that brains operate through repeated transitions between highly-organized, synchronous states and disorganized states with low levels of synchrony. These observations lead to the hypothesis that brains are critical or near-critical systems, which has been proposed by various authors. One prominent approach is based on the concept of self-organized-criticality (SOC) when a high-dimensional complex system organizes itself to a critical point which is an attractor state. SOC demonstrates scale invariance, including power-law behavior with 1/f scaling, where f is the frequency of the events corresponding to the specific problem domains. In the case of neural processes, f could relate, for example, to bursts of spontaneous activity in neural populations, and 1/f shows the number of bursts of the given frequency. SOC has been observed in many disciplines, from earthquakes, to solar flares, sandpiles, etc, and in neural tissues as well (Beggs and Timme, 2012; Shew and Plenz, 2013). SOC is widely used now in the interpretation of brain monitoring data, including the connectome, resting state networks, consciousness, and other areas; see, e.g., Fingelkurts et al. (2013); Tagliazucchi (2017); Nosonovsky and Roy (2020), and Wang et al. (2020). Under certain conditions, deviation from the power-law behavior predicted by SOC are observed in brain dynamics, which justify approaches addressing criticality beyond SOC, e.g., critical integration and soft assemblies (Aguilera and Di Paolo, 2021).
A related approach uses percolation theory to describe criticality of brain operation, by modeling the cortical neuropil (Kozma et al., 2005, 2014; Bollobás et al., 2010; Kozma and Puljic, 2015).
PROPOSITION 3 (Neuropercolation model of criticality and phase transition in brain dynamics; Kozma et al., 2005; Kozma and Puljic, 2015). According to neuropercolation, critical behavior in the cortex is made possible by the filamentous structure of the cortical neuropil, which is the most complex substance in the known universe. Neuropercolation is the generalization of Ising models and lattice cellular automata, and it describes the following aspects of the neuropil:
(i) Presence of rare long axonal connections between neurons, which allow action at distant locations with minimal delay.
(ii) Contribution of astrocytes cells, which have a key role in metabolic processes and in the formation of field effects.
(iii) Incorporation of random noise effects; the model is robust to noise and noise is an important constructive control parameter to tune the system to achieve desired behavior.
(iv) Input-induced and spontaneous phase transitions between states with large-scale synchrony and without synchrony exhibit brief episodes with long-range spatial correlations.
(v) Neuropercolation proposes a constructive algorithm that self-regulates cortical dynamics at criticality following supercritical explosive excursions.
Beyond the theoretical results, neuropercolation has been employed successfully to interpret experiments with Pavlovian conditioning in rabbits (Kozma et al., 2014; Kozma and Puljic, 2015), on entrainment of sensory processing by respiration in rats and human subjects (Heck et al., 2017, 2019), and strategy changes during learning in gerbils (Kozma et al., 2021).
2.3. Intermittent Metastable Brain Oscillations
There is widespread agreement that processing of sensory information in the cortex is associated with complex spatio-temporal patterns of activity (Abeles, 1982). Experimental observations of intermittent brain oscillations with extended metastable periods, interrupted by rapid transients, are widely discussed in the literature (Lehmann et al., 1987; Buzsáki, 1998). This issue is often framed as a choice between opposing views of continuous vs. discrete cognition. Following the wisdom of Kelso's complementarity principle, the likely answer would be that both discrete and continuous aspects are relevant to cognition through the unity of continuity-discreteness (Fingelkurts and Fingelkurts, 2006; Tognoli and Kelso, 2014; Parr and Friston, 2018). Recent reviews by Josipovic (2019), Menétrey et al. (2021), and Lundqvist and Wutz (2021) help to disentangle the arguments.
The hypothesis that perception happens in discrete epochs has been around for decades, and models of brains as dynamical systems with itinerant trajectories over distributed attractor landscapes provided mathematical tools to support the analysis, see, e.g., Babloyantz and Destexhe (1986); Skarda and Freeman (1987); Freeman (2000), and Tsuda (2001). Crick and Koch (2003) described discrete frames as snapshots in visual processing, as well as in consciousness; while Tetko and Villa (2001) provided evidence of cognitive relevance of spatio-temporal neural activity patterns. The sample-and-hold hypothesis expands on the sampling idea and it describes the perceptual and motor processing cycle (Edelman and Moyal, 2017). Spatiotemporal sequences of time-position patterns have been observed in the human brain associated with cognitive tasks (Tal and Abeles, 2018). Recent models describing sequential processing of complex patterns of brain activity are developed in, e.g., Cabessa and Villa (2018); Malagarriga et al. (2019).
EEG data evaluated using Hilbert analysis also display sudden transitions of cognitive relevance (Brennan et al., 2011; Frohlich et al., 2015), while operational architectonics provides a powerful framework for transient synchronization of operational modules underlying mental states (Fingelkurts et al., 2010, 2017). Phase transitions over large-scale brain networks have been applied to describe the switches from one frame to another in the cinematic theory of neurodynamics and cognition (Kozma and Freeman, 2016, 2017). Kozunov et al. (2018) evaluates MEG visual processing data and points to the role of phase transitions and critical phenomena to understand how meaning can emerge from sensory data. The identified cycle length varies depending on the experimental conditions; i.e., it is in the theta/bands in the cinematic theory (Freeman, 2000; Kozma and Freeman, 2017); while Pereira et al. (2017) estimate a very long cycle of consciousness (2 s). The work by Werbos and Davis (2016) is unique by identifying a very precise clock cycle of 153 ms, by analyzing Buzsáki lab data (Fujisawa et al., 2015).
There are various open issues regarding discrete effects in neurodynamics and some questions were raised about their significance in cognition and consciousness. For example, Fekete et al. (2018) states that the involved brain networks cannot produce switching behavior at the rates observed in brain imaging experiments. They lay out a valuable work, but they do admit that their reasoning does not hold for strongly non-linear systems as brains are. Their proposed multi-scale computation near criticality is certainly interesting and it has a lot in common with the edge of criticality described as the result of ontogenetic development in neuropercolation in the past two decades (Kozma et al., 2005). White (2018) does not question the existence of sudden changes observed by Freeman et al. (2006); Brennan et al. (2011), and Kozma and Freeman (2016), rather it misses the established proof that these neurodynamic effects are relevant to conscious perception. Clearly, there is a need for extensive further experiments before confirming or rejecting the central hypothesis on the key role of phase transitions in cognition and consciousness. Some recent experiments lend support to the hypothesis on discontinuities in cognition, such as entrainment of multi-sensory perception by the respiratory cycle (Heck et al., 2017); how breathing shapes memory functions (Heck et al., 2019); the role of state transitions in strategy changes during an aversive learning paradigm and the formation of Hebbian cell assemblies by identifying emergent causal cortical networks (Kozma et al., 2021); and clustering of phase cones during interictal periods over the epileptogenetic brain region (Ramon and Holmes, 2020). Statistical markers of phase transitions show potential use in psychotherapy (Sulis, 2021).
PROPOSITION 4 (Transient processing in perception; Kozma and Freeman, 2017). Phase transitions over large-scale brain networks have been applied to describe the switches from one frame to another in the cinematic theory perception, as follows:
(i) The intermittent emergence and collapse of AM patterns in EEG data is interpreted as the evidence that perceptual information processing happens in discrete steps, aligned with the prominent AM patterns.
(ii) The cinematic theory of perception uses the concept of the frame and the shutter, which follow each other sequentially. There is no exact threshold separating the two phases from each other, rather they transit to each other following the corresponding energy cycle of WP.
(iii) The frames are defined by the dominant AM patterns which are sustained for an extended period of around 100 ms, with significant variation depending on experimental conditions. The frames are selected according to the reinforced contingencies as perceived by the subject. The frame activity is largely synchronous across large cortical areas during the existence of the frame. However, the frame is not a frozen pattern, rather it oscillates at the beta/gamma carrier frequencies.
(iv) The shutter is defined by the relatively short periods (approx. 10 ms) when the AM patterns collapsed and the neural activity is disordered, still not completely random and maintains some trace of the previous dynamics.
The Freeman/Kozma approach has been called cinematic, because the cortical dynamics self organizes into phase plateaus at roughly every ~ 100 ms, followed by a collapse of the phase plateau for about 10 ms. During the brief collapse of synchrony, the cortex is prepared to receive novel perturbations, while the self organized phase synchrony is a time of relative stability and internal processing. This style of functioning plausibly optimizes a balance between receptivity to novelty and stability, pointing to potential evolutionary advantage by the rapid, moment-to-moment adaptivity of the conscious cortex.
Brains are dynamic systems, they can never stop, not even during the relatively quiet periods when frames with metastable amplitude patterns are maintained. Being constrained to a quasi-periodic attractor basin during a frame is just the sign of relative silence, before the explosive impact of the phase transition, which destroys the existing structure and gives rise to the emergence of a new pattern in response to the new sensory input and its meaning to the subject (Freeman, 2000). Dynamical modeling of the brain includes both continuity of the movement along its trajectory, as well as rapid changes as the path leads from one metastable state to another (Tognoli et al., 2018). The switches are not rigid and they have their own rich dynamic structure and a hierarchy with possibly scale-free distribution (Mora-Sánchez et al., 2019). These results show that an integrative approach to identify major features of cognitive dynamics and consciousness is very productive, including the unity of discrete and continuous operating modalities in brains.
2.4. Hypothesis on the Link Between EEG Perceptual Transition and GWT
Phase transitions and criticality in cortical layers may have a profound impact on the nature of consciousness. There have been various attempts to integrate phase transitions with GWT, such as the one by Werner (2013), to model the emergence of multi-level collective behaviors in brain dynamics. Tagliazucchi (2017) describes consciousness as the integration of fragmented, highly differentiated entities into a unified message, and they use percolation model to describe the propagation of conscious access through the brain network medium, with phase transitions when a critical threshold is reached. Josipovic (2019) elaborates the concept of non-dual awareness in the framework of GWT. GWT is hereby linked to perceptual phase transitions (Freeman, 1991, 2000; Kozma and Freeman, 2016).
PROPOSITION 5 (Main Hypothesis on EEG phase transitions as indications of conscious experience Kozma and Freeman, 2016; Baars and Geld, 2019). Phase transitions in the cortex are ignited at a given location of the cortex, according to EEG data. Phase transitions generate laterally propagating phase gradients (phase cones) across the cortical sheet. In the context of GWT, these results are interpreted as follows:
(i) Phase cones are neural signatures of perceptual broadcasts described by GWT.
(ii) The rapid expansion of phase cones, covering large cortical areas within 10-20 ms, are consistent with the propagation of perceptual broadcast postulated by GWT.
(iii) The recurrence time of the cortical phase transitions is about 100 ms, which is consistent with the ~ 100 ms window identified in numerous perceptual and behavioral experiments.
The ~ 100 ms time domain has long been studied in the sensory sciences and proposed as an integration period for conscious cortical information processing (Baars, 1988; Madl et al., 2011; Baars and Geld, 2019). GWT suggests that conscious sensory events are the leading edge of adaptation during waking life. The very fast and highly adaptive role of cortex clearly fits within a Darwinian framework of genetic, epigenetic, and moment-to-moment cortical adaptation (Edelman et al., 2011). Edelman's Neural Darwinism is highly consistent with this approach, and specifies the role of selectionism at multiple time and spatial scales in the brain. Interpreting phase cones as neural manifestations of perceptual broadcasts of GWT is an important step to connect the content of consciousness with the temporal structure of consciousness per se (Menétrey et al., 2021). Next, computational results are introduced to illustrate the hypothesis.
3. Results
3.1. Long-Axonal Connections Facilitate Criticality in the Neuropil
Brain networks analysis has been successful to study anatomical, functional, and effective brain connectivity, using tools of graph theory (Iglesias and Villa, 2007, 2010; Stam and Reijneveld, 2007; Steyn-Ross and Steyn-Ross, 2010; Bullmore and Sporns, 2012; Haimovici et al., 2013). Imamoglu et al. (2012) suggest that frontal and visual brain regions are part of a functional network that supports conscious object recognition by changes in functional connectivity. Zanin et al. (2021) point out that neuroscience of brain networks often emphasizes the extraction of neural connectivity represented by strong links and highly-connected nodes, although weak links can in fact be critical in determining the transition between universality classes. Most of the existing network-based toolsets extract information on the interaction of localized units and nodes (Korhonen et al., 2021). Brains are metastable systems, and their optimal functioning depends upon a delicate metastable balance between local specialized processes and their global integration (Fingelkurts and Fingelkurts, 2010), while minute perturbations and topological changes can lead to significant deviations from the normal operational dynamics (Tozzi et al., 2017), with an impact on synchronization effects in these complex non-linear systems (Brama et al., 2015; Xu et al., 2020). Random graphs and cellular automata models have been developed for cortical dynamics to address the challenges (Balister et al., 2006; Kozma and Puljic, 2015; Ajazi et al., 2019; Turkheimer et al., 2019). Percolation models are especially helpful in the interpretation of experimental findings describing the intermittent emergence of common-mode oscillations in neural cell assemblies (Kozma and Freeman, 2016).
An important theoretical finding describes phase transitions in a graph model of the cortical neuropil with a mix of short and long connections, including long axons (Janson et al., 2019). A random graph is considered over the square grid of size (N+1) × (N+1), and p is the probability describing random long edges, see Equation (1). We assume periodic boundary conditions, for simplicity, thus we have a torus with the short notation . The set of vertices of G consists of all the vertices of . There are two types of edges E, short and long, respectively. Short edges are all the edges from the torus ; i.e., each node has 4 short edges connecting to its 4 direct neighbors. Additionally, we introduce random long edges as follows: for any pair of vertices that are at distance d apart of each other on the lattice, we assign an edge with probability p that depends on the distance:
Here α is a number, e.g., α = 1. An activation process is defined on as follows: Denote by A(t) the set of all active vertices at time t. We say that a vertex v is active at time t if its potential function χv(t) = 1 and inactive if χv(t) = 0. Therefore, A(t) = {v∈V(G) |χ v(t) = 1}. At the start, A(0) consists of all vertices that are active with probability p0. Each vertex may change its potential based on the states of its neighbors as follows:
A vertex can become active if at least k of its neighbors are active. Let ρt be a proportion of active nodes at time t, i.e., then the evolution of ρt can be described in a mean-field approximation, for details, see Janson et al. (2019). A key result has been derived for the existence of phase transition of the activation process over :
PROPOSITION 6 (MAIN THEOREM JKRS219: on phase transitions in the neuropercolation model with short and long connections Janson et al., 2019). For the activation process A(t) over random graph , in the mean-field approximation, there exists a critical probability pc such that for a fixed p, w.h.p.:
1. all vertices will eventually be active if p>pc, while
2. all vertices will eventually be inactive for p<pc.
3. The value of pc is given as the function of k and λ through the solution of some transcendental equations.
The main theorem in Proposition 3.1 rigorously proves the existence of phase transitions in neuropercolation model with long axons; its meaning is illustrated in Figure 1, using numerical evaluation of the precise mathematical formula. In Figure 1, the x-axis shows λ, which scales linearly with the probability of long axons, while the y-axis is the critical probability when the phase transition happens; k indicates the update rule. It is seen that there is a region for small λ values, where the model behaves essentially as a local system. For large λ values, the critical probability diminishes what is expected for a global system without local order. There is a transitionary region when the incremental addition of long connections does matter, as it is expected to be the case in the neuropil. Clearly, this model cannot grasp all the complexity of brain networks, and there are many advancements including inhibitory and excitatory effects, multi-layer architectures with delayed reentrant connections. Still, the introduced effect is very robust and it is a unique property of the neuropil with a mix of short and long projections. Brains can benefit from the transitionary region for tuning their behavior between local fragmentation and overall global dominance, using adaptation and learning effects.
Figure 1. Illustration of the effect of the long edges λ on the critical probability pc; parameter k specifies the type of the update rule; based on Janson et al. (2019).
3.2. Metabolic Processing in the Neuropil Controls Transitions Between States With High and Low Synchrony Based on Hysteresis Dynamics
Following fundamental studies on the brain energy budget (Raichle and Gusnard, 2002; Magistretti, 2006), there are extensive integrative models on metabolic coupling in the neuron-glia ensemble with capillaries (Cloutier et al., 2009; Belanger et al., 2011; Jolivet et al., 2015), and the role of metabolic constraints on spiking activity (Teixeira and Murray, 2015; Zhu et al., 2018; Qian et al., 2019). The models typically use multi-compartmental neuron models, but some simplified still realistic spiking neuron models are popular as well, e.g., Izhikevich (2003).
To describe the emergence of synchronized collective cortical oscillations driven by metabolic constraints, the capillary astrocyte-neuron model (CAN) is introduced, which couples spiking and metabolic processes (Kozma et al., 2018, 2019b). The simplest CAN model has two metabolic variables: g(t) and m(t), where g(t) describes the available glycogen stored in the astrocyte, m(t) models the available ATP in the neuron's mitochondria. Izhikevich (2003) model is used for the spiking neurons, with variables u(t) and v(t), which are the dimensionless membrane potential and the membrane recovery variable, respectively. The following differential equations describe the rate of change of the variables:
Here Φ1(u, v) is membrane potential fitting function; Φ2(u, v, m) describes the recovery variable dynamics, modulated by the available ATP via m(t). Ψ1(g, m) and Ψ2(g, m) describe the attenuation of g(t) and m(t), respectively. I(t) describes the influence of synaptic currents. The integral term in Equation (3b) describes the cumulative effect of spiking on the glutamate concentration in the synaptic cleft, over time period of τ, and κ is a scaling parameter. Izhikevich's model has a sensitivity parameter b regulating spike production inside term Φ2(u, v, m). A nominal value of b = 0.2 assures regular spiking (Izhikevich, 2003). To close the feedback loop between the metabolic and neural parts of the model, b is modulated by m(t) as follows: b+(m(t)) = [ωb+βm(t)]. Here ω is a scaling parameter in the range [0.75, 1.25], directly impacting the spiking density. The 2nd term reflects the contribution of m(t), where β is a control parameter in the range [0, 0.5]. For β = 0, metabolic processes do not impact spiking, while increasing β leads to increasing frequency of spiking. Full elaboration of the model is given in Kozma et al. (2018).
PROPOSITION 7 (Metabolic control of synchrony transitions in neural populations based on hysteresis dynamics Kozma et al., 2019b). The capillary astrocyte-neuron model (CAN) described by Equations (3a)–(3b) demonstrates transitions between synchronized collective cortical oscillations and the absence of synchrony, as illustrated in Figure 2. The process has the following properties:
(i) The amount of available energy modulates the oscillation frequency of neural populations.
(ii) There is a hysteresis effect as the result of cusp bifurcation in the CAN model. The space defined by the forward gain from neural to metabolic subsystems, and the feedback gain from metabolic to neural system has a bifurcation point leading to the split of a stable equilibrium to two stable and one unstable equilibrium.
(iii) The parameters corresponding to the bifurcated states produce self-sustained oscillations between high and low-synchrony states.
(iv) The results reproduce experimentally observed collective neural dynamics in the form of large-scale cortical phase transitions.
Figure 2. Metabolic-neural model; (A) hierarchy of structures from cellular, mini-column, and mesoscopic population levels; (B) hysteresis effect on the transition between states with low (blue/turqoise) and high synchrony (purple); (C) examples of spiking raster plots for low and high synchrony.
It is important to point out that the metabolic processes are required to produce the hysteresis effect and the desired transitions between states with high and low synchrony. Populations of pure spiking neurons without metabolic components are not sufficient to reproduce the experimentally observed transition effects, as it has been remarked by Deschle et al. (2021).
4. Discussion: Human Understanding and Machine Understanding
This work explores what the evolutionary advantage may be of brains utilizing repeated phase transitions at theta/alpha rates, as compared to possible alternatives with smooth dynamics. There are a striking number of regularities that are found over and over again at around 10 Hz. Some of these emerge from the mathematics of neurodynamics described here, and some of them emerge from a century of research in conscious sensory perception. We can call this pattern of convergence the “magic number” near-10 Hz (~100 ms). The flow of conscious events is serial, while unconscious knowledge domains constantly interact with the conscious stream, as EEG data and psychological evidence show over and over again. The ~100 ms Temporal Window has been studied since the 1800s because it keeps on emerging in psychological evidence. In psychology experiments, it is always linked to highly reliable reports of conscious sensory experiences. As we described here, the magic Temporal Window may be explained by the cinematic view of neurodynamics and phase transitions in the cortex. Because the ~100 ms Temporal Window is so common, and clearly appears in association with conscious experiences, this possible link is intriguing.
Some of the empirical phenomena that clearly dwell in the magic Temporal Window:
1. Two sensory inputs fuse into single conscious gestalts if they occur within a ~100 ms time window. This is an enormously general phenomenon in sensory psychophysics, both within and between the major sensory modalities.
2. The motor domain shows a similar Temporal Window. Simple reaction time hovers around ~100 ms. In continuous tasks, the relationship between sensory output and motor outputs works best within the Temporal Window.
3. The ~ 100 ms sensory integration window is found in all the major senses, and also in cross sensory tasks. We should reemphasize the extraordinary generality of this phenomenon across vision, audition, and touch perception in humans and other species. What has been missing is an explanation.
The mathematical properties of cortex, as found by Kozma and Freeman (2016), may therefore explain unconscious-conscious events as they have long been observed in psychology experiments. Phase transitions create the basis for rapid and robust responses to environmental challenges, which provided our ancestors with evolutionary advantage compared to the competitors. As an illustration of these abstract considerations, we can easily imagine a wild rabbit needing to interpret a raptor attack in order to escape it. Under the best possible scenario, it may take ~ 100 ms or more for the rabbit to perceive the attack, and even longer to combine these events with short term and long term memory (Madl et al., 2011). Based on the evolutionary process, this specific time window is sufficient to develop a successful escape strategy while optimizing the finite resources of its brain and body, considering the natural environment, in which the rabbit's ancestors strived for millions of years.
In this work, we outlined a framework for interpreting and modeling brain measurements demonstrating metastable dynamics with rapid transients, which can be used to develop computational devices incorporating brain-inspired principles. Such novel devices have the potential to develop machines which understand the world around us in a way as we humans do, and help us with the challenges we face.
Data Availability Statement
The data analyzed in this study is subject to the following licenses/restrictions: data are available upon request from the authors. Requests to access these datasets should be directed to rkozma@memphis.edu.
Author Contributions
RK coordinated the work with a focus on mathematical and computational modeling and outlined the initial draft manuscript that was modified, and approved by all authors in its final version. BB and NG focused on consciousness and cognitive areas. All authors contributed to the conceptual formulation of this research.
Conflict of Interest
NG founded MedNeuro, Inc.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
This work was based on the ideas of the paper presented at AFOSR AFRL/RY Workshop on Understanding in the Human and the Machine, 24-26 August 2020, Washington DC. The support of the organizers is greatly appreciated.
Footnotes
1. ^In the scientific literature, over 27 thousand relevant abstracts can be found at this link: https://pubmed.ncbi.nlm.nih.gov/?term=conscious+brain.
2. ^Stan Franklin's research group really pushed the world of computer science and AI toward these cognitive architectures and moved the needle into this direction, in a 20+ year strong research program at the University of Memphis.
3. ^We take no position on philosophical questions that are often raised in connection with conscious perception, the brain, and the relevance of quantum mechanics and quantum fields. Global workspace theory and neuropercolation should be considered on their respective merits. Both theories have been fruitful, and here we consider how they may interact in interpreting experimental results.
4. ^A state of a dynamical system is called metastable, if it is not stable, but it maintains its integrity for an extended period of time, which is meaningful for the analyzed problem. In other words, a metastable state is unstable over very long time scales, but it can be considered stable for shorter, still extended time periods. It is of special interest to study metastability in spatially extended systems, when metastability in time is manifested in the emergence of well-defined spatial patterns for some time periods. Transient dynamics from one metastable state to another metastable state has been extensively studied in various mathematical and physical systems. In the present essay, we refer to metastability appearing in the form of intermittent synchronization of cortical activity.
References
Abeles, M., and Gerstein, G. L. (1988). Detecting spatiotemporal firing patterns among simultaneously recorded single neurons. J. Neurophysiol. 60, 909–924. doi: 10.1152/jn.1988.60.3.909
Aguilera, M., and Di Paolo, E. A. (2021). Critical integration in neural and cognitive systems: beyond power-law scaling as the hallmark of soft-assembly. Neurosci. Biobehav. Rev. 123, 230–237. doi: 10.1016/j.neubiorev.2021.01.009
Ajazi, F., Chavez-Demoulin, V., and Turova, T. (2019). Networks of random trees as a model of neuronal connectivity. J. Math. Biol. 79, 1639–1663. doi: 10.1007/s00285-019-01406-8
Amodei, D., Hernandez, D., Sastry, G., Clark, J., Brockman, G., and Sutskever, I. (2018). Ai and Compute. Heruntergeladen von. Available online at: https://blog.openai.com/aiand-compute
Baars, B. J.. (1997). In the theatre of consciousness. Global workspace theory, a rigorous scientific theory of consciousness. J. Conscious. Stud. 4, 292–309. doi: 10.1093/acprof:oso/9780195102659.001.1
Baars, B. J., and Geld, N. (2019). On Consciousness: Science & Subjectivity- Updated Works on Global Workspace Theory. New York, NY: The Nautilus Press Publishing Group.
Baars, B. J., Geld, N., and Kozma, R. (2021). Global workspace theory (GWT) and prefrontal cortex: Recent developments. Front. Psychol. 12:749868. doi: 10.3389/fpsyg.2021.749868
Babloyantz, A., and Destexhe, A. (1986). Low-dimensional chaos in an instance of epilepsy. Proc. Natl. Acad. Sci. U.S.A. 83, 3513–3517. doi: 10.1073/pnas.83.10.3513
Balister, P., Bollobás, B., and Kozma, R. (2006). Large deviations for mean field models of probabilistic cellular automata. Random Struct. Algorithms 29, 399–415. doi: 10.1002/rsa.20126
Barto, A. G., Sutton, R. S., and Anderson, C. W. (1983). Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Trans. Syst. Man Cybernet. 5, 834–846. doi: 10.1109/TSMC.1983.6313077
Beggs, J. M., and Timme, N. (2012). Being critical of criticality in the brain. Front. Physiol. 3:163. doi: 10.3389/fphys.2012.00163
Belanger, M. A., and Magistretti, P. (2011). Brain energy metabolism: focus on astrocyte-neuron metabolic cooperation. Cell Metab. 14, 724–738. doi: 10.1016/j.cmet.2011.08.016
Bishop, C. M.. (1995). Neural Networks for Pattern Recognition. Oxford; New York, NY: Oxford University Press. doi: 10.1201/9781420050646.ptb6
Bollobás, B., Kozma, R., and Miklos, D. (2010). Handbook of Large-Scale Random Networks, Vol. 18. Berlin; Heidelberg; New York, NY; Budapest: Springer Science & Business Media.
Brama, H., Guberman, S., Abeles, M., Stern, E., and Kanter, I. (2015). Synchronization among neuronal pools without common inputs: in vivo study. Brain Struct. Funct. 220, 3721–3731. doi: 10.1007/s00429-014-0886-6
Brennan, B. J., Pockett, S., Bold, G. E. J., and Holmes, M. D. (2011). A possible physiological basis for the discontinuity of consciousness. Front. Psychol. 2:377. doi: 10.3389/fpsyg.2011.00377
Bullmore, E., and Sporns, O. (2012). The economy of brain network organization. Nat. Rev. Neurosci. 13, 336–349. doi: 10.1038/nrn3214
Buzsáki, G.. (1998). Memory consolidation during sleep: a neurophysiological perspective. J. Sleep Res. 7, 17–23. doi: 10.1046/j.1365-2869.7.s1.3.x
Cabessa, J., and Villa, A. E. (2018). Attractor dynamics of a boolean model of a brain circuit controlled by multiple parameters. Chaos Interdiscipl. J. Nonlinear Sci. 28:106318. doi: 10.1063/1.5042312
Cloutier, M., Bolger, F., Lowry, J., and Wellstead, P. (2009). An integrative dynamic model of brain energy metabolism using in vivo neurochemical measurements. J. Comp. Neurosci. 27, 391–414. doi: 10.1007/s10827-009-0152-8
Crick, F., and Koch, C. (2003). A framework for consciousness. Nat. Neurosci. 6, 119–126. doi: 10.1038/nn0203-119
De Groot, A. D.. (2014). Thought and Choice in Chess. The Hague; Paris; New York, NY: De Gruyter Mouton.
Deco, G., Vidaurre, D., and Kringelbach, M. L. (2021). Revisiting the global workspace orchestrating the hierarchical organization of the human brain. Nat. Hum. Behav. 5, 497–511. doi: 10.1038/s41562-020-01003-6
Dehaene, S., Kerszberg, M., and Changeux, J.-P. (1998). A neuronal model of a global workspace in effortful cognitive tasks. Proc. Natl. Acad. Sci. U.S.A. 95, 14529–14534. doi: 10.1073/pnas.95.24.14529
Deschle, N., Gossn, J. I., Tewarie, P., Schelter, B., and Daffertshofer, A. (2021). On the validity of neural mass models. Front. Comput. Neurosci. 14:118. doi: 10.3389/fncom.2020.581040
Di Ventra, M., Pershin, Y. V., and Chua, L. O. (2009). Circuit elements with memory: memristors, memcapacitors, and meminductors. Proc. IEEE 97, 1717–1724. doi: 10.1109/JPROC.2009.2021077
Dreyfus, H.. (2007). Why Heideggerian AI failed and how fixing it would require making it more Heideggerian. Artif. Intell. 171, 1137–1160. doi: 10.1016/j.artint.2007.10.012
Edelman, G. M., Gally, J. A., and Baars, B. J. (2011). Biology of consciousness. Front. Psychol. 2:4. doi: 10.3389/fpsyg.2011.00004
Edelman, S., and Moyal, R. (2017). Fundamental computational constraints on the time course of perception and action. Prog. Brain Res. 236, 121–141. doi: 10.1016/bs.pbr.2017.05.006
Fekete, T., Van de Cruys, S., Ekroll, V., and van Leeuwen, C. (2018). In the interest of saving time: a critique of discrete perception. Neurosci. Conscious. 4:niy003. doi: 10.1093/nc/niy003
Fingelkurts, A. A., and Fingelkurts, A. A. (2006). Timing in cognition and EEG brain dynamics: discreteness versus continuity. Cogn. Process. 7, 135–162. doi: 10.1007/s10339-006-0035-0
Fingelkurts, A. A., and Fingelkurts, A. A. (2010). Alpha rhythm operational architectonics in the continuum of normal and pathological brain states: current state of research. Int. J. Psychophysiol. 76, 93–106. doi: 10.1016/j.ijpsycho.2010.02.009
Fingelkurts, A. A., Fingelkurts, A. A., and Neves, C. F. (2010). Natural world physical, brain operational, and mind phenomenal space-time. Phys. Life Rev. 7, 195–249. doi: 10.1016/j.plrev.2010.04.001
Fingelkurts, A. A., Fingelkurts, A. A., and Neves, C. F. (2013). Consciousness as a phenomenon in the operational architectonics of brain organization: criticality and self-organization considerations. Chaos Solitons Fractals 55, 13–31. doi: 10.1016/j.chaos.2013.02.007
Fingelkurts, A. A., Fingelkurts, A. A., and Neves, C. F. (2017). The legacy of a renaissance man: from mass action in the nervous system and cinematic theory of cognitive dynamics to operational architectonics of brain-mind functioning. Chaos Complex. Lett. 11, 81–91.
Franklin, S., Strain, S., Snaider, J., McCall, R., and Faghihi, U. (2012). Global workspace theory, its Lida model and the underlying neuroscience. Biol. Inspired Cogn. Arch. 1, 32–43. doi: 10.1016/j.bica.2012.04.001
Freeman, W.. (2000). Neurodynamics: An Exploration in Mesoscopic Brain Dynamics. London: Springer Science & Business Media. doi: 10.1007/978-1-4471-0371-4
Freeman, W. J.. (1991). The physiology of perception. Sci. Am. 264, 78–87. doi: 10.1038/scientificamerican0291-78
Freeman, W. J., Holmes, M. D., West, G. A., and Vanhatalo, S. (2006). Dynamics of human neocortex that optimizes its stability and flexibility. Int. J. Intell. Syst. 21, 881–901. doi: 10.1002/int.20167
Friston, K., Kilner, J., and Harrison, L. (2006). A free energy principle for the brain. J. Physiol. Paris 100, 70–87. doi: 10.1016/j.jphysparis.2006.10.001
Friston, K. J., Parr, T., Yufik, Y., Sajid, N., Price, C. J., and Holmes, E. (2020). Generative models, linguistic communication and active inference. Neurosci. Biobehav. Rev. 118, 42–64. doi: 10.1016/j.neubiorev.2020.07.005
Frohlich, J., Irimia, A., and Jeste, S. S. (2015). Trajectory of frequency stability in typical development. Brain Imaging Behav. 9, 5–18. doi: 10.1007/s11682-014-9339-3
Fuchs, A., Kelso, J. S., and Haken, H. (1992). Phase transitions in the human brain: spatial mode dynamics. Int. J. Bifurcat. Chaos 2, 917–939. doi: 10.1142/S0218127492000537
Fujisawa, S., Amarasingham, A., Harrison, M. T., Buzsáki, G., and Peyrache, A. (2015). Simultaneous electrophysiological recordings of ensembles of isolated neurons in rat medial prefrontal cortex and intermediate CA1 area of the hippocampus during a working memory task. CRCNS.org, 10:K01V5BWK.
Furber, S.. (2016). Large-scale neuromorphic computing systems. J. Neural Eng. 13:051001. doi: 10.1088/1741-2560/13/5/051001
Haimovici, A., Tagliazucchi, E., Balenzuela, P., and Chialvo, D. R. (2013). Brain organization into resting state networks emerges at criticality on a model of the human connectome. Phys. Rev. Lett. 110:178101. doi: 10.1103/PhysRevLett.110.178101
Hazan, H., Saunders, D., Khan, J., Sanghavi, D. T., Siegelmann, H. T., and Kozma, R. (2018). BindsNet: a machine learning-oriented spiking neural networks library in Python. Front. Neuroinform. 12:89. doi: 10.3389/fninf.2018.00089
Heck, D. H., Kozma, R., and Kay, L. M. (2019). The rhythm of memory: how breathing shapes memory function. J. Neurophysiol. 122, 563–571. doi: 10.1152/jn.00200.2019
Heck, D. H., McAfee, S. S., Liu, Y., Babajani-Feremi, A., Rezaie, R., Freeman, W. J., et al. (2017). Breathing as a fundamental rhythm of brain function. Front. Neural Circ. 10:115. doi: 10.3389/fncir.2016.00115
Iglesias, J., and Villa, A. E. (2007). Effect of stimulus-driven pruning on the detection of spatiotemporal patterns of activity in large neural networks. Biosystems 89, 287–293. doi: 10.1016/j.biosystems.2006.05.020
Iglesias, J., and Villa, A. E. (2010). Recurrent spatiotemporal firing patterns in large spiking neural networks with ontogenetic and epigenetic processes. J. Physiol. Paris 104, 137–146. doi: 10.1016/j.jphysparis.2009.11.016
Imamoglu, F., Kahnt, T., Koch, C., and Haynes, J.-D. (2012). Changes in functional connectivity support conscious object recognition. Neuroimage 63, 1909–1917. doi: 10.1016/j.neuroimage.2012.07.056
Izhikevich, E.. (2003). Simple model of spiking neurons. IEEE Trans. Neural Netw. 14, 1569–1572. doi: 10.1109/TNN.2003.820440
Janson, S., Kozma, R., Ruszinkó, M., and Sokolov, Y. (2019). A modified bootstrap percolation on a random graph coupled with a lattice. Discrete Appl. Math. 258, 152–165. doi: 10.1016/j.dam.2018.11.006
John, B. E., and Newell, A. (1990). Toward an engineering model of stimulus-response compatibility. Adv. Psychol. 65, 427–479. doi: 10.1016/S0166-4115(08)61233-9
Jolivet, R., Coggan, J. S., Allaman, I., and Magistretti, P. J. (2015). Multi-timescale modeling of activity-dependent metabolic coupling in the neuron-glia-vasculature ensemble. PLoS Comp. Biol. 11:e1004036. doi: 10.1371/journal.pcbi.1004036
Jordan, M. I., and Mitchell, T. M. (2015). Machine learning: trends, perspectives, and prospects. Science 349, 255–260. doi: 10.1126/science.aaa8415
Josipovic, Z.. (2019). Nondual awareness: consciousness-as-such as non-representational reflexivity. Prog. Brain Res. 244, 273–298. doi: 10.1016/bs.pbr.2018.10.021
Korhonen, O., Zanin, M., and Papo, D. (2021). Principles and open questions in functional brain network reconstruction. Hum. Brain Mapp. 42, 3680–3711. doi: 10.1002/hbm.25462
Kozma, R., Alippi, C., Choe, Y., and Morabito, F. C. (2019a). Artificial Intelligence in the Age of Neural Networks and Brain Computing. London; Cambridge, MA: Academic Press.
Kozma, R., and Freeman, W. J. (2009). The kiv model of intentional dynamics and decision making. Neural Netw. 22, 277–285. doi: 10.1016/j.neunet.2009.03.019
Kozma, R., and Freeman, W. J. (2016). Cognitive Phase Transitions in the Cerebral Cortex-Enhancing the Neuron Doctrine by Modeling Neural Fields. Cham: Springer. doi: 10.1007/978-3-319-24406-8
Kozma, R., and Freeman, W. J. (2017). Cinematic operation of the cerebral cortex interpreted via critical transitions in self-organized dynamic systems. Front. Syst. Neurosci. 11:10. doi: 10.3389/fnsys.2017.00010
Kozma, R., Hu, S., Sokolov, Y., Wanger, T., Schulz, A. L., Woldeit, M. L., et al. (2021). State transitions during discrimination learning in the gerbil auditory cortex analyzed by network causality metrics. Front. Syst. Neurosci. 15:641684. doi: 10.3389/fnsys.2021.641684
Kozma, R., Noack, R., and Manjesh, C. (2018). “Neuroenergetics of brain operation and implications for energy-aware computing,” in 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC) (IEEE), 722–727. doi: 10.1109/SMC.2018.00131
Kozma, R., Noack, R., and Siegelmann, H. T. (2019b). “Models of situated intelligence inspired by the energy management of brains,” in 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC) (IEEE), 567–572. doi: 10.1109/SMC.2019.8914064
Kozma, R., Pino, R. E., and Pazienza, G. E. (2012). Advances in Neuromorphic Memristor Science and Applications, Vol. 4. Dordrecht: Springer Science & Business Media. doi: 10.1007/978-94-007-4491-2
Kozma, R., and Puljic, M. (2015). Random graph theory and neuropercolation for modeling brain oscillations at criticality. Curr. Opin. Neurobiol. 31, 181–188. doi: 10.1016/j.conb.2014.11.005
Kozma, R., Puljic, M., Balister, P., Bollobás, B., and Freeman, W. J. (2005). Phase transitions in the neuropercolation model of neural populations with mixed local and non-local interactions. Biol. Cybernet. 92, 367–379. doi: 10.1007/s00422-005-0565-z
Kozma, R., Puljic, M., and Freeman, W. J. (2014). “Thermodynamic model of criticality in the cortex based on EEG/ECoG data,” in Criticality in Neural Systems, eds D. Plenz and E. Niebur (Weinheim: John Wiley & Sons, Ltd.), 153–176. doi: 10.1002/9783527651009.ch7
Kozunov, V., Nikolaeva, A., and Stroganova, T. A. (2018). Categorization for faces and tools-two classes of objects shaped by different experience-differs in processing timing, brain areas involved, and repetition effects. Front. Hum. Neurosci. 11:650. doi: 10.3389/fnhum.2017.00650
LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521:436. doi: 10.1038/nature14539
Lehmann, D., Ozaki, H., and Pál, I. (1987). Eeg alpha map series: brain micro-states by space-oriented adaptive segmentation. Electroencephalogr. Clin. Neurophysiol. 67, 271–288. doi: 10.1016/0013-4694(87)90025-3
Lundqvist, M., and Wutz, A. (2021). New methods for oscillation analyses push new theories of discrete cognition. Psychophysiology e13827. doi: 10.1111/psyp.13827
Madl, T., Baars, B. J., and Franklin, S. (2011). The timing of the cognitive cycle. PLoS ONE 6:e14803. doi: 10.1371/journal.pone.0014803
Magistretti, P. J.. (2006). Neuron-glia metabolic coupling and plasticity. J. Exp. Biol. 209, 2304–2311. doi: 10.1242/jeb.02208
Malagarriga, D., Pons, A. J., and Villa, A. E. (2019). Complex temporal patterns processing by a neural mass model of a cortical column. Cogn. Neurodyn. 13, 379–392. doi: 10.1007/s11571-019-09531-2
Marković, D., Mizrahi, A., Querlioz, D., and Grollier, J. (2020). Physics for neuromorphic computing. Nat. Rev. Phys. 2, 499–510. doi: 10.1038/s42254-020-0208-2
Mashour, G. A., Roelfsema, P., Changeux, J.-P., and Dehaene, S. (2020). Conscious processing and the global neuronal workspace hypothesis. Neuron 105, 776–798. doi: 10.1016/j.neuron.2020.01.026
Menétrey, M. Q., Vogelsang, L., and Herzog, M. H. (2021). A guideline for linking brain wave findings to the various aspects of discrete perception. Eur. J. Neurosci. 2021, 1–10. doi: 10.1111/ejn.15349
Miller, W. T., Werbos, P. J., and Sutton, R. S. (1995). Neural Networks for Control. Cambridge, MA: MIT Press.
Mnih, V., Kavukcuoglu, K., Silver, D., and Rusu, A. e. a. (2015). Human-level control through deep reinforcement learning. Nature 518:529. doi: 10.1038/nature14236
Mora-Sánchez, A., Dreyfus, G., and Vialatte, F.-B. (2019). Scale-free behaviour and metastable brain-state switching driven by human cognition, an empirical approach. Cogn. Neurodyn. 13, 437–452. doi: 10.1007/s11571-019-09533-0
Newell, A.Simon, H. A., et al. (1972). Human Problem Solving, Vol. 104. Englewood Cliffs, NJ: Prentice-Hall.
Nosonovsky, M., and Roy, P. (2020). Scaling in colloidal and biological networks. Entropy 22:622. doi: 10.3390/e22060622
Parr, T., and Friston, K. J. (2018). The discrete and continuous brain: from decisions to movement and back again. Neural Comput. 30, 2319–2347. doi: 10.1162/neco_a_01102
Pereira, A. Jr., Benevides Foz, F., and Freitas da Rocha, A. (2017). The dynamical signature of conscious processing: from modality-specific percepts to complex episodes. Psychol. Conscious. Theory Res. Pract. 4:230. doi: 10.1037/cns0000115
Qian, Y., Liu, F., Yang, K., Zhang, G., Yao, C., and Ma, J. (2019). Spatiotemporal dynamics in excitable homogeneous random networks composed of periodically self-sustained oscillation. Sci. Rep. 7, 1–13. doi: 10.1038/s41598-017-12333-3
Raichle, M., and Gusnard, D. (2002). Appraising the brain's energy budget. Proc. Natl. Acad. Sci. U.S.A. 99, 10237–10239. doi: 10.1073/pnas.172399499
Ramon, C., and Holmes, M. D. (2020). Increased phase cone turnover in 80-250 Hz bands occurs in the epileptogenic zone during interictal periods. Front. Hum. Neurosci. 14:615744. doi: 10.3389/fnhum.2020.615744
Roy, K., Jaiswal, A., and Panda, P. (2019). Towards spike-based machine intelligence with neuromorphic computing. Nature 575, 607–617. doi: 10.1038/s41586-019-1677-2
Schmidhuber, J.. (2015). Deep learning in neural networks: an overview. Neural Netw. 61, 85–117. doi: 10.1016/j.neunet.2014.09.003
Sengupta, B., Stemmler, M. B., and Friston, K. J. (2013). Information and efficiency in the nervous system-a synthesis. PLoS Comput. Biol. 9:e1003157. doi: 10.1371/journal.pcbi.1003157
Shew, W. L., and Plenz, D. (2013). The functional benefits of criticality in the cortex. Neuroscientist 19, 88–100. doi: 10.1177/1073858412445487
Simon, H. A.. (1967). Motivational and emotional controls of cognition. Psychol. Rev. 74:29. doi: 10.1037/h0024127
Skarda, C. A., and Freeman, W. J. (1987). How brains make chaos in order to make sense of the world. Behav. Brain Sci. 10, 161–173. doi: 10.1017/S0140525X00047336
Stam, C. J., and Reijneveld, J. C. (2007). Graph theoretical analysis of complex networks in the brain. Nonlinear Biomed. Phys. 1, 1–19. doi: 10.1186/1753-4631-1-3
Steyn-Ross, A., and Steyn-Ross, M. (2010). Modeling Phase Transitions in the Brain, Vol. 509. New York, NY; Dordrecht; Heidelberg; London: Springer. doi: 10.1007/978-1-4419-0796-7
Stieg, A. Z., Avizienis, A. V., Sillin, H. O., Aguilera, R., Shieh, H.-H., Martin-Olmos, C., et al. (2019). “Self-organization and emergence of dynamical structures in neuromorphic atomic switch networks,” in Handbook of Memristor Networks, eds L. Chua, G. Sirakoulis, and A. Adamatzky (Cham: Springer), 391–427. doi: 10.1007/978-3-319-76375-0_14
Sulis, W.. (2021). The continuum from temperament to mental illness: dynamical perspectives. Neuropsychobiology 80, 135–147. doi: 10.1159/000509572
Tagliazucchi, E.. (2017). The signatures of conscious access and its phenomenology are consistent with large-scale brain communication at criticality. Conscious. Cogn. 55, 136–147. doi: 10.1016/j.concog.2017.08.008
Tal, I., and Abeles, M. (2018). Imaging the spatiotemporal dynamics of cognitive processes at high temporal resolution. Neural Comput. 30, 610–630. doi: 10.1162/neco_a_01054
Teixeira, F. P. P., and Murray, S. (2015). “Local and global criticality within oscillating networks of spiking neurons,” in 2015 International Joint Conference on Neural Networks (IJCNN) (IEEE), 1–7. doi: 10.1109/IJCNN.2015.7280561
Tetko, I. V., and Villa, A. E. (2001). A pattern grouping algorithm for analysis of spatiotemporal patterns in neuronal spike trains. 2. Application to simultaneous single unit recordings. J. Neurosci. Methods 105, 15–24. doi: 10.1016/S0165-0270(00)00337-X
Tognoli, E., Dumas, G., and Kelso, J. S. (2018). A roadmap to computational social neuroscience. Cogn. Neurodyn. 12, 135–140. doi: 10.1007/s11571-017-9462-0
Tognoli, E., and Kelso, J. S. (2014). The metastable brain. Neuron 81, 35–48. doi: 10.1016/j.neuron.2013.12.022
Tononi, G., and Koch, C. (2015). Consciousness: here, there and everywhere? Philos. Trans. R. Soc. B Biol. Sci. 370:20140167. doi: 10.1098/rstb.2014.0167
Tozzi, A., Peters, J. F., Fingelkurts, A. A., Fingelkurts, A. A., and Marijuán, P. C. (2017). Topodynamics of metastable brains. Phys. Life Rev. 21, 1–20. doi: 10.1016/j.plrev.2017.03.001
Tsuda, I.. (2001). Toward an interpretation of dynamic neural activity in terms of chaotic dynamical systems. Behav. Brain Sci. 24, 793–810. doi: 10.1017/S0140525X01000097
Turing, A. M., and Haugeland, J. (1950). Computing Machinery and Intelligence. Cambridge, MA: MIT Press. doi: 10.1093/mind/LIX.236.433
Turkheimer, F. E., Hellyer, P., Kehagia, A. A., Expert, P., Lord, L.-D., Vohryzek, J., et al. (2019). Conflicting emergences. Weak vs. strong emergence for the modelling of brain function. Neurosci. Biobehav. Rev. 99, 3–10. doi: 10.1016/j.neubiorev.2019.01.023
Tversky, A., and Kahneman, D. (2011). Thinking, Fast and Slow. New York, NY: Farrar, Straus and Giroux.
Waldrop, M. M.. (2016). The chips are down for Moore's law. Nature News 530:144. doi: 10.1038/530144a
Wang, R., Lin, P., Liu, M., Wu, Y., Zhou, T., and Zhou, C. (2020). Hierarchical connectome modes and critical state jointly maximize human brain functional diversity. Phys. Rev. Lett. 123:038301. doi: 10.1103/PhysRevLett.123.038301
Werbos, P. J., and Davis, J. J. (2016). Regular cycles of forward and backward signal propagation in prefrontal cortex and in consciousness. Front. Syst. Neurosci. 10:97. doi: 10.3389/fnsys.2016.00097
Werner, G.. (2013). Consciousness viewed in the framework of brain phase space dynamics, criticality, and the renormalization group. Chaos Solitons Fractals 55, 3–12. doi: 10.1016/j.chaos.2012.03.014
White, P. A.. (2018). Is conscious perception a series of discrete temporal frames? Conscious. Cogn. 60, 98–12. doi: 10.1016/j.concog.2018.02.012
Xu, Y., Wu, X., Mao, B., Lü, J., and Xie, C. (2020). Fixed-time synchronization in the pth moment for time-varying delay stochastic multilayer networks. IEEE Trans. Syst. Man Cybernet. Syst. 52, 1–10. doi: 10.1109/TSMC.2020.3012469
Yufik, Y. M.. (2013). Understanding, consciousness and thermodynamics of cognition. Chaos Solitons Fractals 55, 44–59. doi: 10.1016/j.chaos.2013.04.010
Yufik, Y. M.. (2019). The understanding capacity and information dynamics in the human brain. Entropy 21:308. doi: 10.3390/e21030308
Yufik, Y. M., and Friston, K. (2016). Life and understanding: the origins of “understanding” in self-organizing nervous systems. Front. Syst. Neurosci. 10:98. doi: 10.3389/fnsys.2016.00098
Zanin, M., Ivanoska, I., Güntekin, B., Yener, G., Loncar-Turukalo, T., Jakovljevic, N., et al. (2021). A fast transform for brain connectivity difference evaluation. Neuroinformatics. 1–15. doi: 10.1007/s12021-021-09518-7
Keywords: machine understanding, cortex, perception, consciousness, graph theory, neuropercolation, phase transition, criticality
Citation: Kozma R, Baars BJ and Geld N (2021) Evolutionary Advantages of Stimulus-Driven EEG Phase Transitions in the Upper Cortical Layers. Front. Syst. Neurosci. 15:784404. doi: 10.3389/fnsys.2021.784404
Received: 27 September 2021; Accepted: 03 November 2021;
Published: 08 December 2021.
Edited by:
Yan Mark Yufik, Virtual Structures Research Inc., United StatesReviewed by:
Andrew A. Fingelkurts, BM-Science, FinlandAlessandro E. P. Villa, University of Lausanne, Switzerland
Copyright © 2021 Kozma, Baars and Geld. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Robert Kozma, rkozma@memphis.edu