Skip to main content

HYPOTHESIS AND THEORY article

Front. Integr. Neurosci., 04 November 2022

From the origins to the stream of consciousness and its neural correlates

  • Independent Research Center of Brain and Consciousness, Andijan, Uzbekistan

There are now dozens of very different theories of consciousness, each somehow contributing to our understanding of its nature. The science of consciousness needs therefore not new theories but a general framework integrating insights from those, yet not making it a still-born “Frankenstein” theory. First, the framework must operate explicitly on the stream of consciousness, not on its static description. Second, this dynamical account must also be put on the evolutionary timeline to explain the origins of consciousness. The Cognitive Evolution Theory (CET), outlined here, proposes such a framework. This starts with the assumption that brains have primarily evolved as volitional subsystems of organisms, inherited from primitive (fast and random) reflexes of simplest neural networks, only then resembling error-minimizing prediction machines. CET adopts the tools of critical dynamics to account for metastability, scale-free avalanches, and self-organization which are all intrinsic to brain dynamics. This formalizes the stream of consciousness as a discrete (transitive, irreflexive) chain of momentary states derived from critical brain dynamics at points of phase transitions and mapped then onto a state space as neural correlates of a particular conscious state. The continuous/discrete dichotomy appears naturally between the brain dynamics at the causal level and conscious states at the phenomenal level, each volitionally triggered from arousal centers of the brainstem and cognitively modulated by thalamocortical systems. Their objective observables can be entropy-based complexity measures, reflecting the transient level or quantity of consciousness at that moment.

Introduction

What can be said with certainty about the brain is that this is a complex dynamical system: (i) governed by the deterministic laws of nature at the physical (causal) or “hard” level which (ii) implements cognitive processing at the computational (unconscious) or “soft” level while (iii) its conscious manifestations occur at the phenomenal (mental) or “psyche” level. These three levels are also separated across different spatial-temporal scales. Neural activity is presented at the microscale of cellular interactions, cognitive processing occurs at the mesoscale of neural populations, and conscious experience emerges only at the macroscale of brain dynamics (Varela, 1995; Revonsuo and Newman, 1999). Another way to distinguish between the hard and the soft level in terms of network neuroscience is to relate the former to structural (anatomical) connectivity of hard-wired neurons, where causation really occurs. In contrast, the soft level corresponds to functional connectivity, where distant correlations between brain regions take place. Computations at the soft level cannot violate causal interactions at the hard level but are merely imposed upon physically interacting neurons. They cannot change brain dynamics that obey deterministic laws.

Thus, the mind-brain duality can be viewed in the formal terms of “hard/soft” parallelism between causation and computation. While computation is typically defined through symbolic manipulations of information, its processing depends ultimately on causal chains between inputs and outputs. The question of how those symbolic manipulations translate into conscious experience in the brain and why other information-processing systems such as computers and AI systems lack conscious experience at the psyche level is one of the most mysterious problems in neuroscience.

The main postulate of Gestalt psychology is that conscious experience is a unified whole, which is greater than the sum of its parts. More generally, this postulate is known in the context of the spontaneous emergence of unexpected higher-level phenomena that are not reducible to their low-level constituents (Bedau and Humphreys, 2008; Gibb et al., 2019). A related issue in neuroscience is specified as the binding problem: how brain regions composed of billions of (unconscious) neurons can generate a unified conscious experience at a given moment of time (Edelman, 2003). In fact, many (if not all) theories of consciousness originate from or, at least, can be reduced to how they decide this problem. Among the candidates proposed by known theories are integrated information (II) of irreducible causal mechanisms (Tononi, 2008), global workspace (GW) for broadcasting (Baars et al., 2013; Mashour et al., 2020), updating of priors (UP) in predictive processing (Knill and Pouget, 2004; Clark, 2013), meta-representation (MR) by recurrent processing (Lamme, 2006; Rosenthal, 2008), self-organized criticality (SOC) in brain dynamics (Werner, 2009; Kozma and Freeman, 2017), adaptive resonance (AR) of brain structures (Grossberg, 2017; Hunt and Schooler, 2019), and even large-scale quantum entanglement with a consequent collapse (Hameroff and Penrose, 2014; Fisher, 2015).

Those in turn can be grouped by the similarity of mechanisms or processes involved: II + GW by integration-differentiation processes, GW + UP + MR by feedback mechanisms, SOC + AR by spontaneous synchronization and phase transition. However, any grouping is somewhat arbitrary as the underlying mechanisms can converge to a thermodynamic account: the neural binding arises when brain activity is balanced on the edge between order and disorder. Broadly speaking, consciousness emerges in a very special state of matter somewhere between a perfect crystal and an ideal gas (Tegmark, 2015).

The cognitive evolution theory or CET (Yurchenko, 2022), outlined here, adopts the SOC approach for its apparent advantages over the above models in studying consciousness. SOC is neurophysiologically reliable in resolving the binding problem without resorting to exotic physics or mysterious mind-matter dualism. This provides rich mathematical formalisms applicable to brain dynamics. SOC also proposes the avenue for explaining universal dynamical capacities of the brain to account for large-scale emergent phenomena without involving the so-called downward or top-down causation that might make consciousness like a homunculus due to “synergistic emergence” (Lau, 2009; Hoel et al., 2013; Mediano et al., 2022). On the other hand, SOC is abundantly presented in nature (Bak, 1996; Jensen, 1998; Haken, 2004). However, we do not normally assume that an arbitrary physical system exhibiting critical signatures is conscious. Something else must be inherent to a system to generate consciousness.

CET starts from the obvious fact that the only place where consciousness certainly resides is the brain. There are four principled features that make the brain distinct from all other critical systems. First, the brain consists of neurons specialized for transmitting information over spike patterns. The neurons had evolved from autonomous biological cells possessing all properties of life not merely as mechanistic binary devices. Hence, consciousness is a property of living systems. Second, there are arousal mechanisms regulating sleep-wake cycles in these living systems which can be suppressed by anesthetics. It is also known that damage to arousal nuclei causes immediate coma when the rest of the brain can remain intact (Parvizi and Damasio, 2001; Giacino et al., 2014). Hence, consciousness is impossible without involving special neural nuclei in the brain responsible for arousal. Third, the brain learns and accumulates knowledge. Hence, the brain is a cognitively evolving system. However, AI systems can learn and even cognitively evolve without any kind of awareness. Is there something else that is inherent to the conscious brain but absent in unconscious machines?

The fourth and ultimate distinction is volition, the ability to make free decisions not causally predetermined from the past. This valuable property is thought to be intrinsic to many (if not all) brain systems regardless of their conscious features. In contrast, we do not normally grant volition to computers and AI systems even if they can sometimes surpass humans in cognitive performance. The volition of this kind is akin to one that can be ascribed to clockwork’s engine, i.e., it is an ordinary physical process carrying energy out of one place into another. How much our intuition is right by assuming that consciousness and volition are evolutionarily linked?

With the advent of causation neuroscience, the detailed relationship between statistical models of neural activity and actual causation in the brain is intensively debated (e.g., Albantakis et al., 2019; Reid et al., 2019; Weichwald and Peters, 2021). Statistical measures such as Granger Causality (Granger, 1969) or Entropy Transfer (Schreiber, 2000) had been suggested to infer some aspect of causal interaction among neural entities that are then modeled on a particular structural time-directed graph over a set of nodes. Their definition is based entirely on the predictability of some time-series. If Y contains information that helps to predict X beyond the degree to which X already predicts its own future, then Y is said to have a causal influence on X. Causal modeling must reflect dynamic processes irrelevant to the question of whether a system of interest processes information or not. In cognitive neuroscience, causal inference is based on a synthesis of functional and effective connectivity extracted from neuroimaging data (Friston, 1994). The “gold standard” to establishing whether a stimulus variable Y affects the target variable X is to introduce controlled perturbations to the brain.

It must be however emphasized that causality measures do not necessarily reflect physical causal chains (Seth, 2008). Meanwhile, brain dynamics are commonly believed to evolve completely in causal ways over conscious and unconscious volitional repertoires of the brain. Within those volitional repertoires, the ability generally labeled “free will” is associated with the sum of executive functions, self-control, decision-making, and long-term planning. Thus, it makes free will inseparable from the biological function of consciousness, evolution-driven, and implicitly active (Feinberg and Mallatt, 2020). Putting the above question differently, could consciousness supervene on its own physical substrate to choose the course of action free of predetermination from the past?

To answer this question, a unified theory of consciousness should account for brain activity at three hierarchical levels: (i) at the causal (hard) level; (ii) at the computational (soft) level; and (iii) at the phenomenal (psyche) level. The first two levels should explain how subjective experience and self-awareness emerge from the underlying brain dynamics over which cognitive processing is carried out.

Consciousness and Volition in Brain Dynamics

The stochastic account of free volition can be often found in the literature. For example, Rolls (2012) suggests: “in so far as the brain operates with some degree of randomness due to the statistical fluctuations produced by the random spiking times of neurons, brain function is to some extent non-deterministic, as defined in terms of these statistical fluctuations.” Since the brain contains billions of neurons, causal processes can be only estimated with the help of network statistics extracted from different neuroimaging data. However, probabilistic (counterfactual) descriptions, as those derived from causality measures, reflect the state of our knowledge (ignorance) that, by itself, does not violate determinism. Brain dynamics can still be completely deterministic, i.e., predetermined.

Many free-will advocates usually suggest that if even conscious volition cannot violate determinism at the causal level, it can still be involved in long-term planning of actions at the psyche level. They argue that the ability to use optimal algorithms in predictive processing would be a much more important factor than whether the brain operates deterministically or not (Rolls, 2020). Does it mean that by using those algorithms AI systems might suddenly acquire free volition? Relevantly, relying on this pure computational aspect of volition in the context of the hard-soft duality would imply no obstacle in creating machine (hence, copyable) consciousness at the psyche level (Dehaene et al., 2017; VanRullen and Kanai, 2021). This comes from the observation that there can be no operational difference between a perfect computational simulation of, for instance, Alice’s actions and an in silico copy of her consciousness running automatically on many digital clones (Aaronson, 2016). Thus, ignoring the hard-soft duality also entails the problem of the privacy of consciousness: the clones should know what it is like to be Alice.

A typical scenario suggested for manifesting free volition at the computational (soft) level is one where Alice can consciously plan something ahead of time, for example, visiting her friends tonight. Thus, Alice’s freedom to choose can be proven by achieving her goal. Upon a closer examination, all such scenarios are behavior-driven, yet based on uncertain and rough assumptions about what occurs at the microscale of neural interactions. Within a rigorous physical framework, the spatial and temporal locations of action should be specified via the stream of conscious states, each processed by the brain at the hard level.

Suppose Alice plans to visit her friends at the moment t0 when her consciousness is in a state X. Whenever she could reach this goal, her conscious state at that exact moment t would be Y. But the state Y should in turn have been consistently processed from the previous state Y − 1 over ubiquitous causal chains. How could it be done freely? Moreover, the manifestation of the conscious will should be related not to Alice being at her goal state Y but to its mental initiation in X. Indeed, after the decision has been made, her goal-directed behavior could be completely deterministic. However, this state should also be causally processed from the previous state X − 1, and so on. How could mental initiation be free of the past? Hence, if consciousness cannot choose the next state from a given past state, no future state in the stream can be chosen at all.

On the other hand, if the brain cannot make a choice free of the past, the old-fashioned fatalism, also known as superdeterminism in the context of quantum mechanics (‘t Hooft, 2016; Hossenfelder and Palmer, 2020), would prevail. This states that neither consciousness nor even the brain might violate deterministic laws to do otherwise than what has been predetermined by the past. How could free volition be reconciled with computational models of consciousness at the hard level?

Volition in Theories of Consciousness

Consciousness and volition are intrinsically linked, and both are largely ignored in neuroscience. Although there are now a plethora of various theories of consciousness, the free will problem still remains largely neglected. The theories do not explain how the brain integrates consciousness (psyche), cognition (soft), and volition (hard) seamlessly across the three hierarchical levels. Yet, as being static in nature, most of them aim to explain the structure of conscious experience per se without accounting for the successive alternation of conscious states over time. Many authors attempt to compare these theories (Doerig et al., 2021; Del Pin et al., 2021; Signorelli et al., 2021), or even to reconcile some of them (Shea and Frith, 2019; Chang et al., 2020; Graziano et al., 2020; Mashour et al., 2020; Northoff and Lamme, 2020; Mallatt, 2021; Seth and Hohwy, 2021; VanRullen and Kanai, 2021; Niikawa et al., 2022). Another tendency is to incorporate these static theories into a more general dynamical framework such as the Temporo-spatial Theory of Consciousness (Northoff and Zilio, 2022) which is somewhat reminiscent of Operational Architectonics (Fingelkurts et al., 2010), the whole-brain mechanistic models from a bottom-up perspective (Cofré et al., 2020), or the self-organizing harmonic modes coupled with the free energy principle (Safron, 2020).

In general, all these theories are not concerned with brain dynamics at the underlying hard level from which conscious states emerge. As being static in design, they have also missed another fundamental aspect of consciousness, namely, its cognitive evolution over a lifetime going by accumulating new knowledge and skills. Accordingly, the stream of consciousness (though implied) is not properly defined. At the same time, they all ascribe a special, active role to consciousness while being indifferent to the free will problem by adopting the view that consciousness can somehow influence brain dynamics at the psyche level. It is implicitly assumed that consciousness: (i) facilitates learning; (ii) is useful for voluntary decision-making; and (iii) provides stability to the objects of experience in an egocentric framework (Seth and Baars, 2005).

For example, Integrated Information Theory (IIT) defines consciousness as information a system is able to integrate. It does not consider perception, cognition, and action at all by aiming at the quantitative account of phenomenal consciousness due to “irreducible causal mechanisms” (Oizumi et al., 2014). It makes the striking conclusion that “a conscious choice is freer, the more it is causally constrained within a maximally irreducible complex” (Tononi, 2013) without explaining how free volition might be manifested there. Such a kind of volition turns out to be just superdeterministic (see below). It could well account for the individuality of subjective experience as specified by the complex at the psyche level but not for the indeterminism of volition at the hard level. The IIT approach can be even extended to a general idea that the very evolution of life is just the triumph of determinism. It argues that living systems can thrive across the universe as soon as these autonomous systems have more cause-effect power than non-living environments (Marshall et al., 2017).

On the contrary, Predictive Processing Theory (PPT) can well explain perception, cognition, and action by making the brain an error-minimizing machine (Clark, 2013; Seth and Hohwy, 2021). Although PPT is unclear about where exactly, between priors and posteriors, the conscious experience should appear, it can in principle separate discrete conscious states emerging at the psyche level as ultimate decisions of Bayesian learning from unconscious predictive processing at the soft level. Nevertheless, PPT cannot still account for free volition, which is covertly embedded in attentional effort and active inference (Friston et al., 2013; Pezzulo et al., 2018). This is just the point where free will and active consciousness converge. If consciousness is an algorithm for the maximization of resilience (Rudrauf et al., 2017), then PPT has to explain how and why conscious processing in the brain should differ from deep machine learning in AI systems, which can exploit the same computational models but lack both volition and conscious experience.

Another dominant theory, Global Workspace Theory (GWT), relies explicitly on the active role of consciousness at the psyche level, which is required for global access, broadcasting information, and self-monitoring (Dehaene and Changeux, 2011; Baars et al., 2013). According to the theory, a physical system “whose successive states unfold according to a deterministic rule can still be described as having free will, if it is able to represent a goal and to estimate the outcomes of its actions before initiating them” (Dehaene and Naccache, 2001). Thus, GWT adopts just the aforementioned scenario with Alice deciding where she will be tonight. It is therefore not surprising that GWT does not suggest any obstacle to machine consciousness (Dehaene et al., 2017), which by virtue of its cognitive architecture could be spontaneously endowed with free will.

Finally, psychological theories such as Higher-Order Thought (Lau and Rosenthal, 2011), Attention Schema (Graziano et al., 2020), Radical Plasticity Thesis (Cleeremans, 2011), or Self Comes to Mind (Damasio, 2010) argue that self-awareness or metacognition would separate conscious states from unconscious processing via self-referential mechanisms. These mechanisms would make the brain aware of its own states, unlike other biological and artificial networks. Accordingly, conscious will in these theories is similar to “conscious veto” suggested by Libet (1985) to circumvent the findings of his famous free will experiments.

The Libet-type experiments have been based on two temporal measures: the readiness potential detected from the supplementary motor area, and the awareness of wanting to move reported with the clock. The delay observed between neural motor predictors at the hard level and conscious intentions at the psyche level was around several hundred milliseconds (Libet, 1985; Guggisberg and Mottaz, 2013; Schultze-Kraft et al., 2016), thereby making conscious intentions a post-factum phenomenon. Libet had proposed that it could occur due to conscious deliberations before consciousness could block an action with the explicit veto on the movement. Nevertheless, since any kind of intentional veto must also be causally processed, it has been noted that the blocking itself could be predetermined (Soon et al., 2013).

Yet, some authors assume that noisy neural fluctuations can be involved in self-initiated actions (Schurger et al., 2012). They argue that the key precursor process for triggering internally-generated actions could be essentially random in a stochastic framework (Khalighinejad et al., 2018). First, those actions generated internally by the brain from noisy fluctuations would have little relevance to the ability of consciousness to make its own choice. Meanwhile, in the psychological theories mentioned above, conscious will would involve higher-order thoughts to check many moves ahead of time, and then with this information to make a free choice (Rolls, 2020). Second, stochastic noise as being classical (thermal) in nature does not violate determinism to account for freely generated actions even on the brain’s authorship. Something else is needed.

In contrast, quantum-inspired theories of consciousness take seriously the free will problem by adopting quantum indeterminism that might affect brain dynamics at the hard level. They are based on the conceptualization of free will suggested by Bell in his famous theorem (Bell, 1993). The key assumption of Bell was that Alice’s free choice (as defined above) should not be controlled by any hidden (unknown to our modern knowledge) deterministic variables λ. For example, these variables might include all the internal and external information about the past of Alice’s brain and everything in her environment. The choice is then formalized by conditional probability,

p(A|λ)=p(A)(1)

The theorem had shown that under measurements of preciously prepared quantum experiments no such variables should exist in principle unless we had to agree on a “cosmic conspiracy” that constrained Alice to make not any choice but just the one that did not violate standard statistical correlations of the Bell’s inequality (Gallicchio et al., 2014). Note that, as suggested beyond the neuroscientific context of Libet-type experiments, Equation (1) does not discriminate between a choice made by Alice with her conscious will, and a choice internally generated by her brain itself. The only thing required there is that A has not been predetermined from the past by λ. Thus, by ruling out conscious will from Libet-type experiments, only the indeterminism of neural activity can account for the Bell theorem, which has been well-confirmed experimentally (e.g., Aspect et al., 1982).

Orchestrated Objective Reduction (Orch OR) of Hameroff and Penrose (2014), the most known of quantum-inspired theories of consciousness, is explicitly based on the indeterminism of the wavefunction collapse or “objective reduction” (OR), also known as the measurement problem. In quantum mechanics, the measurement problem has the striking property of observer-dependence. In contrast, Penrose argues for a real quantum-gravitational OR that occurs everywhere in the universe, independently of observation, as spontaneous “proto-conscious” events that are then orchestrated in microtubules of the brain to give rise to consciousness and volition. According to Orch OR, consciousness must be in principle incomputable as being orchestrated (Orch) by quantum entanglement with consequent OR. Nonlocal correlations between microtubules (for global binding of dissociated brain regions) and backward time referral (for closed causal loops) are then required “to rescue causal agency and conscious free will” (Hameroff, 2012).

Remarkably, the mathematicians Conway and Kochen in their Free Will Theorem (a modified version of the Bell theorem) make a statement very similar to that of Penrose (though without concerning themselves with OR): “If conscious observers have a certain freedom to choose, then particles already have their own share of this freedom” (Conway and Kochen, 2008). It is often argued that the quantum randomness on which the statement reposes has little to do with “free will” thought to be caused by a reason rather than by chance (Koch, 2009; Aaronson, 2016). Conway and Kochen had however noted that if a subject’s action was indeed free of the past, then it should be difficult to find a testable (i.e., objective from a third-person perspective) difference between physical randomness and the subject’s genuine behavioral freedom, both not being predetermined by the previous history of the universe. In the context of neuroscience, their statement has to be inverted and specified as follows:

If particles have some freedom to respond to the environment independently of the past, then the observers have the same kind of initial freedom due to a quantum mechanism Bell-certified by Equation (1) in their brains.

Of course, ascribing free will to living or non-living systems does not depend anyhow on quantum effects. Volition is typically associated with the action a system is capable of initiating. It necessarily involves causation. However, what is the sense of saying that a clockwork toy has the volition to move, or, that a more sophisticated AI system is self-initiated in performing a cognitive task? To discriminate between volition and trivial energy flow in an arbitrary (natural or artificial) system the former must not be predetermined from the past. Volition is not a matter of consciousness or cognition but a matter of causal freedom. Under this definition, quantum particles can have some sort of freedom but this freedom is washed out in macroscopic systems.

To bind volition with consciousness, we need to translate physical predetermination into the notion of predictability. First of all, deterministic systems can in principle be perfectly predictable. It is already a matter of our current knowledge and technology. Now let us modify the famous Turing test where a machine might mimic a human by answering various questions. Suppose, a future machine can be trained to predict precisely Alice’s every choice ahead of time. The machine can thus perfectly simulate Alice’s stream of consciousness so that we might be fooled by asking any question. Why then should we deny that the machine possesses a copy of Alice’s consciousness and, thus, is itself conscious?

Hence, if we disagree that humans are very sophisticated but still deterministic and copyable machines, there must be something that prevents one from making a deterministic machine conscious. This is volition. Indeed, if Alice’s brain is able to make a choice not predetermined from the past, there cannot be in principle a perfect predictor of Alice’s stream of consciousness. This makes her consciousness unique, i.e., non-copyable for any future technology. Thus, the reason to assume the above statement stems fundamentally from the evolutionary perspective.

CET suggests a dynamical model based on a framework, drawn over diverse neuroscientific domains, with contributions from critical dynamics, predictive processing, and evolutionary neurobiology to approach a unified theory of consciousness grounded in physics. The widespread idea that brains are error-minimizing machines has neglected a crucial ingredient of its adaptive framework: before minimizing an error the brain should already have that error internally triggered under cognitive correction. CET argues that the brain can be viewed more generally as both an error-amplifier and a modulator of the primitive volitional reflexes based on chemo- and photoreceptors of unicellular organisms. In doing so, CET makes the assumption that brains have primarily evolved as volitional (quantum in origin) subsystems of organisms at the hard level. CET emphasizes the importance of randomness in evolution (Yurchenko, 2021a) contrary to the idea that life can be viewed as the triumph of determinism (Marshall et al., 2017).

Volition from The Evolutionary Perspective

In modern science of consciousness, the Cartesian presumption that animals are only biological automata incapable of experiencing conscious and emotional states looks chauvinistic or even perverse (Lamme, 2018; Fields et al., 2021). Nonetheless, the tradition to separate humans from the rest of the animal kingdom is still persistent. Now the division has shifted to the free will debate. It is assumed that only humans enjoy free will due to the sophisticated computations the human brain is able to carry out at the soft level, while other animals, though passively conscious, are deprived of this gift (Doyle, 2009). CET finds this division evolutionarily implausible.

Maintaining homeostasis in the face of dangerous and unfortunate environmental conditions is the basic imperative for the evolution of the brain (Ashby, 1960). The only valuable advantage the living systems might gain over non-living systems is the ability to decide their way freely among stimulus-reaction repertoires. This ability is commonly referred to as volition. Meanwhile, volitional mechanisms of the brain are still ignored in cognitive neuroscience, remaining hidden under attentional effort, enactive cognitive function, and conscious control. CET argues that volition is a key neural mechanism of evolution placed between organisms and non-living systems. Life could not have flourished on Earth without volitional mechanisms which are the only neural candidate that opens a door to the stimulus-reaction repertoires over the “tyranny” of causal chains.

The underlying idea here is the one of evolutionary neurobiology: the ultimate aim of brain evolution over species, from simplest organisms to humans, was to maximize their adaptive success. Under selection pressure, the main function of neural systems resulted in their ability to integrate and process a very large number of sensory inputs and motor responses occurring in parallel (Edelman, 2003). Thus, the cognitive evolution of the brain over every organism’s lifetime (ontogeny) and the general evolution of the brain over species (phylogeny) should go hand in hand to promote each other. CET makes the general assumption that after acquiring free volition mechanisms to overcome the fatalism of cause-effect interactions which govern completely non-living systems, organisms could have advanced their adaptive success only by evolving cognitive functions capable of predicting future events. Cognition and memory should have evolved exclusively as adaptive (computational) abilities of organisms at the soft level to benefit from some underlying volitional mechanism at the hard level. Otherwise, evolution would have had no reason to advance those adaptive properties over rigid causal chains in deterministic systems.

How much is it plausible that evolution had endowed primitive networks of simplest organisms with some kind of freedom which would have been of value for them in survival and reproduction? Can invertebrates have some volitional mechanism, evolutionarily embedded in their neural networks to make a choice not predetermined by the past? CET argues that primitive neural networks should have primarily evolved as free-volitional subsystems of organisms, not as deterministic prediction machines (Knill and Pouget, 2004; Clark, 2013), requiring larger biological resources. Accordingly, their conscious properties, typically related to higher animals, should have appeared later than their unconscious cognitive functions presented in invertebrates (Brembs, 2011). Instead of being an active participant that creates an internal representation of the environment, the conscious experience would be an extension of primitive volitional-emotional motivations (Mashour and Alkire, 2013). Thus, contrary to the idea that evolution had evolved consciousness to perform complex volitional actions (Pennartz et al., 2019; Feinberg and Mallatt, 2020), in CET consciousness is a byproduct of both volitional mechanisms and unconscious cognitive processing grounded in sensorimotor coupling (Engel et al., 2013). With this assumption on the origins of consciousness, CET suggests a radically new and physically rigorous solution to the free will problem.

Unlike the known quantum-inspired models involving the most mysterious quantum-mechanical effects to account for quantum computing and/or quantum memory storage that might directly mediate consciousness (Hameroff and Penrose, 2014; Fisher, 2015; Georgiev, 2020), CET suggests a minimal use of quantum randomness at the sub-cellular level. This refers to the most reliable mechanisms like the Beck-Eccles quantum trigger, proposed as a quantum-mechanical model for the mechanism of exocytosis (Beck and Eccles, 1992, 1998). Exocytosis is a discrete all-or-nothing event consisting of the opening of presynaptic membrane channels with the release of neurotransmitters in the synaptic cleft. The trigger is based on the tunneling of a quasi-particle across a potential energy barrier. Beck and Eccles (1992) argue that “the mental intention (the volition) becomes neurally effective by momentarily increasing the probability of exocytosis in selected cortical areas such as the supplementary motor area.” Thus, they maintain the cortex-centered conceptualization of conscious free will in Libet-type experiments.

Instead, CET places the mechanism in the brainstem to account for the indeterminism of unconscious free volition initiated in the arousal nuclei. This key quantized event could then be physically amplified across many spatiotemporal scales due to neuronal avalanches which are intrinsic to SOC (Blanchard et al., 2000; Beggs and Plenz, 2003; Chialvo, 2010; Tognoli and Kelso, 2014). The brainstem is a phylogenetically ancient brain region that comprises various nuclei executing many vital autonomic functions for maintaining homeostasis such as blood pressure, heartbeat, or blood sugar levels (Paus, 2000; Parvizi and Damasio, 2001). Those autonomic functions should have evolved much early than thalamocortical regions engaged in elaborating cognitive contents and conscious experience.

Remarkably, this ancient region also contains the arousal centers responsible for permanent vigilance upon which the stream of consciousness reposes. Its arousal machinery is a precondition for behavior and conscious experience (Mashour and Alkire, 2013). CET proposes that each conscious state in the stream is volitionally driven at the causal (hard) level from subcortical arousal centers via the ascending reticular activating system (ARAS) to thalamocortical systems, involved in perception and cognition at the computational (soft) level. In CET, conscious experience emerges at the psyche level as a passive phenomenon of cognitive brain evolution that goes by accumulating new knowledge and skills freshly updated in memory networks.

By postulating the quantum mechanism that will be called the “neurophysiological free-volition mechanism” (NFVM) and placed into the brainstem, CET can account for the indeterminism of brain dynamics without resorting to large-scale quantum mysteries1. The corollaries are the following. While brain dynamics will still be presented by classical stochastic processes as those traditionally described in standard models of neuroscience, the stream of consciousness will no longer be predetermined from the past by hidden deterministic variables λ due to the presence of the NFVM. This makes Alice’s consciousness (whose secured privacy is guaranteed by the Bell-certified NFVM) unique but gives her consciousness no power over the brain unlike typical scenarios mentioned above.

In general, the controversy around free will has been inherited by the science of consciousness from the mind-body problem, originally discussed in philosophy. What exactly should be associated with Alice? Is it her brain or her consciousness generated by her brain (leaving aside her body that makes the brain alive and functional)? In CET, if Alice is associated with her brain, she has free volition. On the contrary, if Alice is associated with her consciousness, she has no free will. Consider, for instance, the following sentence: “I can have true free will: I can have true alternatives, true freedom to choose among them, true will to cause what I have decided, and eventually true responsibility” (Tononi et al., 2022). For CET, the validity of this statement depends on how the “I” is conceptualized.

This also implies that not integrated information of irreducible causal mechanisms (Tononi, 2008; Oizumi et al., 2014), architecture peculiarities of neural networks (Dehaene and Naccache, 2001; Baars et al., 2013), cognitive processing (Clark, 2013), or higher-order linguistic thoughts (Lau and Rosenthal, 2011; Rolls, 2012) but free volition, inherited by the brain from fast and random reflexes rooted in chemo-, magneto-, and photoreceptor cells which are very sensitive to quantum effects (Arndt et al., 2009; Brookes, 2017; McFadden and Al-Khalili, 2018) is the main obstacle that prevents computer scientists from making deterministic machines conscious.

The Stream of Consciousness in Brain Dynamics

Consciousness has been a prescientific concept with a number of different connotations that relate to various aspects of conscious experience. The science of consciousness must rely on the fact that consciousness is a dynamic process, not a thing or a capacity (James, 1950). It should also take into account that the observation of single cells activity has little to say about mental processes. Large-scale interactions between neural networks are therefore more important than the contribution of individual neurons per se. Thus, only neural dynamics at the macro-scale can account for global brain states accompanied by conscious experience. Most attempts to understand the neural mechanisms of consciousness have proceeded by searching for the “neural correlates of consciousness” or NCC (Crick and Koch, 2003). However, correlations by themselves do not provide explanations and there is a need for a framework connecting fundamental aspects of conscious experience at the phenomenal (psyche) level to the corresponding aspects of brain dynamics at the underlying causal (hard) level.

The evolution of consciousness depends on the physical continuous dynamics of the brain, comprising about 1011 neurons connected to each other by thousands of synapses. Clearly, consciousness not only requires neural correlates with their appropriate anatomical architecture but also the time to be processed. A wealth of experimental evidence suggests that conscious experience is a discrete phenomenon processed unconsciously at the soft level (see Section “Temporal resolution of the stream of consciousness”). Moreover, a pure biophysical approach to studying neural activity at micro- and mesoscopic scales cannot account for subjective, internally generated mental phenomena without resorting to their contextual emergence at the macroscopic scale (Atmanspacher and beim Graben, 2007). It has been pointed out many times that there should be an optimal spatio-temporal grain size at which the brain can generate phenomenal experience (Tononi, 2008; Chang et al., 2020). In general, the correspondence between conscious states and neural processes should be provided by mapping brain dynamics onto the state space of consciousness.

CET compresses all these requirements into three key prerequisites:

1. Physicalism (mind-brain identity): consciousness depends entirely on brain activity governed by natural laws at the hard level, not on anything else;

2. Dynamism (temporality): consciousness not only requires the NCC but also the time to be cognitively processed at the soft level;

3. Contextuality (scale-dependence): only large-scale brain dynamics can account for the emergence of conscious states at the psyche level.

A principled distinction between CET and classical theories is that CET involves quantum indeterminism. On the other hand, while quantum-inspired theories engage quantum computing across the whole cortex to account for consciousness and free volition, CET makes minimal use of quantum randomness placed into the arousal nuclei of the brainstem to initiate cognitive processing due to stochastic brain dynamics that are classical in nature. Thus, CET is a semi-classical physical theory of consciousness.

Deriving consciousness from brain dynamics

Based on the three prerequisites, CET will model consciousness as the stream of macrostates, each specified by a particular structural-functional configuration of the whole-brain network 𝒩, where NCC ⊆ 𝒩. Here 𝒩 stands for a graph G = (N, E), where N = |𝒩| is the set of nodes (ideally, neurons), and EN × N is the set of edges (ideally, synapses). Because it is computationally impossible to operate on N ≈ 1011, the first step in formalizing the stream of consciousness is to approximate large-scale brain dynamics at the hard level. To derive the stream from the SOC approach, CET refers to the Langevin formalism as a most general description of a system that depends upon a deterministic flow y and stochastic fluctuations ω [which, note, do not discriminate between quantum (i.e., ontic) and statistical (i.e., epistemic) randomness, e.g., in Brownian motion]

dψ=γψ(t)dt+dω(t)(2)

The formalism can then be transformed into different models depending on the way researchers will adopt that formalism in their study. Those models may be biophysical (mean-field) approximations (Breakspear, 2017; Parr et al., 2020) or phenomenological (synchronization) models such as Stuart-Landau, Kuramoto, Haken-Kelso-Bunz, and other models (Cofré et al., 2020; Kelso, 2021; Kraikivski, 2022; Vohryzek et al., 2022).

Here ψ (𝒩, t) is a descriptive function whose representation by the order parameter in a phase space O should account for metastability, avalanches, and SOC of the global neural activity wandering among dynamical attractors (Kelso, 1995). Originally grounded in physics, chemistry, and biology (Bak, 1996; Jensen, 1998; Haken, 2004), SOC is thought to be of crucial importance in neural activity as it poises the brain on the edge between order and disorder near a bifurcation (Blanchard et al., 2000; Chialvo, 2010; Beggs and Timme, 2012; Deco et al., 2015). This allows the neural network 𝒩 to exhibit the properties of scale-free networks and to produce flexible patterns of coordination dynamics at the hard level. This, in turn, can generate a repertoire of different conscious states at the psyche level, thereby increasing an adaptive behavioral response of the organism to given environmental stimuli (Hesse and Gross, 2014; Cocchi et al., 2017; Dahmen et al., 2019).

CET postulates the emergence of consciousness from brain dynamics at critical points as its derivative extracted in discretized time τ,

S(τ)=defdψdτ(3)

The continuous/discrete dichotomy appears naturally between the brain dynamics, described at the causal (hard) level, and the stream S(τ) of conscious states, presented at the phenomenal (psyche) level. In effect, Equation (3) should capture the instantaneous transitions from continuous brain dynamics to discretized conscious states, identified with a single point oO in a phase space. The phase transitions in brain dynamics occurring at discrete moments of time can then be viewed as the manifestation of pulsating consciousness in the framework of cinematic theory of cognition (Freeman, 2006). This approach finds now experimental evidence in many studies (Haimovici et al., 2013; Mediano et al., 2016; Tagliazucchi, 2017; Demertzi et al., 2019; Kim and Lee, 2019) showing that only states integrated near criticality can ignite conscious experience in the brain.

CET suggests a simple mathematical analogy between consciousness and the physical force, derived from the momentum in Newtonian mechanics,

F=dp/dt=ma(4)

Seeing consciousness as a “mental force” seems to be more moderate and accurate, at least ontologically, than viewing consciousness as a fundamental property of matter like mass, charge, and energy (Tononi, 2008). The analogy between conscious experience and mass, advocated by IIT, is based on a quantitative account of the level of consciousness measured by integrated information Φ a single complex of irreducible causal mechanisms might generate (Oizumi et al., 2014). IIT argues: if a complex can generate Φ, no matter whether it is organic or not, it will have consciousness (Tononi and Koch, 2015). Instead, CET brings into focus the dynamism of consciousness. The analogy with force goes in line with the fact there is a tiny but principled distinction between mass and force in physics by Equation (4): the former is a scalar quantity which is indeed constantly intrinsic to a system, whereas the latter is a dynamical characteristic of motion defined by a vector quantity of a system’s action that can vanish in inertial states.

Similarly, Equation (3) represents consciousness as a dynamical characteristic of the neural network 𝒩 not as an intrinsic potency of causal mechanisms in the brain or anywhere else. According to this conceptualization, the brain has no mental force if its dynamics depart from criticality, as it occurs in unconscious states such as coma, sleep, or general anesthesia (Hudetz et al., 2014; Tagliazucchi et al., 2016; Golkowski et al., 2019; Huang et al., 2020), but not in resting states where neural activity preserves criticality (Deco and Jirsa, 2012; Barttfeld et al., 2015). On the other hand, even in critical dynamics, the brain lacks the mental force during some interval Δt until a next conscious state is unconsciously processed.

To make the above analogy more comprehensive, imagine a clockwork toy, say, a jumping frog. The engine of the toy will impel it to iterate the same jump over and over. The force occurs only at discrete moments of jump-initiation between which motion decays. In critical dynamics, the brain exhibits flexible patterns of coordinated dynamics which decay after some critical value (Gollo et al., 2012; Tognoli and Kelso, 2014). Conscious experience can be ignited only on the edge between the two phases as if the brain accumulated information for triggering the next “jump” with some mental force at discrete moments of time (Figure 1A).

FIGURE 1
www.frontiersin.org

Figure 1. Brain criticality and conscious experience. (A) In large-scale brain dynamics, conscious states emerge at critical points near phase transitions between synchronization (order) and desynchronization (disorder) patterns of neural activity at the microscale. (B) The map m transforms each critical point of brain dynamics, described in a phase space, onto the whole-brain network 𝒩 in a vector space as a particular NCC responsible for a certain conscious state at that moment of time. (C) The stream of consciousness can then be formalized as a discrete chain of states (N-dimensional vectors) and studied in causal dynamical modeling as a directed acyclic graph. (D) In neural pleiotropy, many neurons constitute a particular NCC for producing a certain conscious percept, thereby involving a single neuron in generating very different percepts. (E) Conscious experience is a product of unconscious computations initiated by the brainstem at the hard level and accomplished by various thalamocortical systems at the soft level.

Temporal resolution of the stream of consciousness

According to Equation (3), the brain needs some time to process a new conscious state from the previous one. There are two complementary ways to estimate the interval Δt– either by monitoring brain dynamics to calculate phase transitions at the hard level or by obtaining a subjective report at the psyche level. Unfortunately, both approaches do not provide an exact estimate. The monitoring of brain dynamics is non-trivial because of the heterogeneous intrinsic timescales involved. An averaged interval is usually reported to be about 200 ms (Kozma and Freeman, 2017; Deco et al., 2019). The second approach, based on first-person reportability, is affected by the problem of multiple temporal resolutions (White, 2018).

To comprise both approaches, we assign the interval to a wide window Δt ≈ 100–450 ms that encompasses multiple experimental findings—from the periodicity of attentional cycles at approximately 7–13 Hz (VanRullen et al., 2014) to the attentional blink on masked targets separated by 200–450 ms (Shapiro et al., 1997; Drissi-Daoudi et al., 2019). Yet, an important neurophysiological aspect of brain dynamics is that the stream S(τ) cannot be normally delayed for a period longer than about 300 ms according to the timescale proposed to be crucial for the emergence of consciousness (Dehaene and Changeux, 2011). Consciousness spontaneously fades after that period, for example, in anesthetized states (Hudetz et al., 2014; Tagliazucchi et al., 2016).

Only states that emerge globally and are integrated at critical points are conscious. In the stream, each state appears instantaneously as a snapshot accompanied by a phenomenal percept of the “specious present” (Varela, 1999), which provides human (and most likely animal) observers of the external world with an egocentric frame of reference, preserving its stability over time (Seth and Baars, 2005). Due to the cognitive updating of this self-referential frame by the acquisition of new contents, consciousness remains well informed at the psyche level about what is occurring around. Thus, while conscious states emerge only at discrete critical points of brain dynamics, subjects can still feel the continuity of being and acquire the illusion of a persistent presence.

There are also arguments for a continuous flow of conscious experience (Fekete et al., 2018; Kent and Wittmann, 2021). However, these are typically grounded in the phenomenology of consciousness and time instead of being based on the neuroscience of consciousness and the physics of time. Yet, they do it without making a difference between the rigorous notion of mathematical continuum and something that can be continued. If such arguments have any merit, their core can be formulated like this: “If conscious experience is produced by the brain, then it would seem that there must be a lawful relation between the state of the brain at the given time and the percept that the person is experiencing… because experiencing or perceiving is an activity, not something to be looked at” (Brette, 2019). Indeed, such a “lawful relation” must be suggested.

Equation (3) explains how discrete conscious snapshots at the psyche level can be separated from continuous brain dynamics at the hard level. Moreover, because consciousness does not observe how the experience was processed by the brain, awareness requires no time to be ignited. Since the ignition across the brain’s workspace occurs phenomenally due to SOC, there is no dynamical process that should transmit information into a special site of the brain to reach the subject’s awareness. Experience is the information the brain has unconsciously computed at that specious moment τ. The illusion of temporal continuity of consciousness is merely a trivial corollary of its self-evidential nature: consciousness cannot in principle detect its own absence in the brain. Whenever consciousness searches for itself, the search happens successfully for an obvious reason: in the stream, consciousness will always be present in introspection. Likewise, whenever we look in a mirror, we always see ourselves as if we were constantly there. We know that it is not true, but it is impossible to catch the mirror before our image will appear.

In particular, the discreteness of conscious experience explains how the subjective distortion of perceived temporal duration produced by emotionally charged events and causing the sense of dilated time in the stream S(τ) can occur. Overestimation of time is associated with many factors such as enhanced memory encoding or intensive involvement of the limbic system and the interconnection with medial cortical areas (Dirnberger et al., 2012). This depends on how the states were unconsciously processed over multiple timescales in brain dynamics (Merchant et al., 2013) while their sensory-cognitive contents had been compressed over temporal chunks T = ∑ Δt, composed of non-overlapping intervals. Otherwise, for the sake of continuous conscious perception, we should agree on something like space-time effects in Relativity theory as if a perceived chunk T, as compared to a physically real TR, would indeed be dilated by the Lorenz transformations T=TR/1υ2/c2. Thus, the discrete model of the stream S(τ) proposes a more reasonable and parsimonious explanation of the subjective distortion in time perception as merely depending on a ratio TΔt, i.e., on a variation of Δt during which those emotionally charged conscious states were processed.

The passage of time (whatever it is physically) cannot be perceived at all without sensory-cognitive contents processed by different brain systems at the soft level and integrated globally at the psyche level. This also explains why time perception vanishes completely in altered states of consciousness such as sleep, anesthesia, or epileptic seizures. It occurs as brain dynamics deviate from criticality upon which discrete conscious experience entirely reposes (Tagliazucchi et al., 2016).

The stream of consciousness as a causal chain of discrete states

Now let m: OV be a map from the phase space O onto a vector space V over the product N × N of all neurons of the brain network 𝒩. The map returns S(τ) from a point oO to an N-dimensional vector x = [n1, n2,‥., nN], where ni = 1 or ni = 0 stand for neurons active or inactive at the given time. We write (omitting details),

S(τ)mx(5)

Thus, each state S(τ) is represented now by x as a certain structural-functional configuration of 𝒩 that is responsible for that subjective snapshot at the moment τ. The discreteness of the stream signifies that all the conscious states could, at least in principle, be naturally enumerated from a subject’s birth, not merely by a lag in experimental settings.

Let the brain bring consciousness to a state xi at the moment τ = t. We can rewrite Equation (5) as follows.

ψ(𝒩,t)=xi(6)

The next conscious state will emerge over the time interval as

ψ(𝒩,t+Δt)=xi+1(7)

In the timeless description, the stream S(τ) is a discrete causal chain 𝒳 = (X,<) (Figure 1B), where xiX and whose relation < standing for the causal (and temporal) order is transitive and irreflexive. Here the irreflexivity means ixi<xi that forbids closed causal loops and, in particular, instantaneous feedback circuitry in brain dynamics that might somehow enable consciousness with causal power over the brain, for example, due to the presumed quantum temporal back-referral of information in Orch OR (Hameroff and Penrose, 2014). CET strictly rejects the possibility that consciousness could—classically or quantum-mechanically—choose a state to arrive at.

Consciousness is a physically classical macro-phenomenon (though quantum-triggered) that is always local in space and time. For consciousness, planning something, as discussed in Alice’s scenario, does not mean to be already in that state. In other words, every state in S(τ) is actual, and any chosen state should also be actual. However, it is physically impossible for the brain or for the stream of consciousness to be ahead of itself. Physically, the brain is exactly where it is. Mentally, consciousness is what it is just now. Even assuming quantum temporal back-referral, closed causal loops would be involved, as if consciousness would be now in a state it had already chosen in the state before. On any account, this would imply that before choosing a state consciousness should already be in that state despite irreflexivity.

Herzog et al. (2016) suggest a discrete model of consciousness very similar to that of CET. The authors assume that conscious states represent integrated, meaningful outputs of unconscious (modular) computations carried out at the soft level which had been causally provided by dynamical attractors at the hard level of interacting neurons. They compare a conscious snapshot with a k-dimensional feature vector Vi that contains a value for each relevant feature i of an external stimulus (e.g., color, size, shape, or position), which together constitute a meaningful post-hoc representation of the event processed unconsciously during an interval Δt. The interval is the period of sense-making (Herzog et al., 2016). CET transforms their model, based on a statistical presentation of an external stimulus by a feature vector Vi, into an internal neuron-based representation of that stimulus in the brain network 𝒩. Namely, in the chain 𝒳 = (X,<), each snapshot S(τ) is a particular NCC described by a N-dimensional vector xi = [n1, n2,‥., nN] that constitutes a given conscious state which is responsible for the perceived stimulus at that time.

The stream of consciousness can now be represented in terms of dynamical causal networks formalized typically as a directed acyclic graph Gu = (N,E), where edges E indicate not synaptic connections but cause-effect pairs among a set of nodes N which in turn represent a set of associated random variables with a state space Ω=ΠiΩNi, based on a given set of background conditions (states of exogenous variables) U = u (Albantakis et al., 2019). The representation can be made by mapping the temporal evolution of the brain network 𝒩 onto the acyclic graph. Accordingly, the chain 𝒳 can be defined in Gu by a partition of its nodes N into temporally ordered slices, N = N1, ‥., Nk, each interpreted as a particular NCC of a discrete dynamical system starting with an initial slice N1x1 and such that the parents of each next slice are fully contained within the previous slice. This definition prohibits any instantaneous causal loops and signifies that Gu (hence 𝒳) fulfills the Markov property (Figure 1C).

Neural Correlates of Consciousness

According to Equation (5) each conscious state S(τ) in brain dynamics can be represented by xi as a particular NCC. The NCC has been traditionally defined as the minimal set of neuronal events that gives rise to a specific aspect of a conscious percept. In the context of the binding problem, the NCC should then be seen not as the static connectivity patterns over anatomical brain regions but as dynamical and transient networks with synchronous neural activity as those evolve across the whole brain in critical dynamics. Klein et al. (2020) argue for what can be called “neural pleiotropy” in the search for NCC. In biology, pleiotropy refers to the fact that a single gene can affect a number of otherwise distinct phenotypic variables.

Accordingly, it can be said that not only many neurons from different brain areas simultaneously contribute to a single conscious state xi (a particular NCC) at that moment τ but also many conscious states are affected by a single neuron, e.g., via selective participation in producing neuronal avalanches (Bellay et al., 2021). Thus, a conscious experience at the psyche level involves multiple phenomenal elements, many of which can occur in the context of other experiences (Figure 1D). Thus, one might then consider how different NCC might be mapped onto a state space of phenomenal experience to account for qualia at the psyche level (Kraikivski, 2020). However, CET specifies that the problem of the privacy of consciousness cannot be solved abstractly. Secured privacy is guaranteed by the NFVM, preventing the possibility of copying one’s unique consciousness.

Generally, the search for a physical substrate S for some P makes sense if P is a well-defined phenomenon accessible to objective experience. Here, the objective can be replaced by the collectively subjective, something that everyone can experience independently. For example, if P means the Moon, this is well-defined under objective experience: we all understand that the Moon means an object in the night sky which is accessible for our public evidence. Further, physicists tell us that the physical substrate S of this P consists of atoms that are already inaccessible to our public evidence. Instead, we simply believe physicists, and, in principle, everyone can verify their belief in a lab. Yet, some people may also believe in ghosts revealed to them in their subjective experience, but, unlike atoms, these might not even in principle be tested in a lab for public evidence.

Likewise, when one speaks of consciousness, we all understand what it means. But this produces a lot of confusion. Consciousness is ontologically like neither one of these. Unlike the Moon, consciousness cannot be in principle the object of our collective experience. Unlike an atom, we do not even need to believe in its existence because of its self-evidential nature. Unlike a ghost, it is naturally revealed to all of us with no need to verify its presence in a lab. Although consciousness is the necessary prerequisite for the public evidence of the existence of anything, it itself is neither well-defined nor even a phenomenon under objective experience. Thus, the empirical search for NCC cannot be theory-neutral but depends on how consciousness is conceptualized (Pauen, 2021). How might the NCC be detected without knowing what exactly the function of consciousness is and how it has evolved?2

The NCC program, as initiated by Crick and Koch, has been explicitly based on the idea of active consciousness. This allowed the authors to propose the search for neural correlates of visual consciousness which should “produce the best current interpretation of the visual scene in the light of past experience …and to make this interpretation directly available, for a sufficient time, to the parts of the brain that contemplate and plan voluntary motor output” (Crick and Koch, 2003). The program had quickly been divided into two parts: the search for the level of consciousness and the search for the cognitive (reportable) contents of specific NCC (Miller, 2007; Hohwy, 2009). The former was explicitly concentrated on the diagnostic assessments of coma/level of arousal in humans, tested for instance with the Glasgow Coma scale or Coma Recovery Scale-revised, while the latter remained concerned with conscious (mostly visual) perception vs. unconscious processing in experimental studies on awake subjects.

These in turn generated a lot of cognate concepts such as prerequisites of consciousness (preNCC), proper correlates (NCCpr), their consequences (NCCcons), and others (Bachmann and Hudetz, 2014; Northoff and Zilio, 2022). CET is not involved in all of these. CET denies active consciousness, thereby reducing the importance of cortical regions for the study of NCC. CET argues that the search for true NCC suggested as an “easy part” of the hard problem (Crick and Koch, 2003) is an attempt to sidestep the privacy of consciousness in the same way as the Turing test attempts to do it by obtaining a machine report.

Instead, CET decomposes the concept of NCC into separated neural configurations that are responsible for different conscious states. In principle, one can uncover an NCC for any particular state x = [n1, n2,‥., nN] by merely detecting activity patterns in 𝒩 at that moment τ. The minimal neural substrate can then be defined by the intersection of all those states over the stream S(τ), or, more generally, as

NCCmin=i=12Nxi(8)

Here 2N is a set of all possible states (from full vigilance to sleep, or coma) a subject might have during his lifetime. To identify which minimal correlates are necessary for consciousness, we need to associate it with the most primitive core of subcortical consciousness presented in infants born without the telencephalon (Damasio et al., 2012; Aleman and Merker, 2014; Barron and Klein, 2016; Panksepp et al., 2017) or even in patients with the unresponsive wakefulness syndrome (see Section “Mental force in critical brain dynamics”).

Otherwise, if we assume an active role of consciousness (i.e., free will) in an attentional effort, active inference, decision-making, planning, and goal-directed behavior, the NCC would comprise most of the brain, including even the cerebellum (Schmahmann, 2010; Friston and Herreros, 2016) that is generally not considered in the NCC debate,

NCCactive=i=12Nxi(9)

The general idea of linking consciousness to volition and cognition relies on the observation that they seem always to be associated with each other (Bachmann and Hudetz, 2014; Naccache, 2018; Aru et al., 2019). However, the premise that consciousness is active in brain dynamics claims much more than the NCC program has suggested—to account for the minimal neural substrate that is necessary and sufficient to ignite consciousness in critical points of brain dynamics. The premise makes legitimate the use of such terms as “conscious processing” in parallel with unconscious processing.

CET denies the possibility of conscious processing that would be active and continuous over the stream S(τ) of discrete conscious states. This argues that the integration of neural activity, associated with a particular conscious state emerging at the psyche level, cannot be dissociated from underlying cognitive and volitional processes at the soft and hard levels respectively. Some regions in the front or in the back of the cortex that are thought to be relevant in the NCC debate (Boly et al., 2017) are also involved in various functions that are necessary for volition and cognition, e.g., the supplementary motor area in Libet-type free will experiments, or the inferior parietal cortex and many corticothalamic loops in predictive processing. However, in CET, consciousness is derived from brain dynamics only as a mental force, it cannot persist without volition and cognition, and completely fades if there is no space for the cognitive evolution when brain dynamics depart from criticality (as in coma, epileptic seizures, and general anesthesia).

Thus, contrary to the idea of active consciousness, advocated by many theories of consciousness, CET takes an “inverted perspective”: consciousness is a passive phenomenon, ignited at critical points of brain dynamics and resulting from unconscious computational processes implemented by various thalamocortical functional systems. When distilled from those processes, conscious states emerge as a mental marker of brain dynamics, by analogy with the physical force as a dynamical marker of a moving system. This is the subjective experience that makes a difference over time from its own perspective, thereby acquiring the illusion of causal power over neural activity3.

Instead of discussing different brain regions with their contributions to subjective experience, e.g., the prefrontal cortex vs. posterior “hot zones” (Koch et al., 2016; Boly et al., 2017), or dissociating attention from consciousness (Koch and Tsuchiya, 2007; Nani et al., 2019), CET argues: conscious states emerge as neural configurations x = [n1,‥.,nN] at critical points of brain dynamics, represented then as different NCC. Consciousness has not played any special role to have its own evolution-driven neural substrate; it emerges passively from volitional-cognitive substrates distributed across all regions of the brain (and from primitive organisms to humans). There is no special NCC that might causally influence brain dynamics, NCCactive = ∅. Instead, the NFVM initiates a random micro-event which can be stochastically amplified by bottom-up causation via neuronal avalanches to generate a particular conscious state. Such avalanches are intrinsic to and ubiquitous in critical dynamics (Beggs and Plenz, 2003; Hahn et al., 2010; Shew et al., 2011).

Volition in The Stream of Consciousness

To account for quantized neuronal events that might cause avalanches across the brain, CET places the NFVM into the brainstem (Figure 1E). This derives from the fact that the brainstem is responsible for spontaneous arousal and vigilance conducted through the ARAS projecting to the thalamocortical systems (Paus, 2000; Merker, 2007). Although the cortex is mostly responsible for elaborating conscious contents, damage to ARAS and intralaminar nuclei of the thalamus can abolish consciousness (Mhuircheartaigh et al., 2010; Edlow et al., 2020). Moreover, the neuromodulatory influences of the brainstem (due to its anatomical location in the neural hierarchy) act as control parameters of SOC, moving the whole cortex through a broad range of metastable states, responsible for cognitive processing in brain dynamics (Bressler and Kelso, 2001). The NFVM is thus a natural trigger of critical dynamics over which the stream of consciousness evolves.

Placing the NFVM into the brainstem is also related to the fact that the brainstem is the oldest brain region that plays a fundamental role in the maintenance of homeostasis (Parvizi and Damasio, 2001). Evolutionarily, brains have evolved gradually as multilevel systems consisting of anatomical parts that were selection-driven as adaptive modules for executing specific functions. Any brain function requires an appropriate neuronal structure able to generate various dynamical patterns to carry out that function optimally. It is well-known that the global architecture of the brain is not uniformly designed across its anatomical parts, which structural signatures are specialized under corresponding functions. Possibly, the network characteristics of the brainstem with its reticular formation were developed to be especially conducive to small neuronal fluctuations. These small neuronal fluctuations were then amplified to account for reflexes and primary volitional reactions. Later they were projected to higher thalamocortical systems to make computations at the soft level.

Thus, CET reduces the active role of consciousness to the brainstem activation of each conscious state, triggered by the NFVM [this would happen in the same way as, for instance, long-lived quantum entanglements in the cryptochromes of the retina are thought to participate in precise magnetoreception of the avian compass (Ritz, 2011; Hiscock et al., 2016)]. The NFVM is necessary to warrant the indeterminism of the stream based on quantum randomness as emphasized above (Conway and Kochen, 2008), not merely on a stochastic (deterministic) account of brain dynamics. Thus, in the stream S(τ), all conscious states (but not consciousness) must be initially Bell-certified against hidden deterministic variables λ by Equation (1) to be independent from the past.

Many researchers commonly agree that conscious states represent a simultaneous binding of neural modules, but there is no agreement regarding the NCC underlying this form of integration. CET addresses the problem by dividing the NCC into two parts: a first one that initiated a new conscious state, and a second part that integrated the cognitive contents of that state. The neural correlates (NC) of the NFVM (located in the brainstem) provide the neural basis for the integration of all states processed by the thalamocortical systems. This makes the NFVM a necessary and sufficient mechanism to maintain the sleep-wake cycle in patients with unresponsive wakefulness syndrome (UWS) after widespread thalamocortical damage when brainstem activity is more or less preserved (Laureys et al., 2004),

NCCmin=NC(NFVM)=defUWS(10)

Recovery of consciousness depends then upon the functional reemergence of the ARAS, which must provide sufficient input via the thalamic projections to the anterior forebrain mesocircuit and frontoparietal network (Edlow et al., 2020). Indeed, it is known that full recovery from UWS can be accompanied by restoration of activity in frontoparietal areas (Giacino et al., 2014). In contrast, brainstem lesions cause immediate coma by damaging the ARAS and its associated neuromodulatory systems (Parvizi and Damasio, 2001). Thus, if the NFVM is severely damaged, no conscious state could be initiated in brain dynamics, even when thalamocortical systems remained completely or partially intact,

NC(NFVM) = =defcoma(11)

The NFVM could also account for subcortical conscious experience in infants born without the cortex, on the condition of some preserved islands of the limbic system coupled with arousal centers. This goes in line with Merker’s proposal that the stream of consciousness derives from the interactions within the brainstem, supporting the triangle of action-target-motivation, while the thalamocortical systems provide cognitive contents for “immediate, unreflective experience” (Merker, 2007). For example, experimental evidence shows that decorticated rats preserve consciousness driven by volition (NFVM) from the brainstem, but often exhibit hyperactive wandering and more emotional behavior than their neurologically intact siblings whose behaviors are just suppressed and cognitively modulated by the cortex (Panksepp et al., 1994).

Consciousness as A Mental Force

In Newtonian mechanics, the force is a time derivative of the momentum. But there are no things such as the force and the momentum; these are only the dynamical characteristics of a moving system. Similarly, consciousness can be derived—in any meaningful sense—only from the cognitive evolution of the brain over lifetime. Consciousness is thus no more real than the force in physics though the latter can be measured and calculated. Hence, if we will proceed with this conceptualization of consciousness as a mental force, we need to propose objective observables that could measure its magnitude at a given time.

Mental force in critical brain dynamics

Before suggesting a candidate measure for consciousness let us return to the metaphor with a clockwork frog. The only operation a frog can execute is a jump initiated by its engine repeatedly with the same magnitude of physical force, F = ma. The frog will make the same jump over and over regardless of how the environment changes. To continue the analogy between this force, triggered by the frog’s engine at discrete moments of time, and the brain’s mental force, ignited momentarily at critical points of brain dynamics, we need to compare the iteration of the same jump with the alternation of conscious states.

In the stream, each state must make a difference from the past (Edelman, 2003) to gain an insight into what occurs around it. Thus, a mental jump at the psyche level over discretized time τ provides consciousness with new knowledge that can be viewed as Bayesian updating of prior beliefs in the framework of predictive processing models at the soft (unconscious) level. What follows is that the same conscious state, experienced by a subject repeatedly, would be like a freeze-frame on a TV screen, yet, preserving the same mental force. The brain, generating every time the same snapshot, could not make a difference from the past to learn something. Accordingly, memory networks should have nothing to update over the stream of such states.

Perhaps, just these conditions are presented in UWS that, unlike coma, is accompanied by spontaneous eye and limb movements without evidence of goal-oriented behavior or sensory responsiveness (Giacino et al., 2014; Schiff et al., 2014). According to Equation (10), in these conditions, the NFVM should still be preserved within the brainstem to maintain arousal cycles with the lowest level of conscious experience, whereas thalamocortical systems responsible for cognitive (unconscious) processing would be severely suppressed. It also explains why patients who recovered after UWS have no memories of their staying in that state as if time perception was also broken (Gosseries et al., 2014).

Now replace the frog with a Bell-certified random number generator producing a string of numbers a Turing machine could not compute. Let the numbers symbolize conscious states, totally disconnected over the stream. The stream of such states would be unpredictable in favor of free volition. On the other hand, as the brain has to make its own predictions about the world, cognition and logical reasoning should be unguided in this case because a subject having those states could not concentrate on a task to gain a coherent insight into it. The subject could then be viewed as going among many arbitrary frames (thoughts), while the information gain (as estimated by the Kullback-Leibler divergence between the prior and the posterior), would be too big to make a consistent difference from the past.

Perhaps, similar conditions might characterize attention deficit hyperactivity disorder (ADHD) in the literature. Its typical symptoms are disorganization, impulsiveness, trouble with focussing on tasks, and low attention span (Salmi et al., 2020). In this disorder, consciousness, as the way of being (Tononi, 2008), is ill-adaptive to the environment. How can free volition be manifested there? CET predicts that ADHD can occur when the random neural event initiated by the NFVM and transmitted by the ARAS up to the thalamocortical systems cannot be normally modulated by these parts of the NCC. Whatever the neurophysiological reasons of it may be, there is evidence that critical phenomena are involved in ADHD (e.g., Zimmern, 2020; Heiney et al., 2021; Rocha et al., 2022).

Neural complexity as a measure of consciousness

According to Equation (3), conscious states can emerge only near criticality while the brain moves from a subcritical regime where excitation dies out to a supercritical regime accompanied by neuronal avalanches (Haldeman and Beggs, 2005; Chialvo, 2010). SOC maintains a long-term correlation of brain regions to provide optimal information storage and computational power on the edge between order and disorder (Figure 2A). In CET, consciousness is like a river buoy fluctuating on the surface of water regardless of its depth. The behavior of such a float can well characterize underwater dynamical processes (for example, in fishing or navigation) without, however, having any influence on those complex and invisible processes.

FIGURE 2
www.frontiersin.org

Figure 2. Critical dynamics, phase transitions and the NFVM. (A) In critical dynamics, one neuron excites another and causes avalanches of activity spreading across 𝒩 and obeying the power law of distribution. The brain exhibits a broad range of flexible patterns of coordinated dynamics which decay after some critical value. While the NFVM initiates avalanches from arousal centers, a decentralized feedback mechanism arises spontaneously to lead a self-organized system from subcritical to supercritical phases. (B) A simulation of critical phase transitions on the 2D Ising spin lattice model exhibits different degrees of synchronization or coordination (black dots) from subcritical to supercritical states as temperature (a control parameter) increases from left to right. Adopted from Kitzbichler et al. (2009). Complexity as an entropy-based measure placed between two thermodynamic extrema: a perfect crystal at absolute zero temperature and an ideal gas. C𝒩 reflects a mixture of integration/segregation in brain dynamics with maximal values near criticality between subcritical and supercritical phases presented above.

Similarly, conscious experience is what has been just exposed to the psyche level by the brain at a given time; it is a passive marker of the deep neural processes going at hard and soft levels of brain dynamics. Now we can suggest a candidate measure to estimate the mental force of this fluctuating phenomenon, generally referred to as the level or quantity of consciousness. Recently, SOC has been proposed as a determinant for complexity measures of consciousness, mainly with respect to the concept of integrated information in IIT, based on the fact that both statistically reflect the same interplay between integration and segregation in network dynamics exhibiting scale-free patterning (Sporns, 2006; Lee et al., 2010; Tagliazucchi, 2017; Aguilera, 2019; Kim and Lee, 2019). While the critical dynamics are characterized by the order parameter, for example, a mean proportion of activated neurons in 𝒩, with the control parameter depending on connectivity density over time (Hesse and Gross, 2014), the complexity measures evaluate the degree of integration of 𝒩 in a particular state (Timme et al., 2016; Rosas et al., 2018).

In CET, the consciousness’ objective observables, derived from brain dynamics at discrete moments τ, will be the neural complexity measure C𝒩 (Tononi et al., 1994). This would be a more feasible measure, instead of its sophisticated version (integrated information Φ) that requires computing distance measures between probability distributions of a system and finding the minimum information partition (Oizumi et al., 2014). In IIT, Φ is determined by a major complex of irreducible causal mechanisms, which is then identified with NCC. The emphasis on the ontological status of Φ takes some form of panpsychism, and includes the possibility to ignite consciousness in deterministic machines (Tononi and Koch, 2015). Unlike IIT, in which the presence of consciousness itself is conditioned on integrated information a system is able to generate, CET determines the stream of consciousness by SOC in Equation (3). The complexity measure C𝒩 has no fundamental property. It is needed only as an objective (epistemic) observable to evaluate the mental force of brain dynamics at the given time without assuming machine consciousness or its ubiquity in matter like mass.

The main reason to accept C𝒩 as a quantitative measure of consciousness is that the notion of complexity is commonly conditioned on the same balance between order and disorder. In statistical physics, it starts by considering the perfect crystal and the isolated ideal gas as simple classical models of two extrema. The former represents a maximally ordered state of a system with entropy H = 0, and the latter symbolizes its maximally disordered state with H = Hmax, while both have zero complexity (Lòpez-Ruiz et al., 1995). Yet, just like the emergence of consciousness conditioned on large-scale brain dynamics with characteristic temporal scale Δt ≈ 100–450 ms in CET, the complexity is a scale-dependent concept. This cannot be measured uniformly because of the contextuality of statistical analysis itself. According to the third prerequisite (contextuality) in CET, only SOC with contributions of many brain regions can ignite consciousness.

The neural complexity CN (Tononi et al., 1994) starts with Shannon entropy H, defined on a set X which can occupy M states with probability pi, satisfying the condition p=i=1Mpi=1.

H(X)=i=1Mpilogpi(12)

Thus, H(X) = Hmax = log M if X is in equilibrium with ipi=1/M. On the contrary, X has the lowest entropy H(X) = 0 if X will occupy only a single state with p = 1. These both comprise the full spectrum of states in which the system X could be.

In CET, X represents the brain network 𝒩 which can occupy M = 2N states. Accordingly, the extrema can be applied to the two scenarios in cognitive brain evolution described above: the first with a frog iterating the same jump, and the second with the random number generator where all states from 2N configurations are equiprobable. However, it does not propose a quantitative measure of consciousness. To do it, the neural complexity focuses on the structural-functional connectivity of 𝒩. This is mathematically equivalent to the average information exchanged between subsets of a system and the rest of the system, summed over all subset sizes (ideally, beginning with a single neuron). The C𝒩 can be calculated by mutual information I obtained for all possible bipartitions of a system 𝒩 consisting of N elements,

C𝒩(𝒩)=k=1N/2I(𝒩jk;𝒩𝒩jk),(13)

where Νjk is the j’th bipartition running over all subsets of size k, and <· > stands for their average integration. The I is defined as

I(𝒩jk;𝒩𝒩jk)=H(𝒩jk)H(𝒩jk|𝒩𝒩jk)(14)

C𝒩 is highest when synchronization (integration) and desynchronization (segregation) tendencies are balanced in 𝒩, and lowest under either total synchronization (order) or total desynchronization (disorder) of its elements. In CET, neural complexity should provide a measure for information that was integrated by the brain during a time interval Δt. This displays how well the brain is poised near criticality to gain the optimum information over cognitive brain evolution. Thus, conscious states can emerge with different values of C𝒩 over the full spectrum of conscious states ranging from UWS to ADHD (as conditioned above).

The measure will reflect the magnitude of the brain’s mental force at the given time (Figure 2B). Formally, the level of consciousness in a particular state S(τ) equals to the neural complexity of its NCC presented by a corresponding vector xi. Unlike Φ in IIT, consciousness does not depend ontologically on C𝒩. The distinction can be formalized in the language of CET as follows.

IIT:S(τ)def¯¯maxΔtΦ(15)
CET:o(S(τ))=C𝒩(τ)(16)

where o is a physical observable of S(τ)def¯¯dψdτ at given moment τ.

An obvious limitation of C𝒩 is that the measure reflects the degree of short-term coherency of neural activity irrespectively to the type of cognitive processing recruited there. This can apply to disorders of consciousness (Mateos et al., 2018) such as UWS, but not to mental disorders that depend on long-term coherency of many conscious states over brain dynamics where other complexity (algorithmic) measures are preferable (Aboy et al., 2006; Hager et al., 2017; Murray et al., 2018). On the other hand, it can be proposed that the level of consciousness, typically defined by arousal criteria arranged from coma to full vigilance, is itself irrelevant to cognition. Mashour and Alkire (2013) argue that while humans exhibit the most advanced cognitive capabilities, including language, it is difficult to agree that a fully conscious human has a higher level of consciousness (alertness) than a beast in pursuit of prey.

Nevertheless, it seems undoubted that cognitive processing at the soft level of the brain must somehow affect conscious contents at the psyche level (Yurchenko, 2022). Humans can significantly differ in acquiring information from the same environment while preserving equal sensory abilities. It means that information gain is not affected by their capability to process sensory signals but depends only on how the signals are transformed into cognitive contents provided by brain dynamics (which do not still deviate from criticality). Accordingly, neural complexity should also differ between healthy humans. The difference should be especially significant in states of dementia, where the level of consciousness is preserved despite the dramatic disruption of cognitive abilities. This is a reason why C𝒩 cannot be a self-sufficient measure of conscious presence. The problem becomes yet more striking if we draw attention to the fly’s brain, which should apparently provide a very low quantity of C𝒩 compared to the human brain, maybe, even lower than in disorders of consciousness. Should we treat the fly unconscious?

Lamme (2018) recently pointed out that existing theories have missed a “key ingredient” of consciousness. One of its consequences is that many of them while being prone to superdeterminism are at the same time not immune to panpsychism. Thus, on one hand, they endow consciousness with causal power, i.e., free will. On the other hand, they make consciousness a scale-independent phenomenon that can emerge at the atomic or quantum level (Zeki, 2003; Hunt and Schooler, 2019), or that can be inherent to a single neuron or even to a mitochondrion as being ontologically conditioned on Φ (Tononi and Koch, 2015). Therefore, it is not surprising that these theories grant consciousness to future machines (Dehaene et al., 2017). How then should machine consciousness be compatible with free will controlled by hidden deterministic variables λ? Also, what would machine consciousness be if each atom already possessed some level of consciousness? CET is immune to both types of panpsychism: the emergence of consciousness is scale-dependent, whereas C𝒩 is only an epistemic measure of the mental force conditioned on SOC.

Yet, CET asserts that there is no conscious experience without arousal, one of the four principled features (plus SOC) of an arbitrary system should satisfy to be conscious. Recall, these features also comprise: (i) living properties; (ii) cognitive evolution for accumulating knowledge; and (iii) volition not predetermined from the past. In particular, dreams are often taken as evidence that consciousness can be separable from wakefulness (Hobson, 2009; Wamsley, 2013). CET considers dreams differently as the evidence that (passive and discrete) consciousness is complementary to unconscious (active and continuous) cognitive processing, which can persist during REM sleep by involving a parieto-occipital “hot zone” that is typically associated with the cortex-centered NCC (Siclari et al., 2017). CET argues that without activating arousal centers there can be no immediate awareness of something experienced.

As stated, conscious experience is always local in space and time. To be conscious means to be conscious of ourselves here and now while the great gift of imagination can allow us to virtually travel over space and time. Similarly, on retrospection, we can be conscious of what we were doing yesterday or many years ago, but it occurs just now, at the specious present (Varela, 1999). Importantly, we are not conscious while dreaming but we recall what the brain has unconsciously processed just before awakening. Dreams are residual (cortex-centered?) memory the brain can or cannot expose to the psyche level. Indeed, we do not retain dreams every time by awakening.

Yet, we know that all animals, even those endowed with a primitive neutral system, can be anesthetized (Zalucki and van Swinderen, 2016) but there is no “anesthesia” for arbitrary natural or artificial systems like atoms or machines. The arousal centers—even if very different in worms, insects, mollusks, and vertebrates—are that crucial ingredient (Lamme, 2018), without which neither consciousness nor any kind of free will is possible. Thus, consciousness, being inseparable from arousal, could have expanded over the animal kingdom due to the NFVM. Only the level of consciousness would have varied over species depending on the size and architecture of their brain capable of maintaining SOC.

Discussion

CET adheres to the evidence derived from Liber-type experiments that consciousness is a passive post-factum phenomenon of neural activity with no causal power over brain dynamics. Unfortunately, the free-will experiments have little to say about how the neural activity itself might be free of predetermination in unconscious processing of decision-making. The brain could still remain a deterministic machine (‘t Hooft, 2016; Hossenfelder and Palmer, 2020). Apparently, the only alternative is to adopt quantum indeterminism. The question is then: How to rigorously incorporate the quantum effects into classical stochastic brain dynamics? This problem has produced several quantum proposals in the science of consciousness. First, CET differs from all of them because it does not involve any macroscopic quantum effects that might arguably be exploited by consciousness to speed up the computational power of the brain, and make consciousness non-computable (Fisher, 2015) or even transiently separable from the brain due to closed causal loops (Hameroff and Penrose, 2014).

To CET, quantum indeterminism is not a matter of conscious will, but a matter of evolution of life and origins of consciousness. CET argues that consciousness has no free will but brain dynamics can be free of predetermination in a physically meaningful sense by the minimal use of quantum randomness via Beck-Eccles exocytosis. At the same time, Libet-type experiments are all initially cortex-centered by neglecting one obvious neurobiological fact. Wakefulness is a necessary prerequisite of any manipulations with consciousness to detect perception, attention, and cognition as compared to unconscious processing. Neither consciousness nor volition is possible without activating arousal centers.

First, CET rules out conscious will which allegedly might have power over brain dynamics at the soft (computational) level by deciding ahead of time what goal will be achieved (as described above in Alice’s scenario). Second, CET also rejects a statistical (probabilistic) account of volition resulting from noisy neural fluctuations in a stochastic framework (Khalighinejad et al., 2018) because such account of randomness depends ultimately on the state of our knowledge which by itself does not violate determinism. CET then takes the NFVM as a principled argument against superdeterminism, which is the hypothesis that conscious observers cannot act freely or do otherwise than what has been predetermined by deterministic laws of nature. This hypothesis has been suggested with the claim “to remove every single bit of mysticism from quantum theory” (‘t Hooft, 2016) such as a random wavefunction collapse caused by observation or nonlocal (faster than light) correlations between a pair of entangled particles. Nevertheless, contrary to this reasonable claim, superdeterminism would inevitably lead to much more mysterious implications such as “cosmic conspiracies” in Bell-type experiments (Gallicchio et al., 2014) or, more generally, a “designed” universe where conscious beings, controlled by hidden deterministic variables λ, do what the universe wants them to do (Yurchenko, 2021b).

Leaving aside these metaphysical issues, CET looks at this problem from an evolutionary perspective on the origins of life and consciousness. Might life have evolved from completely deterministic cause-effect interactions? What advantages would organisms have gained over non-living systems in following the same predetermined ways? Why should consciousness have then evolved from the simplest forms of life? Contrary to the idea that life is a triumph of determinism (Marshall et al., 2017), CET argues that organisms should initially be free of predetermination to have any meaningful difference from non-living systems to evolve from simplest to complex forms of life. In this way, the evolution of the brain would have been driven by the natural selection from primitive neuronal bundles to higher-level neural systems to supply their freedom to maneuver with computational power in an efficient prediction of the environmental events (Brembs, 2011).

This lays a principled division between the cause-effect interactions, reserved for dynamics of non-living systems of any complexity, and the stimulus-response repertoires accessible for all organisms. The repertoires can emerge only from the corresponding neural activity. A lesson we have learnt from biology is that every useful mechanism will be re-used by evolution to promote new more sophisticated mechanisms for adaptive advantages of life. The fast and random reflexes of the simplest organisms could have evolved from quantized micro-events that are abundantly presented in cellular structures of plants and animals (Arndt et al., 2009; Brookes, 2017). The evolution of the brain might have then used those reflexes to build up a neural mechanism like the NFVM in the same way as random (quantum in origin) gene mutations underlie the biological machinery that is responsible for the astonishing variety of species on the Earth. CET places the NFVM into the brainstem, a phylogenetically ancient brain region that comprises various nuclei responsible for executing many vital autonomic functions to maintain homeostasis.

Remarkably, those autonomic functions are typically called “involuntary” in the literature, while voluntary functions are implicitly reserved for conscious volition. The distinction between voluntary and involuntary functions is the main reason for confusion leading to postulating free will. Indeed, unlike reflex-like actions, voluntary actions are by definition those which were initiated consciously. On the other hand, volition must have by definition causal power. Thus, one naturally comes to free will via the distinction. In contrast, CET considers all autonomic and unconscious functions just voluntary, i.e., not predetermined from the past even in unicellular organisms placed upon the psyche-matter division between cause-effect interactions and stimulus-reaction repertoires (Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. The origins of consciousness. CET starts from the assumption that the brain should have primarily evolved as volitional subsystems of organisms from simplest neural reflexes based on chemoreceptor, magneto- and photoreceptor cells sensitive to quantum effects. At the causal (hard) level, the volitional subsystems should provide a principled psyche-matter division between organisms, exploiting their stimulus-reactions repertoires freely, and non-living systems, governed completely by cause-effects interactions. Placing the NFVM into the brainstem, the oldest brain region that integrates functions of many vital systems and is responsible for arousal and vigilance guarantees that each conscious state will be triggered free of predetermination from the past.

This requires a paradigm shift akin to that of the Copernican heliocentrism that had turned upside down the humans’ belief that the Sun revolves around the Earth as they indeed saw it in everyday experience. Similarly, humans believe they have free will to choose the course of future actions under given conditions. Note, CET does not univocally deny this ambiguous formulation. If humans are associated with what their brain does, the brain can indeed make a choice that had not been predetermined from the past. This scenario is possible on the condition that the choice is causally triggered at the hard level by the NFVM in the arousal centers of the brainstem. The quantized micro-event will then be amplified via neural scale-free avalanches and cognitively modulated at the soft level across thalamocortical systems in unconscious ways, and exposed ultimately to the psyche level.

On the contrary, CET completely denies the assumption that humans are associated with what their consciousness might do to their brain (and their body) by intervening in neural activity at the hard level. In the stream, particular conscious states are what the brain has periodically generated at critical points of brain dynamics as it explores its attractor space (yet bombarded by sensory signals). The only distinction between involuntary and voluntary actions depends on how the brain processes decision-making, which, in turn, depends on the cognitive complexity of a task. Simple or intermediate tasks as those modeled in multilayer predictive processing (Pezzulo et al., 2018) will be computed unconsciously. Only the ultimate decisions of complex tasks will be exposed to the psyche level as particular conscious states (which can in principle be measured as the brain’s mental force at a given moment).

Take for example breathing and heartbeat, both occurring without necessarily requiring consciousness. We can nonetheless stop breathing for some time by focusing our attention on this autonomic process. Advocates of free will take this as an evidence of the causal power of consciousness. They concede, however, that we are still unable to stop the heartbeat. Importantly, in CET, the stream of “pulsating” consciousness is itself like a heartbeat being every time initiated by the NFVM from arousal centers. At the same time, meditators trained in yoga practice are believed to be able to stop their heartbeat too (Krygier et al., 2013; Vinay et al., 2016). Does it mean that their free will is more powerful than that of ordinary humans? If so, does it also mean that free will weakens in humans with mental diseases such as obsessive-compulsive disorder or schizophrenia? Or is it all a matter of the cognitive effort and the corresponding state of the thalamocortical systems?

CET considers wrong saying that “A trained consciousness can control involuntary heartbeat at its will” or “A cognitively disrupted consciousness cannot well control even voluntary behavior at its will”. In both cases, the brain possesses consciousness as its mental force generated at a given moment of time. This is what produces the illusion of free will as if consciousness itself (like a homunculus) “strikes” the force from the brain to command future actions. Hence, all conscious states in the stream are already causally driven by the brain like breathing and heartbeat. They all can be called voluntary on the authority of the brain. For instance, the experience of “acting involuntarily” like suddenly reaching out to catch a falling object, or reflexively removing one’s body from a hot object may arise because such relatively simple actions can be produced at the hard level before its stimulus had been exposed to consciousness at the psyche level.

CET argues that the stream of consciousness (way of being) is initially NFVM-driven from arousal centers, and can be then derived—in any meaningful (formal or informal) sense—only from cognitive brain evolution (way of knowing). The brain continuously accumulates information to make an embodied choice, whereas unitary conscious states appear instantaneously at critical points of brain dynamics as ultimate decisions (snapshots) of unconscious action-oriented cognitive processing (Engel et al., 2013). The stream S(τ) cannot go on, though, it can persist like a freeze-frame on a TV screen if cognitive brain evolution stops and working memory has nothing to update, as it occurs in patients with UWS when brain dynamics depart from criticality (Tagliazucchi et al., 2016; Lee et al., 2019; Huang et al., 2020). Indeed, patients who recovered after UWS have no memories of their time in rehabilitation as if the stream was “frozen” all the time (Gosseries et al., 2014).

While the emergence of consciousness from brain dynamics is commonly accepted, many authors still ascribe to consciousness a special biological function and active role for selective attention, operant learning, and goal-directed behavior as if evolution had equipped the brain with consciousness just to perform complex volitional actions, which might not be performed unconsciously (Pennartz et al., 2019; Feinberg and Mallatt, 2020; Rolls, 2020). Again, CET does not univocally deny this conjecture. On one hand, consciousness has never been the goal of evolution but it has developed gradually from neural activities of organisms as a global byproduct of their volitional-cognitive mechanisms, however, with no causal power in brain dynamics. On the other hand, conscious states are exposed to the psyche level as ultimate decisions of complex cognitive tasks processed at that moment.

The laws of Nature make consciousness a necessary emergent phenomenon of brain dynamics in the same way as some quantity of H2O molecules (but not one or two) placed together will necessarily produce water. Turn now water into a neuronal substance, add SOC, a volitional (quantum in origin) mechanism, arousal centers, and learning for accumulating knowledge: these ingredients will spontaneously produce consciousness.

Does it mean that the hard problem of consciousness can be solved? Generally, nothing in CET prevents one from assuming that future AI systems endowed with these neuromorphic properties may one day become conscious. Nevertheless, these machines would be non-deterministic and, hence, as unpredictable as humans are. In other words, we might not circumvent Nature in order to create conscious automata like selfless lackeys. Moreover, such conscious and selfish machines might not be immune to disorders and psychoses such as “machine schizophrenia” or “machine depression.” CET argues that the hard problem as it is related to the search for NCC (Crick and Koch, 2003) can be better understood by linking it to the secured privacy of a particular consciousness. If a machine might know what it is like to be Alice, the uniqueness of her subjective experience would be uncovered and in principle copyable. Many machines might have the same qualia. In this sense, subjective qualia emerging phenomenally at the psyche level of Alice’s brain are a non-eliminable part of her (and only her) consciousness.

In a conventional view, humans possess free will due to conscious deliberation. The view seems to be maintained by plenty of experimental evidence showing that, in contrast to reflex (autonomic) actions, the cortical function is essential for volitional control of self-initiated actions (Libet, 1985; Haggard, 2008; Fried et al., 2011). However, an obvious thing is neglected in those findings: before executing conscious control in sensorimotor systems of the cortex, the brain should already maintain conscious experience due to arousal centers. The cortex-centered account of free will in Libet-type experiments loses its validity the moment one realizes that no kind of volition can be ascribed to a subject in coma after brainstem lesions (Laureys et al., 2004).

CET proposes a paradigm shift based on an evolutionary and physically rigorous perspective, contrary to the conventional view: humans possess consciousness due to a volitional (Bell-certified) mechanism. While rapid and random reflexes represent the most primitive preconscious form of volitional motor integration, consciousness is the highest phenomenon of global integration. In CET, consciousness is like a river buoy fluctuating on the surface of water regardless of its depth. On the other hand, precisely because of its representative and “superficial” nature in brain dynamics, the stream of consciousness is the marker the brain has exposed to its environment (including external observers) in a meaningful sense.

Conclusion

Assuming consciousness to be active in brain dynamics entails three intrinsically linked corollaries: it should be: (i) temporally continuous to implement its own (ii) conscious processing in parallel with unconscious processing. Brain dynamics would be therefore divided into two interdependent parts, a “highway” for conscious processing and an “underground” for unconscious processing, each requiring some (iii) neural correlates (workplace) for executing their causal function in the brain. Rejecting the active role of consciousness challenges these corollaries as well. First of all, CET operates exclusively on the stream of consciousness as a discrete chain of momentary states ignited at/near criticality without assuming any separate highway for conscious processing. Accordingly, what we experience as a state of perceptual presence has no special NCC in the brain but spreads over many brain regions as those currently involved in volition and cognition.

Moreover, consciousness is not only the phenomenal experience of something; it implies one’s self-awareness. Perhaps, the only biological function assigned to consciousness by evolution is self-awareness, i.e., the ability not only to be but also to have the sense of being. Consciousness gives the meaning of life to organisms—not philosophically but biologically: it is the evolutionary reward of the brain for making its work, over which the variety of all sensations and all values of being alive unfold. Life could not have succeeded on Earth if organisms did not appreciate these values.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Author Contributions

The author confirms being the sole contributor of this work and has approved it for publication.

Funding

I would like to thank Frontiers for funding the APC of this article.

Acknowledgments

I would like to thank the editor and two reviewers for their comments.

Conflict of Interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

  1. ^ Here, the mechanism is called “neurophysiological” to emphasize that CET has nothing to do with any kind of mind-brain dualism or psyche-matter complementarity, advocated by prominent scientists such as von Neumann, Pauli, Wheeler, Eccles, and Popper. CET does not also admit downward (top-down) causation, a covert version of dualism that might endow consciousness with causal power over the brain (Hoel et al., 2013; Mediano et al., 2022). Yet, the word “free” implies that causal power of volition must necessarily be free of predetermination. That is, CET does not adopt a compatibilist account of free will.
  2. ^ Traditionally, affective neuroscience holds that activation of evolutionary ancient subcortical regions is both necessary and sufficient for affective experience, whereas cognitive neuroscience argues that subcortical processes are necessary but not sufficient even for primitive experience. Remarkably, many theories of consciousness are conceived in the context of cognitive neuroscience to account—explicitly or implicitly—for active consciousness, i.e., free will that should be cortex-centered. In contrast, CET discards the active role of consciousness and maintains the claim of affective neuroscience.
  3. ^ Recall, in CET, discrete conscious states are exposed to the psyche level as the ultimate decision of cognitive (predictive and unconscious) processing at the soft level (Yurchenko, 2022). These decisions form the stream S(τ) in the physical time continuum with temporal resolution by Δt. For example, if placed between two almost identical images, excepting a tiny detail, consciousness perceives each as a whole image (gestalt) per state, gaining then an insight into the difference. The whole enterprise of making animation movies relies on this brain’s ability to capture very similar visual scenes (discrete frames about 24 per second) and to make a difference by comparing them, perceived then consciously over stream S(τ) as a dynamical picture of natural motion in time. This is the reason why the illusion of active consciousness and the illusion of continuous consciousness converge and make the use of the term “conscious processing” legitimate in the literature.

References

Aaronson, S. (2016). “The ghost in the quantum Turing machine,” in The Once and Future Turing: Computing the World, eds S. B. Cooper and A. Hodges (Cambridge: Cambridge University Press), 193–294.

Google Scholar

Aboy, M., Abasolo, D., Hornero, R., and Álvarez, D. (2006). Interpretation of the lempel-ziv complexity measure in the context of biomedical signal analysis. IEEE Trans. Biomed. Eng. 53, 2282–2288. doi: 10.1109/TBME.2006.883696

PubMed Abstract | CrossRef Full Text | Google Scholar

Aguilera, M. (2019). Scaling behaviour and critical phase transitions in integrated information theory. Entropy 21:1198. doi: 10.3390/e21121198

CrossRef Full Text | Google Scholar

Albantakis, L., Marshall, W., Hoel, E., and Tononi, G. (2019). What caused what? A quantitative account of actual causation using dynamical causal networks. Entropy (Basel) 21:459. doi: 10.3390/e21050459

PubMed Abstract | CrossRef Full Text | Google Scholar

Aleman, B., and Merker, B. (2014). Consciousness without cortex: a hydranencephaly family survey. Acta Paediatr. 103, 1057–1065. doi: 10.1111/apa.12718

PubMed Abstract | CrossRef Full Text | Google Scholar

Arndt, M., Juffmann, T., and Vedral, V. (2009). Quantum physics meets biology. HFSP J. 3, 386–400. doi: 10.2976/1.3244985

PubMed Abstract | CrossRef Full Text | Google Scholar

Aru, J., Suzuki, M., Rutiku, R., Larkum, M. E., and Bachmann, T. (2019). Coupling the state and contents of consciousness. Front. Syst. Neurosci. 13:43. doi: 10.3389/fnsys.2019.00043

PubMed Abstract | CrossRef Full Text | Google Scholar

Ashby, R. (1960). Design for a Brain. London: Chapman and Hall.

Google Scholar

Aspect, A., Dalibard, J., and Roger, G. (1982). Experimental test of bell’s inequalities using time-varying analyzers. Phys. Rev. Lett. 49:1804. doi: 10.1103/PhysRevLett.49.1804

CrossRef Full Text | Google Scholar

Atmanspacher, H., and beim Graben, P. (2007). Contextual emergence of mental states from neurodynamics. arXiv [Preprint]. doi: 10.48550/arXiv.q-bio/0512034

CrossRef Full Text | Google Scholar

Baars, B. J., Franklin, S., and Ramsoy, T. Z. (2013). Global workspace dynamics: cortical “binding and propagation” enables conscious contents. Front. Psychol. 4:200. doi: 10.3389/fpsyg.2013.00200

PubMed Abstract | CrossRef Full Text | Google Scholar

Bachmann, T., and Hudetz, A. G. (2014). It is time to combine the two main traditions in the research on the neural correlates of consciousness: C = L × D. Front. Psychol. 5:940. doi: 10.3389/fpsyg.2014.00940

PubMed Abstract | CrossRef Full Text | Google Scholar

Bak, P. (1996). How Nature Works: The Science of Self-Organized Criticality. New York, NY: Copernicus.

Google Scholar

Barron, A. B., and Klein, C. (2016). What insects can tell us about the origins of consciousness. Proc. Natl. Acad. Sci. U S A 113, 4900–4908. doi: 10.1073/pnas.1520084113

PubMed Abstract | CrossRef Full Text | Google Scholar

Barttfeld, P., Uhrig, L., Sitt, J. D., Sigman, M., Jarraya, B., and Dehaene, S. (2015). Signature of consciousness in the dynamics of resting-state brain activity. Proc. Natl. Acad. Sci. U S A 112, 887–892. doi: 10.1073/pnas.1418031112

PubMed Abstract | CrossRef Full Text | Google Scholar

Beck, F., and Eccles, J. (1992). Quantum aspects of brain activity and the role of consciousness. Proc. Natl. Acad. Sci. U S A 89, 11357–11361. doi: 10.1073/pnas.89.23.11357

PubMed Abstract | CrossRef Full Text | Google Scholar

Beck, F., and Eccles, J. (1998). Quantum processes in the brain: a scientific basis of consciousness. Cogn. Stud. 5, 95–109. doi: 10.11225/jcss.5.2_95

CrossRef Full Text | Google Scholar

Bedau, M. A., and Humphreys, P. E. (2008). Emergence: Contemporary Readings in Philosophy and Science. Cambridge, MA: MIT Press.

Google Scholar

Beggs, J. M., and Plenz, D. (2003). Neuronal avalanches in neocortical circuits. J. Neurosci. 23, 11167–11177. doi: 10.1523/JNEUROSCI.23-35-11167.2003

PubMed Abstract | CrossRef Full Text | Google Scholar

Beggs, J. M., and Timme, N. (2012). Being critical of criticality in the brain. Front. Physiol. 3:163. doi: 10.3389/fphys.2012.00163

PubMed Abstract | CrossRef Full Text | Google Scholar

Bell, J. (1993). Speakable and Unspeakable in Quantum Mechanics. Cambridge: Cambridge University Press.

Google Scholar

Bellay, T., Shew, W. L., Yu, S., Falco-Walter, J. J., and Plenz, D. (2021). Selective participation of single cortical neurons in neuronal avalanches. Front. Neural Circuits 14:620052. doi: 10.3389/fncir.2020.620052

PubMed Abstract | CrossRef Full Text | Google Scholar

Blanchard, P., Cessac, B., and Krueger, T. (2000). What can one learn about self-organized criticality from dynamical system theory? J. Stat. Phys. 98, 375–404. doi: 10.1023/A:1018639308981

CrossRef Full Text | Google Scholar

Boly, M., Massimini, M., Tsuchiya, N., Postle, B. R., Koch, C., and Tononi, G. (2017). Are the neural correlates of consciousness in the front or in the back of the cerebral cortex? Clinical and neuroimaging evidence. J. Neurosci. 37, 9603–9613. doi: 10.1523/JNEUROSCI.3218-16.2017

PubMed Abstract | CrossRef Full Text | Google Scholar

Breakspear, M. (2017). Dynamic models of large-scale brain activity. Nat. Neurosci. 20, 340–352. doi: 10.1038/nn.4497

PubMed Abstract | CrossRef Full Text | Google Scholar

Brembs, B. (2011). Towards a scientific concept of free will as a biological trait: spontaneous actions and decision-making in invertebrates. Proc. R. Soc. B 278, 930–939. doi: 10.1098/rspb.2010.2325

PubMed Abstract | CrossRef Full Text | Google Scholar

Bressler, S. L., and Kelso, J. A. S. (2001). Cortical coordination dynamics and cognition. Trends Cogn. Sci. 5, 26–36. doi: 10.1016/s1364-6613(00)01564-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Brette, R. (2019). Is coding a relevant metaphor for the brain? Behav. Brain Sci. 42:e215. doi: 10.1017/S0140525X19000049

PubMed Abstract | CrossRef Full Text | Google Scholar

Brookes, J. C. (2017). Quantum effects in biology: golden rule in enzymes, olfaction, photosynthesis and magnetodetection. Proc. R. Soc. A 473:20160822. doi: 10.1098/rspa.2016.0822

PubMed Abstract | CrossRef Full Text | Google Scholar

Chang, A. Y. C., Biehl, M., Yu, Y., and Kanai, R. (2020). Information closure theory of consciousness. Front. Psychol. 11:1504. doi: 10.3389/fpsyg.2020.01504

PubMed Abstract | CrossRef Full Text | Google Scholar

Chialvo, D. R. (2010). Emergent complex neural dynamics: the brain at the edge. Nat. Phys. 6, 744–750. doi: 10.1038/nphys1803

CrossRef Full Text | Google Scholar

Clark, A. (2013). Whatever next? Predictive brains, situated agents and the future of cognitive science. Behav. Brain Sci. 36, 181–204. doi: 10.1017/S0140525X12000477

PubMed Abstract | CrossRef Full Text | Google Scholar

Cleeremans, A. (2011). The radical plasticity thesis: how the brain learns to be conscious. Front. Psychol. 2:86. doi: 10.3389/fpsyg.2011.00086

PubMed Abstract | CrossRef Full Text | Google Scholar

Cocchi, L., Gollo, L. L., Zalesky, A., and Breakspear, M. (2017). Criticality in the brain: a synthesis of neurobiology, models and cognition. Prog. Neurobiol. 158, 132–152. doi: 10.1016/j.pneurobio.2017.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Cofré, R., Herzog, R., Mediano, P. A. M., Piccinini, J., Rosas, F. E., Sanz Perl, Y., et al. (2020). Whole-brain models to explore altered states of consciousness from the bottom up. Brain Sci. 10:626. doi: 10.3390/brainsci10090626

PubMed Abstract | CrossRef Full Text | Google Scholar

Conway, J., and Kochen, S. (2008). “The strong free will theorem,” in Deep Beauty: Understanding the Quantum World through Mathematical Innovation, ed H. Halvorson (Cambridge: Cambridge University Press), 443–454. doi: 10.1017/CBO9780511976971.014

CrossRef Full Text | Google Scholar

Crick, F., and Koch, C. (2003). A framework for consciousness. Nat. Neurosci. 6, 119–126. doi: 10.1038/nn0203-119

PubMed Abstract | CrossRef Full Text | Google Scholar

Dahmen, D., Grün, S., Diesmann, M., and Helias, M. (2019). Second type of criticality in the brain uncovers rich multiple-neuron dynamics. Proc. Natl. Acad. Sci. U S A 116, 13051–13060. doi: 10.1073/pnas.1818972116

PubMed Abstract | CrossRef Full Text | Google Scholar

Damasio, A. (2010). Self Comes to Mind: Constructing the Conscious Brain. London: William Heinemann.

Google Scholar

Damasio, A., Damasio, H., and Tranel, D. (2012). Persistence of feelings and sentience after bilateral damage of the insula. Cereb. Cortex 23, 833–846. doi: 10.1093/cercor/bhs077

PubMed Abstract | CrossRef Full Text | Google Scholar

Deco, G., Cruzat, J., and Kringelbach, M. L. (2019). Brain songs framework used for discovering the relevant timescale of the human brain. Nat. Commun. 10:583. doi: 10.1038/s41467-018-08186-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Deco, G., and Jirsa, V. K. (2012). Ongoing cortical activity at rest: criticality, multistability and ghost attractors. J. Neurosci. 32, 3366–3375. doi: 10.1523/JNEUROSCI.2523-11.2012

PubMed Abstract | CrossRef Full Text | Google Scholar

Deco, G., Tononi, G., Boly, M., and Kringelbach, M. L. (2015). Rethinking segregation and integration: contributions of whole-brain modelling. Nat. Rev. Neurosci. 16, 430–439. doi: 10.1038/nrn3963

PubMed Abstract | CrossRef Full Text | Google Scholar

Dehaene, S., and Changeux, J. P. (2011). Experimental and theoretical approaches to conscious processing. Neuron 70, 200–227. doi: 10.1016/j.neuron.2011.03.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Dehaene, S., Lau, H., and Kouider, S. (2017). What is consciousness and could machines have it? Science 358, 486–492. doi: 10.1126/science.aan8871

PubMed Abstract | CrossRef Full Text | Google Scholar

Dehaene, S., and Naccache, L. (2001). Towards a cognitive permanence of consciousness: basic evidence and a workspace framework. Cognition 79, 1–37. doi: 10.1016/s0010-0277(00)00123-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Del Pin, S. H., Skóra, Z., Sandberg, K., Overgaard, M., and Wierzchoń, M. (2021). Comparing theories of consciousness: why it matters and how to do it. Neurosci. Conscious. 2021:niab019. doi: 10.1093/nc/niab019

PubMed Abstract | CrossRef Full Text | Google Scholar

Demertzi, A., Tagliazucchi, E., Dehaene, S., Deco, G., Barttfeld, P., Raimondo, F., et al. (2019). Human consciousness is supported by dynamic complex patterns of brain signal coordination. Sci. Adv. 5:eaat7603. doi: 10.1126/sciadv.aat7603

PubMed Abstract | CrossRef Full Text | Google Scholar

Dirnberger, G., Hesselmann, G., Roiser, J. P., Preminger, S., Jahanshahi, M., and Paz, R. (2012). Give it time: neural evidence for distorted time perception and enhanced memory encoding in emotional situations. Neuroimage 63, 591–599. doi: 10.1016/j.neuroimage.2012.06.041

PubMed Abstract | CrossRef Full Text | Google Scholar

Doerig, A., Schurger, A., and Herzog, M. H. (2021). Hard criteria for empirical theories of consciousness. Cogn. Neurosci. 12, 41–62. doi: 10.1080/17588928.2020.1772214

PubMed Abstract | CrossRef Full Text | Google Scholar

Doyle, R. O. (2009). Free will: it’s a normal biological property, not a gift or a mystery. Nature 459:1052. doi: 10.1038/4591052c

CrossRef Full Text | Google Scholar

Drissi-Daoudi, L., Doerig, A., and Herzog, M. H. (2019). Feature integration within discrete time windows. Nat. Commun. 10:4901. doi: 10.1038/s41467-019-12919-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Edelman, G. M. (2003). Naturalizing consciousness: a theoretical framework. Proc. Natl. Acad. Sci. U S A 100, 5520–5524. doi: 10.1073/pnas.0931349100

PubMed Abstract | CrossRef Full Text | Google Scholar

Edlow, B. L., Claassen, J., Schiff, N. D., and Greer, D. M. (2020). Recovery from disorders of consciousness: mechanisms, prognosis and emerging therapies. Nat. Rev. Neurol. 17, 135–156. doi: 10.1038/s41582-020-00428-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Engel, A. K., Maye, A., Kurthen, M., and König, P. (2013). Where’s the action? The pragmatic turn in cognitive science. Trends Cogn. Sci. 17, 202–209. doi: 10.1016/j.tics.2013.03.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Feinberg, T. E., and Mallatt, J. (2020). Phenomenal consciousness and emergence: eliminating the explanatory gap. Front. Psychol. 11:1041. doi: 10.3389/fpsyg.2020.01041

PubMed Abstract | CrossRef Full Text | Google Scholar

Fekete, T., Van de Cruys, S., Ekroll, V., and van Leeuwen, C. (2018). In the interest of saving time: a critique of discrete perception. Neurosci. Conscious. 2018:niy003. doi: 10.1093/nc/niy003

PubMed Abstract | CrossRef Full Text | Google Scholar

Fields, C., Glazebrook, J. F., and Levin, M. (2021). Minimal physicalism as a scale-free substrate for cognition and consciousness. Neurosci. Conscious. 2021:niab013. doi: 10.1093/nc/niab013

PubMed Abstract | CrossRef Full Text | Google Scholar

Fingelkurts, A. A., Fingelkurts, A. A., and Neves, C. F. H. (2010). Machine consciousness and artificial thought: an operational architectonics model guided approach. Brain Res. 1428, 80–92. doi: 10.1016/j.brainres.2010.11.079

PubMed Abstract | CrossRef Full Text | Google Scholar

Fisher, M. P. (2015). Quantum cognition: the possibility of processing with nuclear spins in the brain. Ann. Phys. 362, 593–602. doi: 10.1016/j.aop.2015.08.020

CrossRef Full Text | Google Scholar

Freeman, W. J. (2006). A cinematographic hypothesis of cortical dynamics in perception. Int. J. Psychophysiol. 60, 149–161. doi: 10.1016/j.ijpsycho.2005.12.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Fried, I., Mukamel, R., and Kreiman, G. (2011). Internally generated preactivation of single neurons in human medial frontal cortex predicts volition. Neuron 69, 548–562. doi: 10.1016/j.neuron.2010.11.045

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K. J. (1994). Functional and effective connectivity in neuroimaging: a synthesis. Hum. Brain Mapp. 2, 56–78. doi: 10.1002/hbm.460020107

CrossRef Full Text | Google Scholar

Friston, K., and Herreros, I. (2016). Active inference and learning in the cerebellum. Neural. Comput. 28, 1812–1839. doi: 10.1162/NECO_a_00863

PubMed Abstract | CrossRef Full Text | Google Scholar

Friston, K., Schwartenbeck, P., FitzGerald, T., Moutoussis, M., Behrens, T., and Dolan, R. J. (2013). The anatomy of choice: active inference and agency. Front. Hum. Neurosci. 7:598. doi: 10.3389/fnhum.2013.00598

PubMed Abstract | CrossRef Full Text | Google Scholar

Gallicchio, J. S., Friedman, A. S., and Kaiser, D. I. (2014). Testing bell’s inequality with cosmic photons: closing the setting-independence loophole. Phys. Rev. Lett. 112:110405. doi: 10.1103/PhysRevLett.112.110405

CrossRef Full Text | Google Scholar

Georgiev, D. D. (2020). Quantum information theoretic approach to the mind-brain problem. Prog. Biophys. Mol. Biol. 158, 16–32. doi: 10.1016/j.pbiomolbio.2020.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Giacino, J. T., Fins, J. J., Laureys, S., and Schiff, N. D. (2014). Disorders of consciousness after acquired brain injury: the state of the science. Nat. Rev. Neurol. 10, 99–114. doi: 10.1038/nrneurol.2013.279

PubMed Abstract | CrossRef Full Text | Google Scholar

Gibb, S., Findlay, R. H., and Lancaster, T. (2019). The Routledge Handbook of Emergence. London; New York: Routledge.

Google Scholar

Golkowski, D., Larroque, S. K., Vanhaudenhuyse, A., Plenevaux, A., Boly, M., Di Perri, C., et al. (2019). Changes in whole brain dynamics and connectivity patterns during sevoflurane- and propofol-induced unconsciousness identified by functional magnetic resonance imaging. Anesthesiology 130, 898–911. doi: 10.1097/ALN.0000000000002704

PubMed Abstract | CrossRef Full Text | Google Scholar

Gollo, L. L., Mirasso, C., and Eguíluz, V. M. (2012). Signal integration enhances the dynamic range in neuronal systems. Phys. Rev. E 85:040902. doi: 10.1103/PhysRevE.85.040902

CrossRef Full Text | Google Scholar

Gosseries, O., Dim, H., Laureys, S., and Boly, M. (2014). Measuring consciousness in severely damaged brains. Annu. Rev. Neurosci. 37, 457–478. doi: 10.1146/annurev-neuro-062012-170339

PubMed Abstract | CrossRef Full Text | Google Scholar

Granger, C. W. J. (1969). Investigating causal relations by econometric models and crossspectral methods. Econometrica 37, 424–438. doi: 10.1017/CBO9780511753978.002

CrossRef Full Text | Google Scholar

Graziano, M. S. A., Guterstam, A., Bio, B. J., and Wilterson, A. I. (2020). Toward a standard model of consciousness: reconciling the attention schema, global workspace, higher-order thought and illusionist theories. Cogn. Neuropsychol. 37, 155–172. doi: 10.1080/02643294.2019.1670630

PubMed Abstract | CrossRef Full Text | Google Scholar

Grossberg, S. (2017). Towards solving the hard problem of consciousness: the varieties of brain resonances and the conscious experiences that they support. Neural Netw. 87, 38–95. doi: 10.1016/j.neunet.2016.11.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Guggisberg, A. G., and Mottaz, A. (2013). Timing and awareness of movement decisions: does consciousness really come too late? Front. Hum. Neurosci. 7:385. doi: 10.3389/fnhum.2013.00385

PubMed Abstract | CrossRef Full Text | Google Scholar

Hager, B., Yang, A. C., Brady, R., Meda, S., Clementz, B., Pearlson, G. D., et al. (2017). Neural complexity as a potential translational biomarker for psychosis. J. Affect. Disorder. 216, 89–99. doi: 10.1016/j.jad.2016.10.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Haggard, P. (2008). Human volition: towards a neuroscience of will. Nat. Rev. Neurosci. 9, 934–946. doi: 10.1038/nrn2497

PubMed Abstract | CrossRef Full Text | Google Scholar

Hahn, G., Petermann, T., Havenith, M. N., Shan, Y., Singer, W., Plenz, D., et al. (2010). Neuronal avalanches in spontaneous activity in vivo. J. Neurophysiol. 104, 3312–3322. doi: 10.1152/jn.00953.2009

PubMed Abstract | CrossRef Full Text | Google Scholar

Haimovici, A., Tagliazucchi, E., Balenzuela, P., and Chialvo, D. R. (2013). Brain organization into resting state networks emerges at criticality on a model of the human connectome. Phys. Rev. Lett. 110:178101. doi: 10.1103/PhysRevLett.110.178101

PubMed Abstract | CrossRef Full Text | Google Scholar

Haken, H. (2004). Synergetics. Introduction and Advanced Topics. Berlin: Springer.

Google Scholar

Haldeman, C., and Beggs, J. M. (2005). Critical branching captures activity in living neural networks and maximizes the number of metastable states. Phys. Rev. Lett. 94:058101. doi: 10.1103/PhysRevLett.94.058101

PubMed Abstract | CrossRef Full Text | Google Scholar

Hameroff, S. (2012). How quantum brain biology can rescue conscious free will. Front. Integr. Neurosci. 6:93. doi: 10.3389/fnint.2012.00093

PubMed Abstract | CrossRef Full Text | Google Scholar

Hameroff, S., and Penrose, R. (2014). Consciousness in the universe: a review of the ‘Orch OR’ theory. Phys. Life Rev. 11, 39–78. doi: 10.1016/j.plrev.2013.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Heiney, K., Ramstad, O. H., Fiskum, V., Christiansen, N., Sandvig, A., Nichele, S., et al. (2021). Criticality, connectivity and neural disorder: a multifaceted approach to neural computation. Front. Comput. Neurosci. 15:611183. doi: 10.3389/fncom.2021.611183

PubMed Abstract | CrossRef Full Text | Google Scholar

Herzog, M. H., Kammer, T., and Scharnowski, F. (2016). Time slices: what is the duration of a percept? PLoS Biol. 14:e1002433. doi: 10.1371/journal.pbio.1002433

PubMed Abstract | CrossRef Full Text | Google Scholar

Hesse, J., and Gross, T. (2014). Self-organized criticality as a fundamental property of neural systems. Front. Syst. Neurosci. 8:166. doi: 10.3389/fnsys.2014.00166

PubMed Abstract | CrossRef Full Text | Google Scholar

Hiscock, H. G., Worster, S., Kattnig, D. R., Steers, C., Jin, Y., Manolopoulos, D. E., et al. (2016). The quantum needle of the avian magnetic compass. Proc. Natl. Acad. Sci. U S A 113, 4634–4639. doi: 10.1073/pnas.1600341113

PubMed Abstract | CrossRef Full Text | Google Scholar

Hobson, J. A. (2009). REM sleep and dreaming: towards a theory of protoconsciousness. Nat. Rev. Neurosci. 10, 803–813. doi: 10.1038/nrn2716

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoel, E., Albantakis, L., and Tononi, G. (2013). Quantifying causal emergence shows that macro can beat micro. Proc. Natl. Acad. Sci. U S A 110, 19790–19795. doi: 10.1073/pnas.1314922110

PubMed Abstract | CrossRef Full Text | Google Scholar

Hohwy, J. (2009). The neural correlates of consciousness: new experimental approaches needed? Conscious. Cogn. 18, 428–438. doi: 10.1016/j.concog.2009.02.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Hossenfelder, S., and Palmer, T. (2020). Rethinking superdeterminism. Front. Phys. 8:139. doi: 10.3389/fphy.2020.00139

CrossRef Full Text | Google Scholar

Huang, Z., Zhang, J., Wu, J., Mashour, G. A., and Hudetz, A. G. (2020). Temporal circuit of macroscale dynamic brain activity supports human consciousness. Sci. Adv. 6:eaaz0087. doi: 10.1126/sciadv.aaz0087

PubMed Abstract | CrossRef Full Text | Google Scholar

Hudetz, A. G., Humphries, C. J., and Binder, J. R. (2014). Spin-glass model predicts metastable brain states that diminish in anesthesia. Front. Syst. Neurosci. 8:234. doi: 10.3389/fnsys.2014.00234

PubMed Abstract | CrossRef Full Text | Google Scholar

Hunt, T., and Schooler, J. W. (2019). The easy part of the hard problem: a resonance theory of consciousness. Front. Hum. Neurosci. 13:378. doi: 10.3389/fnhum.2019.00378

PubMed Abstract | CrossRef Full Text | Google Scholar

James, W. (1950). The Principles of Psychology. Vol. I. New York, NY: Dover Publication Inc.

Google Scholar

Jensen, H. J. (1998). Self-Organized Criticality. Cambridge, UK: Cambridge University Press.

Google Scholar

Kelso, J. S. (1995). Dynamic Patterns: The Self-Organization of Brain and Behavior. Cambridge, MA: MIT Press.

Google Scholar

Kelso, J. A. S. (2021). Unifying large- and small-scale theories of coordination. Entropy (Basel) 23:537. doi: 10.3390/e23050537

PubMed Abstract | CrossRef Full Text | Google Scholar

Kent, L., and Wittmann, M. (2021). Time consciousness: the missing link in theories of consciousness. Neurosci. Conscious. 2021:niab011. doi: 10.1093/nc/niab011

PubMed Abstract | CrossRef Full Text | Google Scholar

Khalighinejad, N., Schurger, A., Desantis, A., Zmigrod, L., and Haggard, P. (2018). Precursor processes of human self-initiated action. Neuroimage 165, 35–47. doi: 10.1016/j.neuroimage.2017.09.057

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, H., and Lee, U. (2019). Criticality as a Determinant of integrated information Φ in human brain networks. Entropy (Basel) 21:981. doi: 10.3390/e21100981

CrossRef Full Text | Google Scholar

Kitzbichler, M. G., Smith, M. L., Christensen, S. R., and Bullmore, E. (2009). Broadband criticality of human brain network synchronization. PLoS Comput. Biol. 5:e1000314. doi: 10.1371/journal.pcbi.1000314

PubMed Abstract | CrossRef Full Text | Google Scholar

Klein, C., Hohwy, J., and Bayne, T. (2020). Explanation in the science of consciousness: from the neural correlates of consciousness (NCCs) to the difference makers of consciousness (DMCs). Philos. Mind Sci. 1:4. doi: 10.33735/phimisci.2020.II.60

CrossRef Full Text | Google Scholar

Knill, D. C., and Pouget, A. (2004). The bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci. 27, 712–719. doi: 10.1016/j.tins.2004.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Koch, C. (2009). “Free will, physics, biology and the brain,” in Downward Causation and the Neurobiology of Free Will, eds N. Murphy, G. F. R. Ellis, and T. O’Connor (Berlin: Springer), 31–52. doi: 10.1007/978-3-642-03205-9_2

CrossRef Full Text | Google Scholar

Koch, C., Massimini, M., Boly, M., and Tononi, G. (2016). Neural correlates of consciousness: progress and problems. Nat. Rev. Neurosci. 17, 307–321. doi: 10.1038/nrn.2016.22

PubMed Abstract | CrossRef Full Text | Google Scholar

Koch, C., and Tsuchiya, N. (2007). Attention and consciousness: two distinct brain processes. Trends Cogn. Sci. 11, 16–22. doi: 10.1016/j.tics.2006.10.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Kozma, R., and Freeman, W. J. (2017). Cinematic operation of the cerebral cortex interpreted via critical transitions in self-organized dynamic systems. Front. Syst. Neurosci. 11:10. doi: 10.3389/fnsys.2017.00010

PubMed Abstract | CrossRef Full Text | Google Scholar

Kraikivski, P. (2020). Systems of oscillators designed for a specific conscious percept. New Math. Nat. Comput. 16, 73–88. doi: 10.1142/S1793005720500052

CrossRef Full Text | Google Scholar

Kraikivski, P. A. (2022). Dynamic mechanistic model of perceptual binding. Mathematics 10:1135. doi: 10.3390/math10071135

CrossRef Full Text | Google Scholar

Krygier, J. R., Heathers, J. A. J., Shahrestani, S., Abbott, M., Gross, J. J., and Kemp, A. H. (2013). Mindfulness meditation, well-being and heart rate variability: a preliminary investigation into the impact of intensive Vipassana meditation. Int. J. Psychophysiol. 89, 305–313. doi: 10.1016/j.ijpsycho.2013.06.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Lamme, V. A. F. (2006). Towards a true neural stance on consciousness. Trends Cogn. Sci 10, 494–501. doi: 10.1016/j.tics.2006.09.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Lamme, V. A. F. (2018). Challenges for theories of consciousness: seeing or knowing, the missing ingredient and how to deal with panpsychism. Philos. Trans. R. Soc. Lond. B Biol. Sci. 373:20170344. doi: 10.1098/rstb.2017.0344

PubMed Abstract | CrossRef Full Text | Google Scholar

Lau, H. C. (2009). “Volition and the function of consciousness,” in Downward Causation and the Neurobiology of Free Will, eds N. Murphy, G. F. R. Ellis, and T. O’Connor (Berlin: Springer), 153–169. doi: 10.1007/978-3-642-03205-9_9

CrossRef Full Text | Google Scholar

Lau, H. C., and Rosenthal, D. (2011). Empirical support for higher-order theories of conscious awareness. Trends Cogn. Sci. 15, 365–373. doi: 10.1016/j.tics.2011.05.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Laureys, S., Owen, A. M., and Schiff, N. D. (2004). Brain function in coma, vegetative state and related disorders. Lancet Neurol. 3, 537–546. doi: 10.1016/S1474-4422(04)00852-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, H., Golkowski, D., Jordan, D., Berger, S., Ilg, R., Lee, J., et al. (2019). Relationship of critical dynamics, functional connectivity and states of consciousness in large-scale human brain networks. Neuroimage 188, 228–238. doi: 10.1016/j.neuroimage.2018.12.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, U., Oh, G., Kim, S., Noh, G., Choi, B., and Mashour, G. A. (2010). Brain networks maintain a scale-free organization across consciousness, anesthesia and recovery: evidence for adaptive reconfiguration. Anesthesiology 113, 1081–1091 doi: 10.1097/ALN.0b013e3181f229b5

PubMed Abstract | CrossRef Full Text | Google Scholar

Libet, B. (1985). Unconscious cerebral initiative and the role of conscious will in voluntary action. Behav. Brain Sci. 8, 529–566. doi: 10.1017/S0140525X00044903

CrossRef Full Text | Google Scholar

Lòpez-Ruiz, R., Mancini, H. L., and Calbet, X. (1995). A statistical measure of complexity. Phys. Let. A 209, 321–326. doi: 10.1016/0375-9601(95)00867-5

CrossRef Full Text | Google Scholar

Mallatt, J. A. (2021). Traditional scientific perspective on the integrated information theory of consciousness. Entropy (Basel) 23:650. doi: 10.3390/e23060650

PubMed Abstract | CrossRef Full Text | Google Scholar

Marshall, W., Kim, H., Walker, S. I., Tononi, G., and Albantakis, L. (2017). How causal analysis can reveal autonomy in models of biological systems. Philos. Trans. A Math. Phys. Eng. Sci. 375:20160358. doi: 10.1098/rsta.2016.0358

PubMed Abstract | CrossRef Full Text | Google Scholar

Mashour, G. A., and Alkire, M. T. (2013). Evolution of consciousness: phylogeny, ontogeny and emergence from general anesthesia. Proc. Natl. Acad. Sci. U S A 110, 10357–10364. doi: 10.1073/pnas.1301188110

PubMed Abstract | CrossRef Full Text | Google Scholar

Mashour, G. A., Roelfsema, P., Changeux, J. P., and Dehaene, S. (2020). Conscious processing and the global neuronal workspace hypothesis. Neuron 105, 776–798. doi: 10.1016/j.neuron.2020.01.026

PubMed Abstract | CrossRef Full Text | Google Scholar

Mateos, D. M., Guevara Erra, R., Wennberg, R., and Perez Velazquez, J. L. (2018). Measures of entropy and complexity in altered states of consciousness. Cogn. Neurodyn. 12, 73–84. doi: 10.1007/s11571-017-9459-8

PubMed Abstract | CrossRef Full Text | Google Scholar

McFadden, J., and Al-Khalili, J. (2018). The origins of quantum biology. Proc. Math. Phys. Eng. Sci. 474:20180674. doi: 10.1098/rspa.2018.0674

PubMed Abstract | CrossRef Full Text | Google Scholar

Mediano, P. A. M., Farah, J. C., and Shanahan, M. P. (2016). Integrated information and metastability in systems of coupled oscillators. arXiv [Preprint]. doi: 10.48550/arXiv.1606.08313

CrossRef Full Text | Google Scholar

Mediano, P. A. M., Rosas, F. E., Luppi, A. I., Jensen, H. J., Seth, A. K., Barrett, A. B., et al. (2022). Greater than the parts: a review of the information decomposition approach to causal emergence. Philos. Trans. A Math. Phys. Eng. Sci. 380:20210246. doi: 10.1098/rsta.2021.0246

PubMed Abstract | CrossRef Full Text | Google Scholar

Merchant, H., Harrington, D. L., and Meck, W. H. (2013). Neural basis of the perception and estimation of time. Annu. Rev. Neurosci. 36, 313–336. doi: 10.1146/annurev-neuro-062012-170349

PubMed Abstract | CrossRef Full Text | Google Scholar

Merker, B. (2007). Consciousness without a cerebral cortex: a challenge for neuroscience and medicine. Behav. Brain Sci. 30, 63–134. doi: 10.1017/S0140525X07000891

PubMed Abstract | CrossRef Full Text | Google Scholar

Mhuircheartaigh, R. N., Rosenorn-Lanng, D., Wise, R., Jbabdi, S., Rogers, R., and Tracey, I. (2010). Cortical and subcortical connectivity changes during decreasing levels of consciousness in humans: a functional magnetic resonance imaging study using propofol. J. Neurosci. 30, 9095–9102. doi: 10.1523/JNEUROSCI.5516-09.2010

PubMed Abstract | CrossRef Full Text | Google Scholar

Miller, S. M. (2007). On the correlation/constitution distinction problem (and other hard problems) in the scientific study of consciousness. Acta Neuropsychiatr. 19, 159–176. doi: 10.1111/j.1601-5215.2007.00207.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Murray, J. D., Demirtaş, M., and Anticevic, A. (2018). Biophysical modeling of large-scale brain dynamics and applications for computational psychiatry. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 3, 777–787. doi: 10.1016/j.bpsc.2018.07.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Naccache, L. (2018). Why and how access consciousness can account for phenomenal consciousness. Philos. Trans. R. Soc. Lond B Biol. Sci. 373:20170357. doi: 10.1098/rstb.2017.0357

PubMed Abstract | CrossRef Full Text | Google Scholar

Nani, A., Manuello, J., Mancuso, L., Liloia, D., Costa, T., and Cauda, F. (2019). The neural correlates of consciousness and attention: two sister processes of the brain. Front. Neurosci. 13:1169. doi: 10.3389/fnins.2019.01169

PubMed Abstract | CrossRef Full Text | Google Scholar

Niikawa, T., Miyahara, K., Hamada, H. T., and Nishida, S. (2022). Functions of consciousness: conceptual clarification. Neurosci. Conscious. 2022:niac006. doi: 10.1093/nc/niac006

PubMed Abstract | CrossRef Full Text | Google Scholar

Northoff, G., and Lamme, V. (2020). Neural signs and mechanisms of consciousness: is there a potential convergence of theories of consciousness in sight? Neurosci. Biobehav. Rev. 118, 568–587. doi: 10.1016/j.neubiorev.2020.07.019

PubMed Abstract | CrossRef Full Text | Google Scholar

Northoff, G., and Zilio, F. (2022). Temporo-spatial theory of consciousness (TTC) - bridging the gap of neuronal activity and phenomenal states. Behav. Brain Res. 424:113788. doi: 10.1016/j.bbr.2022.113788

PubMed Abstract | CrossRef Full Text | Google Scholar

Oizumi, M., Albantakis, L., and Tononi, G. (2014). From the phenomenology to the mechanisms of consciousness: integrated information theory 3.0. PLoS Comput. Biol. 10:e1003588. doi: 10.1371/journal.pcbi.1003588

PubMed Abstract | CrossRef Full Text | Google Scholar

Panksepp, J., Lane, R. D., Solms, M., and Smith, R. (2017). Reconciling cognitive and affective neuroscience perspectives on the brain basis of emotional experience. Neurosci. Biobehav. Rev. 76, 187–215. doi: 10.1016/j.neubiorev.2016.09.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Panksepp, J., Normansell, L., Cox, J. F., and Siviy, S. M. (1994). Effects of neonatal decortication on the social play of juvenile rats. Physiol. Behav. 56, 429–443. doi: 10.1016/0031-9384(94)90285-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Parr, T., Sajid, N., and Friston, K. J. (2020). Modules or mean-fields? Entropy (Basel) 22:552. doi: 10.3390/e22050552

PubMed Abstract | CrossRef Full Text | Google Scholar

Parvizi, J., and Damasio, A. (2001). Consciousness and the brainstem. Cognition 79, 135–160. doi: 10.1016/s0010-0277(00)00127-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Pauen, M. (2021). Why NCC research is not theory-neutral. Philos. Mind Sci. 2:10. doi: 10.33735/phimisci.2021.9188

CrossRef Full Text | Google Scholar

Paus, T. (2000). Functional anatomy of arousal and attention systems in the human brain. Prog. Brain Res. 126, 65–77. doi: 10.1016/S0079-6123(00)26007-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Pennartz, C. M. A., Farisco, M., and Evers, K. (2019). Indicators and criteria of consciousness in animals and intelligent machines: an inside-out approach. Front. Syst. Neurosci. 13:25. doi: 10.3389/fnsys.2019.00025

PubMed Abstract | CrossRef Full Text | Google Scholar

Pezzulo, G., Rigoli, F., and Friston, K. (2018). Hierarchical active inference: a theory of motivated control. Trends Cogn. Sci. 22, 294–306. doi: 10.1016/j.tics.2018.01.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Reid, A. T., Headley, D. W., Mill, R. D., Sanchez-Romero, R., Uddin, L. Q., Marinazzo, D., et al. (2019). Advancing functional connectivity research from association to causation. Nat. Neurosci. 22, 1751–1760. doi: 10.1038/s41593-019-0510-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Revonsuo, A., and Newman, J. (1999). Binding and consciousness. Conscious. Cogn. 8, 123–127. doi: 10.1006/ccog.1999.0393

PubMed Abstract | CrossRef Full Text | Google Scholar

Ritz, T. (2011). Quantum effects in biology: bird navigation. Procedia Chem. 3, 262–275. doi: 10.1016/j.proche.2011.08.034

CrossRef Full Text | Google Scholar

Rocha, R. P., Koçillari, L., Suweis, S., Grazia, M. D. F. D., de Schotten, M. T., Zorzi, M., et al. (2022). Recovery of neural dynamics criticality in personalized whole-brain models of stroke. Nat. Commun. 13:3683. doi: 10.1038/s41467-022-30892-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Rolls, E. T. (2012). Willed action, free will and the stochastic neurodynamics of decision-making. Front. Integr. Neurosci. 6:68. doi: 10.3389/fnint.2012.00068

PubMed Abstract | CrossRef Full Text | Google Scholar

Rolls, E. T. (2020). Neural computations underlying phenomenal consciousness: a higher order syntactic thought theory. Front. Psychol. 11:65. doi: 10.3389/fpsyg.2020.00655

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosas, F., Mediano, P. A. M., Ugarte, M., and Jensen, H. J. (2018). An information-theoretic approach to self-organisation: emergence of complex interdependencies in coupled dynamical systems. Entropy (Basel) 20:793. doi: 10.3390/e20100793

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenthal, D. M. (2008). Consciousness and its function. Neuropsychologia 46, 829–840. doi: 10.1016/j.neuropsychologia.2007.11.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Rudrauf, D., Bennequin, D., Granic, I., Landini, G., Friston, K., and Williford, K. (2017). A mathematical model of embodied consciousness. J. Theor. Biol. 428, 106–131. doi: 10.1016/j.jtbi.2017.05.032

PubMed Abstract | CrossRef Full Text | Google Scholar

Safron, A. (2020). An integrated world modeling theory (IWMT) of consciousness: combining integrated information and global neuronal workspace theories with the free energy principle and active inference framework; toward solving the hard problem and characterizing agentic causation. Front. Artif. Intell. 3:30. doi: 10.3389/frai.2020.00030

PubMed Abstract | CrossRef Full Text | Google Scholar

Salmi, J., Metwaly, M., Tohka, J., Alho, K., Leppämäki, S., Tani, P., et al. (2020). ADHD desynchronizes brain activity during watching a distracted multi-talker conversation. Neuroimage 216:116352. doi: 10.1016/j.neuroimage.2019.116352

PubMed Abstract | CrossRef Full Text | Google Scholar

Schiff, N. D., Nauvel, T., and Victor, J. D. (2014). Large-scale brain dynamics in disorders of consciousness. Curr. Opin. Neurobiol. 25, 7–14. doi: 10.1016/j.conb.2013.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Schmahmann, J. D. (2010). The role of the cerebellum in cognition and emotion. Neuropsychol. Rev. 20, 236–260. doi: 10.1007/s11065-010-9142-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Schreiber, T. (2000). Measuring information transfer. Phys. Rev. Lett. 85, 461–464. doi: 10.1103/PhysRevLett.85.461

CrossRef Full Text | Google Scholar

Schultze-Kraft, M., Birman, D., Rusconi, M., Allefeld, C., Görgen, K., Dähne, S., et al. (2016). The point of no return in vetoing self-initiated movements. Proc. Natl. Acad. Sci. U S A 113, 1080–1085. doi: 10.1073/pnas.1513569112

PubMed Abstract | CrossRef Full Text | Google Scholar

Schurger, A., Sitt, J. D., and Dehaene, S. (2012). An accumulator model for spontaneous neural activity prior to self-initiated movement. Proc. Natl. Acad. Sci. U S A 109, E2904–E2913. doi: 10.1073/pnas.1210467109

PubMed Abstract | CrossRef Full Text | Google Scholar

Seth, A. K. (2008). Causal networks in simulated neural systems. Cogn. Neurodyn. 2, 49–64. doi: 10.1007/s11571-007-9031-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Seth, A. K., and Baars, B. J. (2005). Neural Darwinism and consciousness. Conscious. Cogn. 14, 140–168. doi: 10.1016/j.concog.2004.08.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Seth, A. K., and Hohwy, J. (2021). Predictive processing as an empirical theory for consciousness science. Cogn. Neurosci. 12, 89–90. doi: 10.1080/17588928.2020.1838467

PubMed Abstract | CrossRef Full Text | Google Scholar

Shapiro, K. L., Raymond, J. E., and Arnell, K. M. (1997). The attentional blink. Trends Cogn. Sci. 1, 291–296. doi: 10.1016/S1364-6613(97)01094-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Shea, N., and Frith, C. D. (2019). The global workspace needs metacognition. Trends Cogn. Sci. 23, 560–571. doi: 10.1016/j.tics.2019.04.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Shew, W. L., Yang, H., Yu, S., Roy, R., and Plenz, D. (2011). Information capacity and transmission are maximized in balanced cortical networks with neuronal avalanches. J. Neurosci. 31, 55–63. doi: 10.1523/JNEUROSCI.4637-10.2011

PubMed Abstract | CrossRef Full Text | Google Scholar

Siclari, F., Baird, B., Perogamvros, L., Bernardi, G., LaRocque, J. J., Riedner, B., et al. (2017). The neural correlates of dreaming. Nat. Neurosci. 20, 872–878. doi: 10.1038/nn.4545

PubMed Abstract | CrossRef Full Text | Google Scholar

Signorelli, C. M., Szczotka, J., and Prentner, R. (2021). Explanatory profiles of models of consciousness - towards a systematic classification. Neurosci. Conscious. 2021:niab021. doi: 10.1093/nc/niab021

PubMed Abstract | CrossRef Full Text | Google Scholar

Soon, C. S., He, A. H., Bode, S., and Haynes, J. D. (2013). Predicting free choices for abstract intentions. Proc. Natl. Acad. Sci. U S A 110, 6217–6222. doi: 10.1073/pnas.1212218110

PubMed Abstract | CrossRef Full Text | Google Scholar

Sporns, O. (2006). Small-world connectivity, motif composition and complexity of fractal neuronal connections. Biosystems 85, 55–64. doi: 10.1016/j.biosystems.2006.02.008

PubMed Abstract | CrossRef Full Text | Google Scholar

‘t Hooft, G. (2016). The Cellular Automaton Interpretation of Quantum Mechanics. Berlin: Springer. doi: 10.48550/arXiv.1405.1548

CrossRef Full Text | Google Scholar

Tagliazucchi, E. (2017). The signatures of conscious access and its phenomenology are consistent with large-scale brain communication at criticality. Conscious. Cogn. 55, 136–147. doi: 10.1016/j.concog.2017.08.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Tagliazucchi, E., Chialvo, D. R., Siniatchkin, M., Amico, E., Brichant, J. F., Bonhomme, V., et al. (2016). Large-scale signatures of unconsciousness are consistent with a departure from critical dynamics. J. R. Soc. Interface 13:20151027. doi: 10.1098/rsif.2015.1027

PubMed Abstract | CrossRef Full Text | Google Scholar

Tegmark, M. (2015). Consciousness as a state of matter. Chaos Solit. Fract. 76, 238–270. doi: 10.1016/j.chaos.2015.03.014

CrossRef Full Text | Google Scholar

Timme, N. M., Marshall, N. J., Bennett, N., Ripp, M., Lautzenhiser, E., and Beggs, J. M. (2016). Criticality maximizes complexity in neural tissue. Front. Physiol. 7:425. doi: 10.3389/fphys.2016.00425

PubMed Abstract | CrossRef Full Text | Google Scholar

Tognoli, E., and Kelso, J. A. (2014). The metastable brain. Neuron 81, 35–48. doi: 10.1016/j.neuron.2013.12.022

PubMed Abstract | CrossRef Full Text | Google Scholar

Tononi, G. (2008). Consciousness as integrated information: a provisional manifesto. Biol. Bull. 215, 216–242. doi: 10.2307/25470707

PubMed Abstract | CrossRef Full Text | Google Scholar

Tononi, G. (2013). “On the irreducibility of consciousness and its relevance to free will,” in Is Science Compatible with Free Will?, eds A. Suarez and P. Adams (NY: Springer), 147–176. doi: 10.1007/978-1-4614-5212-6_11

CrossRef Full Text | Google Scholar

Tononi, G., Albantakis, L., Boly, M., Cirelli, G., and Koch, C. (2022). Only what exists can cause: an intrinsic view of free will. arXiv [Preprint]. doi: 10.48550/arXiv.2206.02069

CrossRef Full Text | Google Scholar

Tononi, G., and Koch, C. (2015). Consciousness: here, there and everywhere? Philos. Trans. R. Soc. Lond. B Biol. Sci. 370:20140167. doi: 10.1098/rstb.2014.0167

PubMed Abstract | CrossRef Full Text | Google Scholar

Tononi, G., Sporns, O., and Edelman, G. M. (1994). A measure for brain complexity: relating functional segregation and integration in the nervous system. Proc. Natl. Acad. Sci. U S A 91, 5033–5037. doi: 10.1073/pnas.91.11.5033

PubMed Abstract | CrossRef Full Text | Google Scholar

VanRullen, R., and Kanai, R. (2021). Deep learning and the global workspace theory. Trends Neurosci. 44, 692–704. doi: 10.1016/j.tins.2021.04.005

PubMed Abstract | CrossRef Full Text | Google Scholar

VanRullen, R., Zoefel, B., and Ilhan, B. (2014). On the cyclic nature of perception in vision versus audition. Philos. Trans. R. Soc. Lond. B Biol. Sci. 369:20130214. doi: 10.1098/rstb.2013.0214

PubMed Abstract | CrossRef Full Text | Google Scholar

Varela, F. J. (1995). Resonant cell assemblies: a new approach to cognitive functions and neuronal synchrony. Biol. Res. 28, 81–95.

PubMed Abstract | Google Scholar

Varela, F. J. (1999). “The specious present: a neurophenomenology of time consciousness,” in Naturalizing Phenomenology, eds J. Petitot, F. J. Varela, B. J. Pachoud and J.-M. Roy (Stanford, CA: Stanford University Press), 266–314.

Google Scholar

Vinay, A. V., Venkatesh, D., and Ambarish, V. (2016). Impact of short-term practice of yoga on heart rate variability. Int. J. Yoga 9, 62–66. doi: 10.4103/0973-6131.171714

PubMed Abstract | CrossRef Full Text | Google Scholar

Vohryzek, J., Cabral, J., Vuust, P., Deco, G., and Kringelbach, M. L. (2022). Understanding brain states across spacetime informed by whole-brain modelling. Philos. Trans. A Math. Phys. Eng. Sci. 380:20210247. doi: 10.1098/rsta.2021.0247

PubMed Abstract | CrossRef Full Text | Google Scholar

Wamsley, E. J. (2013). Dreaming, waking conscious experience and the resting brain: report of subjective experience as a tool in the cognitive neurosciences. Front. Psychol. 4:637. doi: 10.3389/fpsyg.2013.00637

PubMed Abstract | CrossRef Full Text | Google Scholar

Weichwald, S., and Peters, J. (2021). Causality in cognitive neuroscience: concepts, challenges and distributional robustness. J. Cogn. Neurosci. 33, 226–247. doi: 10.1162/jocn_a_01623

PubMed Abstract | CrossRef Full Text | Google Scholar

Werner, G. (2009). Viewing brain processes as critical state transitions across levels of organization: neural events in cognition and consciousness and general principles. Biosystems 96, 114–119. doi: 10.1016/j.biosystems.2008.11.011

PubMed Abstract | CrossRef Full Text | Google Scholar

White, P. A. (2018). Is conscious perception a series of discrete temporal frames? Conscious. Cogn. 60, 98–126. doi: 10.1016/j.concog.2018.02.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Yurchenko, S. B. (2021a). The importance of randomness in the universe: superdeterminism and free will. Axiomathes 31, 453–478. doi: 10.1007/s10516-020-09490-y

CrossRef Full Text | Google Scholar

Yurchenko, S. B. (2021b). Why the quantum brain? OBM Neurobiol. 5:103. doi: 10.21926/obm.neurobiol.2103103

CrossRef Full Text | Google Scholar

Yurchenko, S. B. (2022). A systematic approach to brain dynamics: cognitive evolution theory of consciousness. Cogn. Neurodyn. doi: 10.1007/s11571-022-09863-6

CrossRef Full Text | Google Scholar

Zalucki, O., and van Swinderen, B. (2016). What is unconsciousness in a fly or a worm? A review of general anesthesia in different animal models. Conscious. Cogn. 44, 72–88. doi: 10.1016/j.concog.2016.06.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Zeki, S. (2003). The disunity of consciousness. Trends Cogn. Sci. 7, 214–218. doi: 10.1016/s1364-6613(03)00081-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Zimmern, V. (2020). Why brain criticality is clinically relevant: a scoping review. Front. Neural Circuits 14:54. doi: 10.3389/fncir.2020.00054

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: stream of consciousness, volition, criticality, complexity, brain dynamics, quantum, evolution

Citation: Yurchenko SB (2022) From the origins to the stream of consciousness and its neural correlates. Front. Integr. Neurosci. 16:928978. doi: 10.3389/fnint.2022.928978

Received: 26 April 2022; Accepted: 12 October 2022;
Published: 04 November 2022

Edited by:

Elena Monai, University of Wisconsin-Madison, United States

Reviewed by:

Pavel Kraikivski, Virginia Tech, United States
Richard Jones, University of Otago, Christchurch, New Zealand

Copyright © 2022 Yurchenko. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sergey B. Yurchenko, cy55dWNrb0BnbWFpbC5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.