- Cognitive Robotics, Mechanical Engineering, Delft University of Technology, Delft, Netherlands
Background: To counteract gravity, trunk motion, and other perturbations, the human head–neck system requires continuous muscular stabilization. In this study, we combine a musculoskeletal neck model with models of sensory integration (SI) to unravel the role of vestibular, visual, and muscle sensory cues in head–neck stabilization and relate SI conflicts and postural instability to motion sickness.
Method: A 3D multisegment neck model with 258 Hill-type muscle elements was extended with postural stabilization using SI of vestibular (semicircular and otolith) and visual (rotation rate, verticality, and yaw) cues using the multisensory observer model (MSOM) and the subjective vertical conflict model (SVC). Dynamic head–neck stabilization was studied using empirical datasets, including 6D trunk perturbations and a 4 m/s2 slalom drive inducing motion sickness.
Results: Recorded head translation and rotation are well matched when using all feedback loops with MSOM or SVC or assuming perfect perception. A basic version of the model, including muscle, but omitting vestibular and visual perception, shows that muscular feedback can stabilize the neck in all conditions. However, this model predicts excessive head rotations in conditions with trunk rotation and in the slalom. Adding feedback of head rotational velocity sensed by the semicircular canals effectively reduces head rotations at mid-frequencies. Realistic head rotations at low frequencies are obtained by adding vestibular and visual feedback of head rotation based on the MSOM or SVC model or assuming perfect perception. The MSOM with full vision well captures all conditions, whereas the MSOM excluding vision well captures all conditions without vision. The SVC provides two estimates of verticality, with a vestibular estimate SVCvest, which is highly effective in controlling head verticality, and an integrated vestibular/visual estimate SVCint which can complement SVCvest in conditions with vision. As expected, in the sickening drive, SI models imprecisely estimate verticality, resulting in sensory conflict and postural instability.
Conclusion: The results support the validity of SI models in postural stabilization, where both MSOM and SVC provide credible results. The results in the sickening drive show imprecise sensory integration to enlarge head motion. This uniquely links the sensory conflict theory and the postural instability theory in motion sickness causation.
1 Introduction
Experimental and simulation studies have shown that vestibular, visual, and proprioceptive information contributes to postural stabilization of the full body in upright standing (1–6), of the unsupported lumbar spine (7, 8); and the neck (9–15). In models of postural stabilization, it is typically assumed that different sensory modalities act as parallel (additive) pathways, with contributions adapted to the task and sensing uncertainty through sensory reweighting (2, 16). However, experiments in humans and primates indicate that multisensory interactions are more complex: Integration of otolith and semicircular vestibular signals shows consistent differences between self-generated or externally imposed motions (17) and involves prior knowledge or experience (18, 19); and interactions between visual and vestibular signals appear to also include assessments of signal causality (20, 21) that can be likened to evaluations of sensory conflict. We are, however, not aware of models of postural stabilization that incorporate more complex aspects of sensory integration.
Postural stabilization has also been linked to motion sickness (MS). According to the sensory conflict theory of MS, conflicting information from different sensory systems or a mismatch between sensation and expectation is what provokes MS (22). This theory has been further refined to state that it is actually a specific conflict between the perceived vertical and expectations thereof that results in MS (subjective vertical mismatch theory) (23, 24). Although conflict-based theories are the most widely accepted explanation of MS, the exact nature of this conflict remains elusive (25). In an alternative theory, the concept of sensory conflict is rejected on the basis that it derives from an assumption that sensory signals can be ambiguous or non-specific, which is hypothetical [postural instability theory (26–29)]. These authors argue that patterns of stimulation are unique, when considered across multiple sensory systems, and therefore are not ambiguous. Instead, it is proposed that prolonged periods of postural instability, where humans must exert effort to maintain balance, are the cause of MS. However, it has also been suggested that postural instability is not the cause of MS, but rather that they have a common underlying, perceptual, cause (30, 31). For instance, ref. (32) demonstrated a strong correlation between the individual subjective vertical time constant and motion sickness susceptibility.
Advanced models of vestibular and visual sensory integration have been developed to explain motion sickness causation through sensory conflict (33–36). Such models have also been shown to explain conscious self-motion perception experiments when humans are deprived of visual information. We have shown that the subjective vertical conflict model (SVC) and the multisensory observer model (MSOM) quite well predict motion sickness and motion perception in conditions without vision, but best fits were obtained with different parameter sets tuned either for experimental data from perception studies or for sickness studies (37). We further validated MSOM and SVC models by adding visual perception (38) and showed that the SVC with visual perception of rotation rate (SVC-VR) best predicted sickness, but did not (yet) predict perception in all conditions. Adding visual perception of verticality (SVC-VR + VV) did not improve sickness prediction but somewhat improved perception prediction. The MSOM with visual perception of rotation velocity and verticality (MSOM-VR + VV) best predicted all motion perception experiments used for validation.
In the study presented here, we investigate whether these sensory integration models can also capture postural stabilization in relation to MS. The underlying hypothesis is that unified models of vestibular and visual sensory integration can (at least partially) explain and predict (1) postural stabilization, (2) sensory conflicts leading to motion sickness, and (3) conscious self-motion perception.1 We evaluate this hypothesis in an analysis of head–neck stabilization in seated healthy humans, which matches conditions where sensory integration models were validated for motion sickness and self-motion perception. We adopt a biomechanical neck model presented and validated for anterior/posterior stabilization (15) and other directions including frontal impact (39). These articles combined vestibular and visual feedback with joined loops and assumed perfect 3D perception of head orientation in space including verticality and yaw. This assumption of perfect perception is a major simplification, in particular with eyes closed, where verticality perception is confounded by sustained acceleration through the somatogravic illusion (40–43). In the current study, we employ models of sensory integration using physiologically plausible vestibular and visual motion percepts. Vestibular perception consists of otoliths sensing specific force resulting from acceleration and gravity as well as the semicircular canals sensing rotational velocity. Visual perception captures rotation velocities (visual rotation, VR), verticality (visual verticality, VV), and yaw. As motivated in our recent validation study (38), we select the latest versions of MSOM and SVC models of sensory integration of vestibular and visual motion perception and integrate these to capture neck postural stabilization.
Insights and models capturing postural stabilization can be of value in the medical field for research, diagnosis, and treatment and in fields, such as vehicle comfort and impact biomechanics. This study addresses neck postural stabilization in the frequency domain with small loading amplitudes and high bandwidth, illustrating the ability of the models to predict head motion in frequency and amplitude ranges relevant to motion comfort. In addition, we validate the models for a highly dynamic sickening drive eliciting motion sickness.
2 Methods
2.1 Biomechanical head–neck model
A wide range of neuromuscular neck models has been presented in the literature, ranging from 1-pivot models (44–46) to detailed multisegment models (47–55) and partial finite element models (56–65). These models were primarily designed for high-severity road accident loading and/or captured only few motion directions. To address these limitations, we adopted a three-dimensional (3D) multisegment non-linear neck model (66–68) extended with a postural controller stabilizing the head–neck system in the presence of gravity and trunk motion [(15, 39); Figure 1].
Figure 1. Neural control model of the neck. Trunk motion is applied at the base of the neck (T1). Vestibular and visual motion cues, after sensory integration, result in 3D estimates of head angular rate , angle , and acceleration in space. Muscle spindles signal contractile element (CE) length and velocity . Orange blocks contain the feedback sensitivity (gain) and co-contraction parameters controlling the head angular rate , angle , and acceleration with feedback sensitivity parameters Gang.rate, Gang, Gacc, and controlling muscle length with sensitivity parameters kp (position) and kv (velocity) where the reference length L0 represents the desired posture, which is modulated to obtain the desired head angles . Green blocks are muscle synergy vectors converting scalar control signals to an appropriate activation of multiple muscle segments for flexion (Naflex-r for rotation and Naflex-t for translation), extension (Naext-r for rotation and Naext-t for translation), co-contraction (Nacc), and postural activity counteracting gravity (Napost). Blue blocks contain sensory delays for vestibular/visual (τvcr) and muscle feedback (τccr) and muscular activation dynamics (Hact) transforming neural excitation (e) into muscle active state (a). In head angle control a first-order low-pass filter Hang with time constant τang lumps additional delays for visual contributions, neural processing, and control strategies emphasizing lower frequencies. Thick lines indicate multiple signals for all 258 muscle segments. This figure shows anterior–posterior stabilization through neck flexion and extension muscle synergies. Equivalent loops have been added for lateral and yaw motion control.
The neck model contains nine rigid bodies representing the head, seven cervical vertebrae (C1–C7), and the first thoracic vertebra (T1). The eight intervertebral joints allow 3D rotational and translational motion, resulting in a total of 48 degrees of freedom (DOF). Passive joint properties are captured with non-linear force models representing ligaments, intervertebral disks, and facet joints. Muscles (34 muscles, totaling 129 elements per body side) are implemented as line elements based on dissection (69) with ‘via points’ connecting muscles to adjacent vertebrae to ensure the muscles take on a curved path during head–neck displacement, and with non-linear Hill type contractile elastic and series elastic dynamics. Gravity is simulated as a 9.81 m/s2 gravitational field acting on the skull and the vertebrae. The neck model was validated in passive bending and twist and in isometric loading where the ligamentous spine stiffness, instantaneous joint centers of rotation, muscle moment arms, isometric strength, and muscle activation patterns were in general agreement with biomechanical data (68).
The postural stabilization model, parameter estimation, and validation in anterior–posterior loading can be found in (15). Validation for other directions, including frontal impact, is presented in ref. (39). Feedback loops were added for head lateral motion and yaw, equivalent to the anterior–posterior loops. Lateral loops provide feedback on head roll angular velocity and roll angle in space. Yaw loops provide feedback on head yaw angular velocity and yaw angle in space. As described in the Results section, we explore feedback of head rotational velocity, taking into account the dynamics of the semicircular canals, and explore models of sensory integration to provide feedback of head angles in space. Details on the neck model and muscle dynamics can be found in the Appendix.
The biomechanical neck model was implemented in the simulation software MADYMO 2022. Sensor dynamics, neuromuscular control, delays, and muscle dynamics were implemented in MATLAB R2022b. Euler integration (ode1) was applied with a fixed time step set to 10 μs, resulting in a computation time of approximately 100 times real-time on a 2.8-GHz processor. The ISO coordinate system is applied (x = forward, y = left, z = up).
2.2 Models of sensory integration
To capture integration of vestibular and visual motion information, we employed the latest versions of the multisensory observer model (MSOM, see Supplementary Figure A3) (36) and the subjective vertical conflict model (SVC, see Supplementary Figure A4) (33, 34) as described in ref. (38). As described in ref. (38), we use the so-called SVCI model which contains an integrator (dotted box in Supplementary Figure A4d) to process the acceleration conflict Δa. Both models have two vestibular inputs, being the specific force resulting from gravity and acceleration sensed by the otolith organs, and the rotation rate sensed by the semicircular canals. Both models have two visual inputs which are the 3D rotation rate resulting from optical flow, and head orientation consisting of verticality perceived through horizontal structures, such as the horizon or vertical structures such as buildings, and the yaw angle in space. Both models take into account the semicircular dynamics as a high pass filter, which is first order in MSOM, and second order in SVC, while otolith and visual dynamics and delays are ignored. The neck stabilization model also includes “direct” feedback of semicircular motion perception (see Section 3.2) using more advanced semicircular dynamics described in the Supplementary Material. For MSOM and SVC, we apply their original vestibular dynamics which were previously used to select the model parameters and to validate these models.
Both MSOM and SVC employ a state estimation approach using an internal model of 6D head motion and semicircular dynamics. Estimated head motion states are adapted using correction loops comparing predicted and actual vestibular and visual motion perception. In this study, we use MSOM and SVC models with different levels of complexity labeled as: NV (no vision or eyes closed) with all visual loops disabled, VR (visual rotation rate) adding visual perception of rotation velocity, and VR + VV also adding visual perception of verticality. The VR option is evaluated to assess the importance of VV comparing VR to VR + VV. Furthermore, the SVC is proposed with only VR by ref. (34) and we recently showed SVC-VR to best predict motion sickness (38).
We applied the MSOM parameters presented in the original publication by Newman (36) and SVC parameters from ref. (33, 34) as summarized in Table 1. We also applied two-parameter sets which we retuned (37) to optimally match motion perception without vision (MSOM-NV-PERC) and motion sickness without vision (MSOM-NV-MS). These two sets include vestibular gains only as they were adapted to conditions without vision.
The MSOM and SVC do not estimate the head orientation angles in space but estimate the perceived verticality vector v in the head coordinate system. We use v to derive the perceived head pitch and roll angles in space as:
For the MSOM, we use the verticality estimate (Supplementary Figure A3c). The SVC model generates two estimates of verticality, with vs derived from vestibular otolith and semicircular information, and derived from the internal model integrating vestibular and visual information (Supplementary Figures A4b,d). Therefore, is the more plausible percept, in particular in conditions with vision. However, both percepts may play a role in postural stabilization. Hence, we evaluate both options, which will be further referred to as SVCint using integrating visual and vestibular, and SVCvest using vs integrating vestibular inputs only. Here, it should be realized that SVCint and SVCvest use different estimates of verticality but are actually derived from the same model.
The MSOM and SVC estimate yaw rotational velocity but do not estimate the head yaw angle. We derive the head yaw angle by 3D integration of the estimated rotational velocity. This integrates vestibular and visual perception of yaw rotation. Such an integration will result in inaccuracy and drift due to sensor imperfections (70, 71). This is realistic for conditions without vision, but availability of vision will correct such imperfections. To describe this, a visual correction loop for the perceived yaw angle in space can be added in future models.
2.2.1 Discrimination of tilt from acceleration
The MSOM and SVC respond to otolith, semicircular, and visual inputs with complex dynamics. Empirical data indicate that the perception of pitch is frequency-dependent. Without vision, slow, low-frequency pitch is perceived as translational acceleration, whereas fast, high-frequency pitch is perceived as such (i.e., rotation), with a crossover frequency of approximately 0.2 Hz (32). Simulations of MSOM and SVC were performed to illustrate these dynamics. Here, we directly prescribed the head motion and did not use the neck model. Figure 2 shows the dynamics of verticality perception by applying a step pitch head rotation (left) and a sustained acceleration designed to elicit the somatogravic illusion (right). Hence, these two cases demonstrate how well the sensory integration models discriminate tilt from horizontal acceleration. The MSOM follows the applied head rotation with a negligible delay and with a precise and stable pitch estimate even without vision. This is expected as the otoliths will sense the constant pitch. For MSOM without vision, the three-parameter sets show marginal differences. The sickness-tuned MSOM-NV-MS deviates a bit but reconverges to 10 degrees after 50 s. Much larger differences between MSOM variants emerge with sustained horizontal acceleration (Figure 2 right). Visual verticality (VV) perception strongly reduces the perceived pitch, while the visual rotation rate (VR) loop hardly affects the result. Without vision, the MSOM predicts persistent pitch perception, matching the somatogravic illusion [(e.g., 40–43)]. The sickness-tuned parameters provide a slowly developing somatogravic illusion converging to 10 degrees after approximately 50 s.
Figure 2. Verticality perception for (from top to bottom) MSOM, SVCint using the integrated verticality vint and SVCvest using vestibular verticality vvest. Left: A step pitch rotation with ramp onset (10 degrees in 0.1 s) was applied to the head. Right: A sustained rearward acceleration with ramp onset (0.1 s) was applied to the head with an amplitude of 1.7 m/s2 , which affects the specific force with 10 degrees rotation and therefore induces a steady-state pitch perception of 10 degrees for all models without vision. SVCvest is not affected by vision and therefore the three model lines coincide.
2.2.1.1 SVC parameter tuning
The SVC using the integrated vestibular and visual verticality (SVCint) follows the applied head rotation with a substantial delay. This delay is smallest with full vision including visual verticality (SVCint-VR + VV). As shown in the following sections, this delay hampered effective neck stabilization in conditions without vision. Hence, we explored whether tuning of the SVC parameters could enhance the perception dynamics. This was successful for SVCint-VR + VV as illustrated by the lines “high Kgvis” in Figure 2. The increased gain Kgvis contributed to a faster and more precise verticality perception both with head rotation and with sustained acceleration. Tuning other gains marginally affected these responses, and hence, other parameters were left unchanged. For SVCint-VR and for SVCint-NV, tuning was hardly effective, and hence, results are not shown. SVCint also elicited a somatogravic illusion, which was strongly reduced with vision (Figure 2 right).
The SVC using the vestibular estimate of verticality (SVCvest) rapidly follows the applied head rotation and is, as expected, not affected by the visual loops. Due to its fast response, SVCvest is a credible percept, which may effectively contribute to postural stabilization. SVCvest also creates a somatogravic illusion with horizontal acceleration.
The somatogravic illusion (Figure 2 right) develops most rapidly with MSOM followed by SVCint and SVCvest. Recent estimates of the time constant of this illusion range between 2 s (42) and 9.2 s (SD 7.17 s) (32), which appear to be matched reasonably well by model predictions. The perception of pitch should be suppressed when vision is available, which is partially achieved by the MSOM and SVCint.
Roll and other tilt directions yield an identical response (not shown) as MSOM and SVC models and the applied parameters are identical for pitch and roll. Figure 3 presents a similar step response for head yaw. Here, all models show a rapid yaw response, which remains with vision. In yaw, the two vision models (VR + VV and VR) respond identically as, as explained above, MSOM and SVC do not use the visually perceived yaw angle and simply integrate the perceived rotation rate. Without vision, the perceived yaw decays to zero where this decay depends strongly on the model parameters as illustrated by the three lines without vision for MSOM. The predicted fading of perceived rotation without vision matches empirical findings where the sense of rotation gradually fades out, with an average time constant of 17.2 s (SD = 6.8 s), in participants rotated around an Earth–vertical yaw axis (32).
Figure 3. Yaw perception for (from top to bottom) MSOM, SVCint using the integrated verticality vint, and SVCvest using vestibular verticality vvest. A step yaw rotation with ramp onset (10 degrees in 0.1 s) was applied to the head.
2.3 Neck postural stabilization
The neck model was validated using postural stabilization data from eight experimental studies with seated healthy adult human subjects as summarized in Table 2. In the first five experiments, subjects were restrained by a harness belt on a rigid seat mounted on a motion platform. In the lateral (Lat) tests, the subjects were also laterally supported with adaptable cushioned plates (73). In all experiments, head motion was recorded in the direction in which seat motion was applied. 3D head motion in both translation and rotation was available for the experiments AP, Lat, Roll, XYZ-compliant, and the slalom. In the anterior–posterior (AP) tests, the T1 translation (base of the neck) was recorded and applied as input to the neck model and used to derive transfer functions from trunk motion to head motion. For Lat, Pitch, and Roll conditions, trunk motion was reported to be close to the seat motion and the seat motion was applied to T1 in the neck model. Checking the transmission from seat motion to T1 in our own data (72, 73), we found gains close to one for trunk horizontal translation and roll, but we also found some phase shifts which shall be considered when interpreting the results. For the yaw conditions, we used recent data (75), which repeated experiments by Keshner (78). The recent dataset was selected as it includes more subjects (17 instead of 7) and describes head global motion as a function of trunk motion recorded at T2. We compared the experimental T2-to-head transmission to the model-based T1-to-head transmission as the model does not include the joint between T1 and T2. The dataset XYZ-compliant was collected on a motion platform with a car seat with a compliant configurable backrest, using the condition with erect posture, high backrest, and eyes open (76). Here, the recorded trunk motion was used to prescribe the T1 motion of the neck model. The slalom was measured in a vehicle on the compliant back seat, and the recorded trunk motion was used to prescribe the T1 motion of the neck model. All reported signals represent motion at the head center of gravity.
3 Results
Models of increasing complexity in terms of postural stabilization feedback and sensory integration were fitted to the experimental data. Comparison of model fits can illustrate the relevance of feedback loops and sensory integration across motion conditions. For each condition and model version, the postural control parameters were estimated by fitting the model to the experimental data. The postural feedback gains and co-contraction (see Figure 1) were fitted to optimally match the model response with the human response data in the frequency domain (see Supplementary Material). This generally resulted in a good fit, and hence, the parameters of sensory integration were not fitted and remained as defined in Table 1. As described below, for vertical loading, the data were not very informative and we applied parameter sets estimated for horizontal seat translation from the same dataset, and do not report the model error.
Table 3 shows the resulting model error for postural stabilization and perception models of increasing complexity. Here, the model error was scaled toward the model error assuming perfect perception of verticality and yaw, which shows a (near) optimal fit. Hence, Table 3 allows a rapid comparison of the ability of all models studied to match the human response data. Figures 4–10 show validation results for the most relevant models. Figures 4–9 show the results for the 6-seat motion directions in the frequency and time domain. In each figure, the head response is shown for the perturbed seat motion direction and other relevant head motion directions. For instance, Figure 4 shows relevant head pitch motion in response to AP seat translation and Figure 5 shows head roll and yaw in response to lateral seat translation. Figure 10 shows the validation for the slalom drive in the time domain as for this dataset power is concentrated approximately 0.2 Hz for lateral and yaw motion making the slalom unsuitable for frequency domain analysis.
Table 3. Model error in predicting head motion for models of varying complexity in feedback loops (left) and sensory integration (right).
Figure 4. Validation for anterior–posterior seat translation with eyes open (APEO) for muscle/semi/perfect perception (upper 3 graphs), and MSOM and SVC (lower 3 graphs). Only the most relevant models are shown as lines largely coincide.
Figure 5. Validation for lateral seat translation with eyes closed (LatEC) for muscle/semi/perfect perception (upper 3 graphs) and MSOM and SVC (lower 3 graphs). Only the most relevant models are shown as lines largely coincide.
Figure 6. Validation for vertical seat translation on a compliant seat. High-frequency experimental results deviate due to the imprecision of the applied T1 motion, which was based on measured trunk motion. Only the most relevant models are shown as lines largely coincide.
Figure 7. Validation for lateral seat rotation with eyes closed (RollEC) for muscle/semi/perfect perception (upper 3 graphs) and MSOM and SVC (lower 3 graphs). Only the most relevant models are shown as lines largely coincide.
Figure 8. Validation for anterior–posterior seat rotation (pitch) for muscle/semi/perfect perception (upper 2 graphs), MSOM (middle 2 graphs), and SVC (lower 2 graphs) with 3 tasks: PitchMA no vision with mental arithmetic (left), PitchNV no vision active control (mid), PitchVS with vision active control (right).
Figure 9. Validation for left/right axial seat rotation (yaw) for muscle/semi/perfect perception (upper 2 graphs), the most relevant MSOM and SVC results as other results overlap (lower 2 graphs) with 3 tasks: YawMA no vision with mental arithmetic (left), YawNV no vision active control (mid), YawVS with vision active control (right).
Figure 10. Validation in slalom, showing only the most relevant models, for head rotation (upper 3 graphs) and acceleration (lower graph—responses for muscle and muscle+semi are not shown as they overlap with perfect).
3.1 Muscle feedback
Muscle feedback was implemented representing muscle spindle feedback of relative length and velocity, jointly fitting two parameters for all 258 individual muscle elements. A feedback delay of 13 ms was selected based on ref. (79). Models with muscle length and velocity feedback and without any vestibular and visual motion perception could well stabilize the neck in all conditions. However, a very poor fit of the experimental response was obtained, in particular for conditions with trunk pitch and roll, where the model predicted excessive head rotations (lines muscle in Figures 4–10). The model error was on average 16.41 times larger (Table 3) than for the model assuming perfect perception of head verticality and yaw described in Section 3.3. Muscle feedback can apparently well stabilize the head on the trunk, but as expected, muscle feedback cannot effectively reduce head rotation in space in conditions with trunk rotation. Likewise, in the highly dynamic slalom drive, the model with only muscle feedback resulted in excessive head pitch and roll. In some conditions, adding co-contraction (~1%) slightly improved the model fit, but the effects were very similar to the effects of increasing muscle velocity feedback. In the slalom, co-contraction (4%) improved the model fit and allowed higher muscle length and velocity feedback gains while not inducing oscillations. This can be explained by the dynamics of the applied Hill-type muscle model where co-contraction increases muscle damping through the force–velocity relationship and reduces the phase lag of muscular dynamics through preload of the series elastic element.
3.2 Muscle and semicircular feedback
Vestibular feedback of head rotational velocity was implemented taking into account the dynamics of the semicircular canals as motivated in ref. (15) and the Supplementary Material using vestibular dynamics from ref. (80). Therefore, we assumed a direct contribution of semicircular perception to head rotation control independent of any model of sensory integration. A vestibular feedback delay of 13 ms was applied based on human studies (81, 82). Adding semicircular feedback was highly effective in reducing head-in-space rotation at the mid-frequencies (lines muscle + semi in Figures 4–10), and the average model error was reduced from 16.41 to 6.71 (Table 3). However, at low frequencies and in the slalom, the model still predicted excessive head rotations. We also simulated perfect perception of head rotational velocity, and an estimation of head rotational velocity using the MSOM model integrating semicircular and otolith information. The results for these different estimates of rotational velocity were highly similar and are therefore not shown. The similarity can be explained by the fact that semicircular dynamics show a limited frequency sensitivity in the range from 0.5 to 6 Hz (see Supplementary Figure A2 left in the Appendix) where semicircular feedback most strongly contributed to head stabilization. Even with perfect rotation velocity feedback, the low-frequency head rotation largely exceeded the experimental head rotation in conditions with trunk pitch and roll rotation and in the 0.2 Hz slalom.
3.3 Muscle, semicircular, and perfect rotation angle feedback
To enhance the model’s ability to control head orientation at lower frequencies, feedback of head orientation angles was added (verticality and yaw), assuming perfect perception. The vestibular feedback delay of 13 ms was also applied for this loop, combined with a first-order low-pass filter Hang with time constant τang, lumping additional delays for visual contributions, neural processing, and control strategies emphasizing lower frequencies. For AP, Lat, and Roll, τang was not very sensitive and was set to 100 ms being representative of visual delays. For pitch and the slalom, 30 ms was selected to enhance the model fit, suggesting a strong vestibular contribution to the perception of head rotation angles. For yaw perception τang was increased to enhance the model fit, suggesting a control strategy emphasizing the lower frequencies (YawVS: 400 ms, YawNV: 300 ms, YawMA: 150 ms). The two rotational feedback loops (rate and angle) minimized head rotation in space in all conditions. The only exception is the slalom, where we evaluated both head-in-space and head-in-vehicle control strategies. For the slalom, vehicle roll and pitch were limited, and hence, both strategies yielded a good fit, and we present results for head-in-space control. The slalom showed vehicle yaw up to 30 degrees, and the best results were obtained with head-in-vehicle control for yaw angle while using head-in-space control for the direct semicircular feedback.
Perfect rotation angle feedback led to a very good fit in all validation sets (lines perfect in Figures 4–10 and column perfect angle in Table 3). Adding feedback of head orientation in space was most effective in conditions with trunk rotation and visual feedback, where the model error was reduced by a factor of 12.68 for PitchVS and 11.23 for YawVS. In the slalom, the model error was reduced by a factor of 2.19 due to reduced head pitch and roll and due to a better alignment in time for head yaw, where head-in-vehicle control (line perfect) yielded much better results than muscular head-on-trunk control (line muscle).
The assumed perfect perception of head orientation is plausible in conditions with visual feedback. The condition APEC without visual feedback was well described without such feedback, but for all other conditions without visual feedback, rotation angle feedback was needed to match the experimental data. To provide a neurologically plausible control model in conditions with and without vision, we explored models integrating vestibular and visual sensory information as explained in the following sections.
3.4 Otolith (OTO) feedback
As a first step, we assumed verticality perception simply using low-pass-filtered otolith information (OTO) to control head pitch and roll. Such a low-pass filter aligns with the concept of otolith information to reflect acceleration at high frequencies and verticality at low frequencies (40–43, 83–86). A first-order low-pass filter was used with time constants τang between 0 and 10 s. With low values (τang ≤ 0.5 s) OTO feedback improved results for the slalom and RollEC, but the most demanding Pitch cases showed marginal improvement (see results for τang = 0.03 s in Table 3). As compared to perfect perception of pitch and roll, with OTO feedback, gains had to be reduced to prevent oscillations and this was even more detrimental with larger τang (see the results for τang = 5 s in Table 3). Apparently, low-pass-filtered OTO feedback is hardly usable to dynamically control head pitch and roll. Results for Yaw conditions were hardly affected by OTO feedback, which is not surprising as the presented OTO results simply assumed perfect yaw perception.
3.5 MSOM and SVC
The above results illustrate that a more advanced estimation of head rotation angles is needed. This was achieved using the MSOM and SVC models of sensory integration described in Section 2.2, which provide plausible estimates of verticality (head pitch and roll) and yaw. Instead of perfect perception, now head verticality and yaw derived using MSOM or SVC were used for head orientation feedback. Feedback of head rotational velocity remained based on direct semicircular feedback as described in Section 3.2. The feedback delays and time constants introduced above were kept as MSOM and SVC do not include delays. Future models could redistribute delays and time constants across feedback and sensory integration models.
Table 3 shows that in particular the roll, pitch, and slalom conditions are well captured by several sensory integration model variants but not by others. The AP and Lat cases are hardly sensitive because the gain of angular feedback is limited. In the Yaw cases angular feedback is very important, but apparently, all MSOM and SVC model variants adequately predicted head yaw. The only exception is SVCint-NV where the head is not well stabilized leading to complex 3D head motion including substantial roll. As only yaw data were available, we could not assess the validity of head roll and pitch in response to trunk yaw. The similarity of results for MSOM and SVC variants concurs with Figure 3, which also shows MSOM and SVC model variants to yield almost identical responses in yaw. Head yaw was derived from MSOM and SVC through integration which is overly simplistic, and the scope for further validation and improvement of the MSOM and SVC models in predicting head yaw will be addressed in the discussion.
The MSOM model with full vision (MSOM-VR + VV) provided a good fit in all conditions and was generally close to results with perfect perception of verticality and yaw (Table 3). The MSOM no vision (NV) model well captured all conditions without vision. In most conditions, MSOM results were not very sensitive to the applied visual loops. However, the PitchVS case was well fitted by the MSOM-VR + VV model and not by the MSOM-NV and MSOM-VR models, highlighting the importance of visual verticality perception. In the slalom drive, the MSOM with full vision (MSOM-VR + VV) yielded realistic head rotations, whereas the MSOM-NV yielded larger head pitch and roll. We did not find similar experimental driving data comparing external vision to no vision, so we collect such data in ongoing experiments. The MSOM-NV results were sensitive to the applied parameters in roll, pitch, and slalom where the perception-tuned parameters and the motion sickness-tuned parameters (37) were less effective in stabilizing head rotation. This is most apparent in the two trunk tilt conditions without vision (RollEC and PitchNV) where the standard MSOM parameters (MSOM-NV) provide a good fit and the perception-tuned and sickness-tuned MSOM parameters show a substantial model error. The SVC provides two estimates of verticality (see Section 2.2). SVCint using vestibular and visual information shows a considerable time lag (Figure 2), in particular without vision. Hence, SVCint-NV was hardly effective in controlling head pitch and roll. For SVCint-NV, the head pitch and roll angle feedback gains had to be substantially reduced in order to achieve stability, and in the PitchVS, and PitchNV cases, the SVCint-NV model did not improve results. This can be seen comparing columns in Table 3, where the model error for SVCint-NV shows no improvement compared with model muscle + semi, which has no feedback of head rotation angle. The SVC with full vision (SVCint-VR + VV) provided good results in most conditions with the published parameters (33, 34). However, results in pitch improved substantially (Table 3) with the high Kgvis, which was tuned for the step pitch perception simulation in Figure 2. Hence, SVCint-VR + VV well captures conditions with vision, but SVCint-NV poorly describes conditions without vision. The much faster vestibular estimate of verticality SVCvest provides good results in most conditions, which are very close to results with perfect perception of verticality. Apparently, SVCvest is highly effective in stabilizing head verticality and is suitable for capturing conditions without vision. Possibly SVCint and SVCvest are jointly used to control head verticality. However, the current data are not suitable to validate such separate contributions as a good fit was already obtained with SVCvest. SVCvest is not designed to estimate yaw, and we estimated yaw by 3D integration, which proved to be effective in the current validation.
3.6 Six degrees of freedom neck dynamics
The above results show model fits for head motion in the applied seat motion direction, and several other (interacting) head degrees of freedom. Available datasets were limited in bandwidth, but the model allows us to extrapolate the human response to a larger frequency range. Figure 11 shows such results for all six-seat perturbation and head response directions. In all cases, motion was applied directly at T1. As the model and the adopted posture are left/right symmetric, several interaction terms in Figure 11 are zero; for instance, AP motion (top row) does not induce lateral, roll, and yaw motion. Other interactions show zero gains as the linearized transmission is zero, but non-linear behavior will induce higher harmonics. For instance, lateral seat motion (second row) will induce some vertical head motion, but this will be identical to left or right seat motion, leading to a zero linearized transmission.
Figure 11. Six degrees of freedom head–neck model response. Rows describe applied T1 motion conditions from top to bottom: AP, lateral and vertical translation roll, pitch, and yaw rotation. Columns describe the corresponding head motion. Lines represent different sensory integration models. The line “perfect high gain” uses the high feedback gains estimated for the slalom with perfect estimation of head rotation. All other lines use low feedback gains jointly estimated for the horizontal loading conditions APEO and Lat. The high gain substantially affected the results, whereas the results for perfect perception, MSOM, and SVC hardly differed.
The three upper left diagonal cells in Figure 11 show that T1 translation induces head motion in the corresponding direction with a gain close to one at low frequencies. At mid-frequencies, some amplification (gain>1) is shown for all three translation motion directions, in particular with high feedback gains. As expected the high feedback gains effectively reduce head rotations in all three directions (columns roll, pitch, and yaw) in all loading conditions (all rows). The sensory integration models have limited effects.
3.7 Motion sickness prediction
Table 4 shows the sensory conflict between the perceived and estimated verticality vector in the slalom. This conflict is limited for MSOM given the magnitude of applied vehicle acceleration of 4 m/s2. Larger conflicts are found for SVC, indicating non-perfect estimation of verticality (as associated with motion sickness causation). For MSOM, the conflict is slightly larger with full vision (VR + VV) as compared to NV, which is unrealistic. The SVC predicts the expected larger conflict with eyes closed (NV) as compared to full vision (VR + VV), which aligns with higher sickness being driven without vision (77, 87). The SCVint-VR + VV model with high Kgvis improves the model fit (Table 3) but hardly affects the conflict (Table 4). The SVCint-VR is promoted as model to predict sickness in several publications and was indeed found to best predict sickness in our evaluation of MSOM and SVC models (38). Comparing SVCint-VR to SVCint-NV also shows the expected trend of a larger conflict without vision.
For the SVC, Table 4 shows more head roll with a larger sensory conflict comparing SVCint-VR + VV to SVCint-NV and comparing SVCint-VR to SVCint-NV. This larger head roll results from an inaccurate estimation of the head rotation, making feedback of the estimated head rotation less effective. This shows that inaccurate perception can jointly induce sensory conflict and postural instability, joining both the sensory conflict and the postural instability theory of motion sickness causation.
4 Discussion
A biomechanical neck model was uniquely extended with postural stabilization using SI of vestibular (semicircular and otolith), and visual (rotation rate, verticality, and yaw) cues using the multisensory observer model (MSOM) and the subjective vertical conflict model (SVC). The ability of the combined model to capture postural stabilization and motion sickness causation was evaluated using existing empirical datasets, including 6D trunk perturbations and a slalom drive inducing motion sickness.
4.1 Insights gained in neck postural stabilization
The neck model with postural stabilization well matches experimental head translation and rotation responses. When omitting vestibular and visual perception, muscular length and velocity feedback could stabilize the neck in all conditions. Muscle feedback can stiffen the neck supporting a head-on-trunk control strategy. This provided a reasonable fit for low-amplitude horizontal seat translation conditions (AP and Lat). This also concurs with observations in vestibular loss patients where “there are no dramatic differences between patients and controls” in conditions similar to APEC (88). However, this resulted in excessive head rotations in conditions with trunk pitch and roll and in the highly dynamic slalom drive. Here, realistic results could only be obtained by adding two rotational loops: Gang.rate providing direct feedback of head rotation velocity as sensed by the semicircular organs using vestibular sensitivity functions (80) and Gang representing feedback of rotation angles (verticality and yaw). These loops support a head-in-space control strategy for all conditions studied, with an exception for yaw in the slalom where visual feedback aligned the head with the vehicle driving direction. Direct feedback of 3D head rotational velocity sensed by the semicircular canals effectively reduced head rotations at the mid-frequencies. Realistic head rotations at low frequencies were obtained by adding sensory integration (SI) based feedback of head rotation in space (verticality and yaw). This supports the validity of the MSOM and SVC SI models in postural stabilization. A good fit was also obtained assuming perfect perception of verticality and yaw. Hence, this assumption of perfect perception is adequate when developing models predicting head motion. However, this assumption is not justified when developing models explaining neck postural stabilization, where SI models can help unravel the role of vestibular and visual information.
Conditions with small amplitude trunk translation (AP and Lat) were hardly sensitive to the applied SI model. This can be explained by the low gain for angular feedback estimated for these conditions, making the response insensitive toward the selected pitch and roll perception model. In experimental conditions with trunk pitch and roll and in the highly dynamic slalom, high angular feedback gains were needed to fit the data, and head rotation varied strongly with SI model type and parameters. In particular, the pitch condition with three instruction and vision conditions (VS, NV, and MA) was highly suitable for validation. For roll, only one instruction was tested (NV), and it may be interesting to collect VS and MA data. MSOM and SVC both use the same parameters for pitch and roll and therefore respond identically to pitch, roll, and other tilt directions. However, quasistatic experiments show more precise tilt perception in roll (89). In yaw, three instruction and vision conditions (VS, NV, and MA) were tested, and like in pitch, feedback of head rotation angles in space proved essential to match the experimental data. In yaw, all MSOM and SVC variants except SVCint-NV well captured the data. However, MSOM and SVC have not been designed for yaw angle perception and are only validated for yaw rate perception for the cases of earth–vertical and off–vertical axis rotations at a constant yaw rate (38). Here, it was observed that only the models with visual rotation (VR and VR + VV models) are able to simulate human yaw rate perception. For this study, we extended the MSOM and SVC to estimate the yaw angle through the integration of yaw rotation velocity. This well captured the current data but will be inadequate in cases with sensory imperfections, calling for further validation and addition of a visual yaw angle perception loop.
The MSOM-VR + VV with full vision well captured all conditions, whereas the MSOM-NV without vision captured all conditions without vision. This illustrates that the MSOM is a plausible model to capture and explain neck postural stabilization. The SVM provides two estimates of verticality, with a vestibular estimate SVCvest, and an integrated vestibular/visual estimate SVCint. The vestibular SVCvest well captures all conditions, including those with vision. This makes SVCvest a plausible contributor to head–neck stabilization but fails to explain the role of vision. The integrated SVCint shows plausible results but follows the actual head rotation with a substantial delay, in particular without vision. With vision, this delay could be effectively reduced increasing the verticality perception gain Kgvis. This resulted in a faster and more precise verticality perception both with head rotation and with sustained acceleration (Figure 2) and a better fit of human head motion data (Table 3). With high Kgvis, the SVC approximates perfect verticality perception, which is actually equivalent to a direct feedback of the visually sensed verticality. Without vision, tuning of the SVC parameters was not effective, and the slow response of SVCint-NV led to oscillations with high angular feedback gains, and a poor fit, in particular with trunk pitch. Hence, SVCint alone cannot explain head–neck stabilization without vision. However, SVCint may well complement SVCvest, and with such a combined approach, the SVC will also be a plausible model to capture and explain neck postural stabilization.
MSOM and SVC were tested with full vision (VR + VV) using both visual rotation rate (VR), and visual verticality (VV) as well as with rotation rate only (VR). Adding the VV loop enhanced the model fit both for MSOM and SVC showing the relevance of visual verticality perception in head–neck stabilization.
Muscle feedback stabilizes the head on the trunk. Muscle feedback also proved essential for the stabilization of the individual intervertebral joints and to prevent neck buckling. Without muscle feedback, static stability could not be achieved, resulting in excessive static flexion or extension of the individual neck joints and the entire neck (15). Neck muscle co-contraction was estimated to be up to 1% of maximal muscle activation and 4% in the slalom. Co-contraction contributed to head-on-trunk stabilization up to 1 Hz and allowed higher feedback gains in the slalom. This highlights a relevant contribution of neck muscle co-contraction, in particular in high-acceleration conditions.
4.2 Modulation of postural stabilization
Experimental studies have shown the ability of the central nervous system (CNS) to modulate neck afferent feedback in response to changing external environments (9, 90–94). We demonstrated modulation of neck afferent feedback with the frequency bandwidth of anterior–posterior trunk perturbations (72), with modest effects of the presence of vision. The neck model enabled the estimation of postural control parameters for these conditions (15). Control strategies employed during low-bandwidth perturbations most effectively reduced head rotation and head relative displacement up to 3 Hz, while control strategies employed during high-bandwidth perturbations reduced head global translation between 1 and 4 Hz. This indicates a shift from minimizing head-on-trunk rotation and translation during low-bandwidth perturbations to minimizing head-in-space translation during high-bandwidth perturbations. This modulation of control may well be beneficial in terms of comfort, limiting the transfer of 1–4 Hz horizontal seat motions to the head, where comfort standards for whole body vibration attribute considerable weight to these frequencies (95).
The current study evaluated fundamentally different motion conditions and tasks resulting in a stronger modulation of postural gains. High gains for pitch angle were needed in the pitch and slalom conditions (gains were APEO:0.5; PitchMA:1.9; PitchNV:4.8; PitchVS:7.1; slalom:5.2). Similarly, the gains for roll angle were modulated strongly (LatEC:0.6; RollEC:1.3; slalom:6.5). Figure 11 illustrates the importance of feedback gain modulation across 6D perturbation and response directions, where the “high gain” response matching the dynamic slalom drive differed profoundly from the response with low gains estimated for low-amplitude horizontal acceleration. This postural feedback modulation may be beneficial affecting comfort and muscular effort, and we are currently exploring optimal control strategies to explain and predict modulation of postural stabilization.
The SI models have now been applied with the published parameter sets and with two alternative parameter sets for MSOM (NV-PERC and NV-MS) and one for SVC (VR + VV high Kgvis) (Table 1), showing marked effects of SI parameters on postural stabilization. The SI parameters may also be modulated with motion conditions and tasks. As the current data were already well captured, we did not attempt to fit the SI parameters with the current validation sets. Here, perception experiments are presumably more informative and suitable to validate SI models and parameters (37, 38).
4.3 Sensory conflict, postural instability, motion sickness, and motion perception
This study uniquely links models of sensory integration to postural stabilization, with a comprehensive validation for postural stabilization complemented with an exploration of motion sickness. As expected, in the sickening slalom, SI models could not precisely estimate the actual head rotation (verticality) resulting in sensory conflict. Thus, our results support the sensory conflict theory in motion sickness causation. The SVC predicted larger conflicts than the MSOM, which provides a close to perfect prediction of verticality, indicating the SVC to be more promising for sickness prediction. When removing vision, only the SVC predicts the expected increased sickness (77, 87). These results are in line with our recent study (38) showing SVC to be more suitable for motion sickness prediction, whereas MSOM best captured motion perception. We feel that both MSOM and SVC can be enhanced to further explain postural stabilization, motion perception, and motion sickness. Here, we will aim for common sensorimotor integration models but will also take into account evidence toward partially different processes (96).
Our results also show that imprecise sensory integration can enlarge head motion. This shows that inaccurate perception can jointly induce sensory conflict and postural instability, relating both the sensory conflict and the postural instability theory in motion sickness causation.
4.4 Limitations and future study
The current models assume group-based perception and postural stabilization parameters to fit group-based postural stabilization data. However, individuals show marked differences in kinematic, perception, and motion sickness responses. Interpersonal variability in model parameters, or even fundamental differences in neural processing, may potentially explain such individual differences in responses, and link the domains of postural stabilization, motion perception, and motion sickness. For instance, individual motion perception time constants were recently shown to relate to sickness (32). Therefore, the next challenge will be to estimate individual parameters for perception, postural stabilization, and motion sickness susceptibility (97). This can combine dedicated experiments measuring perception, postural stabilization, and sickness on a substantial pool of participants, thereby disclosing relations across these domains and explaining individual differences. The validation can also be extended in frequency range, in particular for pitch and yaw where current data are limited to 3 Hz while Figure 11 shows effects of feedback up to approximately 8 Hz. Effects of SI models and parameters are most apparent at low frequencies (<1 Hz) and can be studied using low-frequency or quasistatic experiments.
The current model is deterministic with a non-linear Hill-type muscle model, non-linear passive structures, and linear idealized models of sensors and sensory integration. Stochastic and more detailed non-linear models may further explain sensory integration and neck stabilization including postural sway. Vestibular dynamics are described as lumping regular and irregular afferents (see Supplementary Material), using a linear second-order semicircular function for direct feedback, a first-order function in MSOM, and a second-order function in SVC, while ignoring otolith dynamics. More advanced models could discriminate regular and irregular afferents taking into account non-linearities and the stochastic nature of vestibular motion perception (98).
The biomechanical neck model contains detailed structures leading to a high computational demand, taking days to fit experimental datasets. To enable individual modeling, we explore computationally efficient (simplified) biomechanical neck and full body models running faster than real time and explore optimal control strategies to explain and predict modulation of postural stabilization.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
Ethical approval was not required for the study involving humans in accordance with the local legislation and institutional requirements. Written informed consent to participate in this study was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and the institutional requirements.
Author contributions
RH: Conceptualization, Data curation, Funding acquisition, Methodology, Validation, Writing – original draft. VK: Data curation, Formal analysis, Investigation, Software, Writing – review & editing. KW: Methodology, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. VK was supported by Toyota Motor Europe. Earlier research on neck stabilization has been performed as part of the Dutch NWO project 10736 (Torticollis).
Acknowledgments
Research on the comfort of automated driving has been performed as part of the German-Dutch Interreg project Interregional Automated Transport (I-AT) http://www.euregio.org/action/projects/item/126/i-at---interregional-automated-transport and the European project Deployment of Higher Automation (HiDrive) www.hi-drive.eu.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2023.1266345/full#supplementary-material
Footnotes
1. ^Similar models may address vestibular-ocular reflexes but this is beyond the scope of the current paper.
References
1. Carver, S, Kiemel, T, van der Kooij, H, and Jeka, JJ. Comparing internal models of the dynamics of the visual environment. Biol Cybern. (2005) 92:147–63. doi: 10.1007/s00422-004-0535-x
2. De Winkel, KN, Edel, E, Happee, R, and Bulthoff, HH. Multisensory interactions in head and body centered perception of verticality. Front Neurosci. (2020) 14:599226. doi: 10.3389/fnins.2020.599226
3. Oie, KS, Kiemel, T, and Jeka, JJ. Multisensory fusion: simultaneous re-weighting of vision and touch for the control of human posture. Brain Res Cogn Brain Res. (2002) 14:164–76. doi: 10.1016/S0926-6410(02)00071-X
4. Peterka, RJ. Sensorimotor integration in human postural control. J Neurophysiol. (2002) 88:1097–118. doi: 10.1152/jn.2002.88.3.1097
5. van der Kooij, H, Jacobs, R, Koopman, B, and Grootenboer, H. A multisensory integration model of human stance control. Biol Cybern. (1999) 80:299–308. doi: 10.1007/s004220050527
6. van der Kooij, H, Jacobs, R, Koopman, B, and van der Helm, F. An adaptive model of sensory integration in a dynamic environment applied to human stance control. Biol Cybern. (2001) 84:103–15. doi: 10.1007/s004220000196
7. van Dieen, JH, van Drunen, P, and Happee, R. Sensory contributions to stabilization of trunk posture in the sagittal plane. J Biomech. (2018) 70:219–27. doi: 10.1016/j.jbiomech.2017.07.016
8. van Drunen, P, Koumans, Y, van der Helm, FC, van Dieen, JH, and Happee, R. Modulation of intrinsic and reflexive contributions to low-back stabilization due to vision, task instruction, and perturbation bandwidth. Exp Brain Res. (2015) 233:735–49. doi: 10.1007/s00221-014-4151-2
9. Keshner, EA, Hain, TC, and Chen, KJ. Predicting control mechanisms for human head stabilization by altering the passive mechanics. J Vestib Res. (1999) 9:423–34. doi: 10.3233/VES-1999-9604
10. Keshner, EA. Vestibulocollic and cervicocollic control. Encyclopedia of neuroscience In: MD Binder, N Jirokawa, and U Windhorst, editors. Encyclopedia of neuroscience. Berlin, Heidelberg: Springer-Verlag (2009). 4222–4.
11. Cullen, KE. The vestibular system: multimodal integration and encoding of self-motion for motor control. Trends Neurosci. (2012) 35:185–96. doi: 10.1016/j.tins.2011.12.001
12. Goldberg, JM, and Cullen, KE. Vestibular control of the head: possible functions of the vestibulocollic reflex. Exp Brain Res. (2011) 210:331–45. doi: 10.1007/s00221-011-2611-5
13. Forbes, PA, Dakin, CJ, Vardy, AN, Happee, R, Siegmund, GP, Schouten, AC, et al. Frequency response of vestibular reflexes in neck, back, and lower limb muscles. J Neurophysiol. (2013) 110:1869–81. doi: 10.1152/jn.00196.2013
14. Peng, GC, Hain, TC, and Peterson, BW. A dynamical model for reflex activated head movements in the horizontal plane. Biol Cybern. (1996) 75:309–19. doi: 10.1007/s004220050297
15. Happee, R, de Bruijn, E, Forbes, PA, and van der Helm, FCT. Dynamic head-neck stabilization and modulation with perturbation bandwidth investigated using a multisegment neuromuscular model. J Biomech. (2017) 58:203–11. doi: 10.1016/j.jbiomech.2017.05.005
16. Zupan, LH, Merfeld, DM, and Darlot, C. Using sensory weighting to model the influence of canal, otolith and visual cues on spatial orientation and eye movements. Biol Cybern. (2002) 86:209–30. doi: 10.1007/s00422-001-0290-1
17. Carriot, J, Jamali, M, Brooks, JX, and Cullen, KE. Integration of canal and otolith inputs by central vestibular neurons is subadditive for both active and passive self-motion: implication for perception. J Neurosci. (2015) 35:3555–65. doi: 10.1523/JNEUROSCI.3540-14.2015
18. Angelaki, DE, Klier, EM, and Snyder, LH. A vestibular sensation: probabilistic approaches to spatial perception. Neuron. (2009) 64:448–61. doi: 10.1016/j.neuron.2009.11.010
19. Laurens, J, and Droulez, J. Bayesian processing of vestibular information. Biol Cybern. (2007) 96:389–404. doi: 10.1007/s00422-006-0133-1
20. de Winkel, KN, Katliar, M, and Bulthoff, HH. Causal inference in multisensory heading estimation. PLoS One. (2017) 12:e0169676. doi: 10.1371/journal.pone.0169676
21. de Winkel, KN, Katliar, M, Diers, D, and Bulthoff, HH. Causal inference in the perception of verticality. Sci Rep. (2018) 8:8. doi: 10.1038/s41598-018-23838-w
22. Reason, JT. Motion sickness adaptation - neural mismatch model. J Roy Soc Med. (1978) 71:819–29. doi: 10.1177/014107687807101109
23. Bles, W, Bos, JE, de Graaf, B, Groen, E, and Wertheim, AH. Motion sickness: only one provocative conflict? Brain Res Bull. (1998) 47:481–7. doi: 10.1016/S0361-9230(98)00115-4
24. Bos, JE, and Bles, W. Modelling motion sickness and subjective vertical mismatch detailed for vertical motions. Brain Res Bull. (1998) 47:537–42. doi: 10.1016/S0361-9230(98)00088-4
25. Nooij, SAE, Pretto, P, Oberfeld, D, Hecht, H, and Bulthoff, HH. Vection is the main contributor to motion sickness induced by visual yaw rotation: implications for conflict and eye movement theories. PLoS One. (2017) 12:e0175305. doi: 10.1371/journal.pone.0175305
26. Riccio, GE, and Stoffregen, TA. An ecological theory of motion sickness and postural instability. Ecol Psychol. (1991) 3:195–240. doi: 10.1207/s15326969eco0303_2
27. Stoffregen, TA, and Riccio, GE. An ecological theory of orientation and the vestibular system. Psychol Rev. (1988) 95:3–14. doi: 10.1037/0033-295X.95.1.3
28. Stoffregen, TA, and Smart, LJ Jr. Postural instability precedes motion sickness. Brain Res Bull. (1998) 47:437–48. doi: 10.1016/S0361-9230(98)00102-6
29. Smart, LJ Jr, Stoffregen, TA, and Bardy, BG. Visually induced motion sickness predicted by postural instability. Hum Factors. (2002) 44:451–65. doi: 10.1518/0018720024497745
30. Bos, JE. Nuancing the relationship between motion sickness and postural stability. Displays. (2011) 32:189–93. doi: 10.1016/j.displa.2010.09.005
31. Kennedy, RS, and Stanney, KM. Postural instability induced by virtual reality exposure: development of a certification protocol. Int J Hum Comput Interact. (1996) 8:25–47. doi: 10.1080/10447319609526139
32. Irmak, T, de Winkel, KN, Pool, DM, Bülthoff, HH, and Happee, R. Individual motion perception parameters and motion sickness frequency sensitivity in fore-aft motion. Exp Brain Res. (2021) 239:1727–45. doi: 10.1007/s00221-021-06093-w
33. Liu, H, Inoue, S, and Wada, T. Motion sickness modeling with visual vertical estimation and its application to autonomous personal mobility vehicles. The 33nd IEEE intelligent vehicles symposium (IV2022), Aachen, Germany (2022).
34. Wada, T, Kawano, J, Okafuji, Y, Takamatsu, A, and Makita, M, (Eds.) A computational model of motion sickness considering visual and vestibular information. 2020 IEEE international conference on systems, man, and cybernetics (SMC), Melbourne (2020) IEEE.
35. Clark, TK, Newman, MC, Karmali, F, Oman, CM, and Merfeld, DM. Mathematical models for dynamic, multisensory spatial orientation perception. Prog Brain Res. (2019) 248:65–90. doi: 10.1016/bs.pbr.2019.04.014
36. Newman, MC. A multisensory observer model for human spatial orientation perception [PhD thesis]:. Cambridge, MA: Massachusetts Institute of Technology (2009).
37. Irmak, T, Pool, DM, de Winkel, KN, and Happee, R. Validating models of sensory conflict and perception for motion sickness prediction. Biol Cybern. (2023) 117:185–209. doi: 10.1007/s00422-023-00959-8
38. Kotian, V, Irmak, T, Pool, DM, and Happee, R. The role of vision in sensory integration models for predicting motion perception and sickness. Experimental brain research. In press.
39. Happee, R, de Bruijn, E, Forbes, PA, van Drunen, P, van Dieën, JH, and van der Helm, FCT. Neck postural stabilization, motion comfort, and impact simulation. DHM Posturogr. (2019):243–60. doi: 10.1016/B978-0-12-816713-7.00019-2
40. Guedry, FE. Psychophysics of vestibular sensation. HH Kornhuber, Handbook of sensory physiology vestibular system part 2: psychophysics applied aspects and general interpretations. VI. Berlin: Springer (1974). p. 3–154.
41. Mayne, R. A systems concept of the vestibular organs In: HH Kornhuber, editor. Handbook of sensory physiology. VI. Berlin: Springer (1974). 493–580.
42. Correia Gracio, BJ, de Winkel, KN, Groen, EL, Wentink, M, and Bos, JE. The time constant of the somatogravic illusion. Exp Brain Res. (2013) 224:313–21. doi: 10.1007/s00221-012-3313-3
43. Bos, JE, and Bles, W. Theoretical considerations on canal-otolith interaction and an observer model. Biol Cybern. (2002) 86:191–207. doi: 10.1007/s00422-001-0289-7
44. Peng, GCY, Hain, TC, and Peterson, BW, (Eds.) How is the head held up? Modeling mechanisms for head stability in the sagittal plane. Proceedings of the 18th annual international conference of the IEEE engineering in medicine and biology society, 18, Pts 1–5; (1997).
45. Peng, GC, Hain, TC, and Peterson, BW. Predicting vestibular, proprioceptive, and biomechanical control strategies in normal and pathological head movements. IEEE Trans Biomed Eng. (1999) 46:1269–80. doi: 10.1109/10.797986
46. Fard, MA, Ishihara, T, and Inooka, H. Dynamics of the head-neck complex in response to the trunk horizontal vibration: modeling and identification. J Biomech Eng. (2003) 125:533–9. doi: 10.1115/1.1589777
47. Van Ee, CA, Nightingale, RW, Camacho, DL, Chancey, VC, Knaub, KE, Sun, EA, et al. Tensile properties of the human muscular and ligamentous cervical spine. Stapp Car Crash J. (2000) 44:85–102.
48. Yoganandan, N, Pintar, FA, and Cusick, JF. Biomechanical analyses of whiplash injuries using an experimental model. Accident Anal Prev. (2002) 34:663–71. doi: 10.1016/S0001-4575(01)00066-5
49. Stemper, BD, Yoganandan, N, and Pintar, FA. Validation of a head-neck computer model for whiplash simulation. Med Biol Eng Comput. (2004) 42:333–8. doi: 10.1007/BF02344708
50. Meijer, R, Broos, J, Elrofai, H, De Bruijn, E, Forbes, P, and Happee, R, editors. Modelling of bracing in a multi-body active human model. 2013 IRCOBI conference proceedings – international research council on the biomechanics of injury (2013).
51. Brolin, K, Hedenstierna, S, Halldin, P, Bass, C, and Alem, N. The importance of muscle tension on the outcome of impacts with a major vertical component. Int. J. Crashworth. (2008) 13:487–98. doi: 10.1080/13588260802215510
52. Wittek, A, Kajzer, J, and Haug, E. Hill-type muscle model for analysis of mechanical effect of muscle tension on the human body response in a car collision using an explicit finite element code. JSME Int. J. Ser. A Solid Mech. Mater. Eng. (2000) 43:8–18. doi: 10.1299/jsmea.43.8
53. Chancey, VC, Nightingale, RW, Van Ee, CA, Knaub, KE, and Myers, BS. Improved estimation of human neck tensile tolerance: reducing the range of reported tolerance using anthropometrically correct muscles and optimized physiologic initial conditions. Stapp Car Crash J. (2003) 47:135–53.
54. Hedenstierna, S. 3D finite element modeling of cervical musculature and its effect on neck injury prevention. PhD thesis. Stockholm: KTH (2008).
55. Almeida, J, Fraga, F, Silva, M, and Silva-Carvalho, L. Feedback control of the head-neck complex for nonimpact scenarios using multibody dynamics. Multibody Syst Dyn. (2009) 21:395–416. doi: 10.1007/s11044-009-9148-4
56. Meyer, F, Bourdet, N, Gunzel, K, and Willinger, R. Development and validation of a coupled head-neck FEM – application to whiplash injury criteria investigation. Int. J. Crashworthiness. (2013) 18:40–63. doi: 10.1080/13588265.2012.732293
57. Hedenstierna, S, and Halldin, P. How does a three-dimensional continuum muscle model affect the kinematics and muscle strains of a finite element neck model compared to a discrete muscle model in rear-end, frontal, and lateral impacts. Spine. (2008) 33:E236–45. doi: 10.1097/BRS.0b013e31816b8812
58. Hedenstierna, S, Halldin, P, and Brolin, K. Evaluation of a combination of continuum and truss finite elements in a model of passive and active muscle tissue. Comput Methods Biomech Biomed Engin. (2008) 11:627–39. doi: 10.1080/17474230802312516
59. Meyer, F, and Willinger, R. Three-year-old child head-neck finite element modelling: simulation of the interaction with airbag in frontal and side impact. Int J Veh Saf. (2009) 4:285–99. doi: 10.1504/IJVS.2009.032757
60. Meyer, F, Bourdet, V, Willinger, R, Legall, F, and Deck, C. Finite element modelling of the human head-neck: modal analysis and validation in the frequency domain. Int J Crashworth. (2004) 9:535–45. doi: 10.1533/ijcr.2004.0309
61. Correia, MA, McLachlin, SD, and Cronin, DS. Vestibulocollic and cervicocollic muscle reflexes in a finite element neck model during multidirectional impacts. Ann Biomed Eng. (2021) 49:1645–56. doi: 10.1007/s10439-021-02783-2
62. Correia, MA, McLachlin, SD, and Cronin, DS. Optimization of muscle activation schemes in a finite element neck model simulating volunteer frontal impact scenarios. J Biomech. (2020) 104:109754. doi: 10.1016/j.jbiomech.2020.109754
63. Putra, IPA, Iraeus, J, Sato, F, Svensson, MY, Linder, A, and Thomson, R. Optimization of female head-neck model with active reflexive cervical muscles in low severity rear impact collisions. Ann Biomed Eng. (2021) 49:115–28. doi: 10.1007/s10439-020-02512-1
64. Putra, IPA, Iraeus, J, Thomson, R, Svensson, MY, Linder, A, and Sato, F. Comparison of control strategies for the cervical muscles of an average female head-neck finite element model. Traffic Inj Prev. (2019) 20:S116–22. doi: 10.1080/15389588.2019.1670818
65. Putra, IPA, and Thomson, R. Analysis of control strategies for VIVA OpenHBM with active reflexive neck muscles. Biomech Model Mechanobiol. (2022) 21:1731–42. doi: 10.1007/s10237-022-01616-y
66. de Jager, MKJ. Mathematical head-neck models for acceleration impacts. [PhD Thesis]. Netherlands: Eindhoven University of Technology (1996).
67. van der Horst, MJ. Human head and neck response in frontal, lateral, and impact loading. [PhD Thesis]. Netherlands: Eindhoven University of Technology (2002).
68. de Bruijn, E, van der Helm, FCT, and Happee, R. Analysis of isometric cervical strength with a nonlinear musculoskeletal model with 48 degrees of freedom. Multib Syst Dyn. (2015) 36:339–62. doi: 10.1007/s11044-015-9461-z
69. Borst, J, Forbes, PA, Happee, R, and Veeger, DH. Muscle parameters for musculoskeletal modelling of the human neck. Clin Biomech (Bristol, Avon). (2011) 26:343–51. doi: 10.1016/j.clinbiomech.2010.11.019
70. Nesti, A, de Winkel, K, and Bulthoff, HH. Accumulation of inertial sensory information in the perception of whole body yaw rotation. PLoS One. (2017) 12:e0170497. doi: 10.1371/journal.pone.0170497
71. Diaz-Artiles, A, and Karmali, F. Vestibular precision at the level of perception, eye movements, posture, and neurons. Neuroscience. (2021) 468:282–320. doi: 10.1016/j.neuroscience.2021.05.028
72. Forbes, PA, de Bruijn, E, Schouten, AC, van der Helm, FC, and Happee, R. Dependency of human neck reflex responses on the bandwidth of pseudorandom anterior-posterior torso perturbations. Exp Brain Res. (2013) 226:1–14. doi: 10.1007/s00221-012-3388-x
73. Forbes, PA. Heads up – sensorimotor control of the head-neck system. PhD thesis. The Netherlands: Delft University of Technology (2014).
74. Keshner, EA, Cromwell, RL, and Peterson, BW. Mechanisms controlling human head stabilization. II. Head-neck dynamics during random rotations in the vertical plane. J Neurophysiol. (1995) 73:2302–12. doi: 10.1152/jn.1995.73.6.2302
75. Stensdotter, AK, Dinhoff Pedersen, M, Meisingset, I, Vasseljen, O, and Stavdahl, O. Mechanisms controlling human head stabilization during random rotational perturbations in the horizontal plane revisited. Physiol Rep. (2016) 4. doi: 10.14814/phy2.12745
76. Mirakhorlo, M, Kluft, N, Shyrokau, B, and Happee, R. Effects of seat back height and posture on 3D vibration transmission to pelvis, trunk and head. Int J Ind Ergon. (2022) 91:103327. doi: 10.1016/j.ergon.2022.103327
77. Irmak, T, Pool, DM, and Happee, R. Investigation of objective and subjective responses to motion sickness: the group and the individual. Exp Brain Res. (2020) 239:515–31. doi: 10.1007/s00221-020-05986-6
78. Keshner, FA, and Peterson, BW. Mechanisms controlling human head stabilization. I. Head-neck dynamics during random rotations in the horizontal plane. J Neurophysiol. (1995) 73:2293–301. doi: 10.1152/jn.1995.73.6.2293
79. Peterson, BW, Goldberg, J, Bilotto, G, and Fuller, JH. Cervicocollic reflex: its dynamic properties and interaction with vestibular reflexes. J Neurophysiol. (1985) 54:90–109. doi: 10.1152/jn.1985.54.1.90
80. Schneider, AD, Jamali, M, Carriot, J, Chacron, MJ, and Cullen, KE. The increased sensitivity of irregular peripheral canal and otolith vestibular afferents optimizes their encoding of natural stimuli. J Neurosci. (2015) 35:5522–36. doi: 10.1523/JNEUROSCI.3841-14.2015
81. Colebatch, JG, Halmagyi, GM, and Skuse, NF. Myogenic potentials generated by a click-evoked vestibulocollic reflex. J Neurol Neurosurg Psychiatry. (1994) 57:190–7. doi: 10.1136/jnnp.57.2.190
82. Forbes, PA, Siegmund, GP, Happee, R, Schouten, AC, and Blouin, JS. Vestibulocollic reflexes in the absence of head postural control. J Neurophysiol. (2014) 112:1692–702. doi: 10.1152/jn.00343.2014
83. Clement, G, Moore, ST, Raphan, T, and Cohen, B. Perception of tilt (somatogravic illusion) in response to sustained linear acceleration during space flight. Exp Brain Res. (2001) 138:410–8. doi: 10.1007/s002210100706
84. Holly, JE, Vrublevskis, A, and Carlson, LE. Whole-motion model of perception during forward- and backward-facing centrifuge runs. J Vestib Res. (2008) 18:171–86. doi: 10.3233/VES-2008-18401
85. Merfeld, DM, Zupan, LH, and Gifford, CA. Neural processing of gravito-inertial cues in humans. II. Influence of the semicircular canals during eccentric rotation. J Neurophysiol. (2001) 85:1648–60. doi: 10.1152/jn.2001.85.4.1648
86. Seidman, SH, Telford, L, and Paige, GD. Tilt perception during dynamic linear acceleration. Exp Brain Res. (1998) 119:307–14. doi: 10.1007/s002210050346
87. Turner, M, and Griffin, MJ. Motion sickness in public road transport: the relative importance of motion, vision and individual differences. Br J Psychol. (1999) 90:519–30. doi: 10.1348/000712699161594
88. Keshner, EA. Head-trunk coordination during linear anterior-posterior translations. J Neurophysiol. (2003) 89:1891–901. doi: 10.1152/jn.00836.2001
89. Keriven Serpollet, D, Hartnagel, D, James, Y, Buffat, S, Vayatis, N, Bargiotas, I, et al. Tilt perception is different in the pitch and roll planes in human. Physiol Rep. (2023) 11:e15374. doi: 10.14814/phy2.15374
90. Goldberg, J, and Peterson, BW. Reflex and mechanical contributions to head stabilization in alert cats. J Neurophysiol. (1986) 56:857–75. doi: 10.1152/jn.1986.56.3.857
91. Gillies, GT, Broaddus, WC, Stenger, JM, and Taylor, AG. A biomechanical model of the craniomandibular complex and cervical spine based on the inverted pendulum. J Med Eng Technol. (1998) 22:263–9. doi: 10.3109/03091909809010009
92. Fard, MA, Ishihara, T, and Inooka, H. Identification of the head-neck complex in response to trunk horizontal vibration. Biol Cybern. (2004) 90:418–26. doi: 10.1007/s00422-004-0489-z
93. Liang, CC, and Chiang, CF. Modeling of a seated human body exposed to vertical vibrations in various automotive postures. Ind Health. (2008) 46:125–37. doi: 10.2486/indhealth.46.125
94. Reynolds, JS, Blum, D, and Gdowski, GT. Reweighting sensory signals to maintain head stability: adaptive properties of the cervicocollic reflex. J Neurophysiol. (2008) 99:3123–35. doi: 10.1152/jn.00793.2007
95. ISO-2631-1. Mechanical vibration and shock: evaluation of human exposure to whole-body vibration (1997).
96. Dalton, BH, Rasman, BG, Inglis, JT, and Blouin, JS. The internal representation of head orientation differs for conscious perception and balance control. J Physiol. (2017) 595:2731–49. doi: 10.1113/JP272998
97. Kotian, V, Pool, DM, and Happee, R. Modelling individual motion sickness accumulation in vehicles and driving simulators. Driving simulator conference, Antibes, France (2023).
Keywords: neck, stabilization, sensory integration, sensory conflict, motion sickness
Citation: Happee R, Kotian V and De Winkel KN (2023) Neck stabilization through sensory integration of vestibular and visual motion cues. Front. Neurol. 14:1266345. doi: 10.3389/fneur.2023.1266345
Edited by:
Jean Laurens, Max Planck Society, GermanyReviewed by:
Robert Peterka, United States Department of Veterans Affairs, United StatesGianfranco Bosco, University of Rome Tor Vergata, Italy
Copyright © 2023 Happee, Kotian and De Winkel. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Riender Happee, r.happee@tudelft.nl