Skip to main content

ORIGINAL RESEARCH article

Front. Virtual Real., 05 November 2024
Sec. Virtual Reality and Human Behaviour
This article is part of the Research Topic A Metaverse for the Good: Design, Application and Understanding View all 7 articles

Estimation of horizontal spatial specifications for ideal head-mounted displays in practical conditions

  • Science and Technology Research Laboratories, Japan Broadcasting Corporation, Tokyo, Japan

This study presents the horizontal spatial specifications required for developing ideal head-mounted displays (HMDs) that provide visual experiences indistinguishable from those obtained without wearing HMDs. We investigated the minimum specifications for pixel density and field of view (FoV) such that users cannot perceive any degradation in these aspects. Conventional studies have measured visual acuity in the periphery and FoV size without eye movement. However, these results are not sufficient to determine the spatial specifications because users may notice a degradation in the quality of images displayed in the periphery when shifting their gaze away from the front. In this study, we measured visual characteristics under practical conditions, wherein participants moved their eyes naturally in coordination with their head movements, as observed when viewing natural scenes. Using a cylindrical display that covered the participant’s entire horizontal visual field with high resolution images, we asked participants to identify degraded spatial resolution or narrowed FoV. Results showed that ideal HMDs do not have to provide spatial frequency components above 0.5 and 2 cycles per degree outside the visual fields of approximately 210° and 120° horizontally, respectively. These results suggest that a resolution equivalent to one-fifteenth of the frontal area is sufficient for the lateral areas of the head. In addition, we found that ideal HMDs require a FoV of at least 240°. This is considerably smaller than a naive estimate of 310° based on previous studies, obtained by integrating the FoV size during fixation and the movable range of the eyes. We believe that these findings will be helpful in developing ideal HMDs.

1 Introduction

What is the exact goal of head-mounted display (HMD) development? Visual experiences through HMDs have improved substantially since they were initially proposed (Sutherland, 1968). Here, we assume that the goal of evolving HMDs is to provide visual experiences that are indistinguishable from those obtained when users do not wear HMDs and refer to such HMDs as “ideal HMDs.” This study aimed to estimate the minimum spatial specifications required for developing ideal HMDs, focusing in particular on pixel density and field of view (FoV).

Ideal HMDs do not need to uniformly fill pixels with high density and completely surround the user’s head with a display. The spatial specifications of the pixel density and FoV size can be degraded to the extent that users unable to perceive the degradation. For example, as the visual acuity deteriorates with retinal eccentricity (Anstis, 1974), the pixel density at the periphery of ideal HMDs can be suppressed relative to that at the center, which is in the frontal direction of the head. Additionally, previous results of perimetry tests, which determine the edge of the temporal visual field, have shown that people cannot perceive visual stimuli in the far periphery beyond 110° (Traquair, 1938; Lestak et al., 2021). Therefore, it is not necessary to cover the entire head with display panels but rather to present visual information in a limited FoV.

Our research question is to determine the pixel density and FoV required for developing ideal HMDs. To assess this, we measured visual characteristics in the periphery with natural eye movements that occur when using HMDs. In the next subsection, we first describe two representative coordinate systems: retina-centered and head-centered coordinate systems, which have been used to measure visual characteristics. We then justify why the head-centered coordinate system is appropriate for this study, as its origin remains relatively fixed to an HMD even when a user moves the eyes. Further details on eye movements are described in the following subsection. Finally, we present our hypotheses and study design.

1.1 Coordinate systems for measuring visual characteristics

1.1.1 Retina-centered coordinate system

Previous studies have primarily employed retina-centered coordinate systems to measure the capabilities of the human visual system (Traquair, 1938; Anstis, 1974; Lestak et al., 2021). A retina-centered coordinate system is typically represented as a polar coordinate system. Its origin corresponds to the direction of the visual axis, as indicated by the dashed arrows in Figure 1. The radial coordinate reflects the retinal eccentricity of the visual target, as shown by the dashed lines. The angular coordinate specifies the direction of the visual target, e.g., temporal, nasal, upward, or downward. For example, Figure 1A shows a boundary at a certain eccentricity in the temporal direction, depicted by the dashed line, beyond which a person cannot perceive a blurred image.

Figure 1
www.frontiersin.org

Figure 1. Retina and head-centered coordinate systems for assessing visual characteristics in the periphery. (A) Conventional measurements in retina-centered coordinate system. The dashed arrow represents the origin of the retina-centered coordinate system, corresponding to the direction of fixation. The dashed line indicates the minimum retinal eccentricity beyond which the quality degradation is unnoticeable. (B) Eye movements expected in actual use. When the eyes move, the quality degradation becomes more noticeable. (C) Measurements in head-centered coordinate system. As shown in (C-left), the origin of the head-centered coordinate system is fixed, as represented by the solid arrow even when the eyes move. The solid line indicates the minimum eccentricity from the origin for users not noticing the quality degradation. As shown in (C-right), our investigation focused on determining the minimum intact region beyond which users do not notice quality degradation on both the left and right sides.

However, the capabilities measured in the retina-centered coordinate system are insufficient to estimate the specifications required for ideal HMDs because the retina-centered coordinate system moves relative to the display along with eye movements, as shown in Figure 1B. Eye movements while using HMDs enlarge the perceptual FoV. Consequently, the degraded regions on the display screen captured by the periphery of the visual field become more noticeable. Therefore, eye movements must be considered when measuring the visual capability to estimate the specifications required for ideal HMDs.

1.1.2 Head-centered coordinate system

We employed a head-centered coordinate system to estimate the specifications required for developing ideal HMDs. The head-centered coordinates also represent the direction and eccentricity of the stimulus, but the origin is not on the retina; instead, it is relative to a specific point on the head, usually taken to be the midpoint of the eyes, i.e., the point of the cyclopean eye. The origin of the head-centered coordinate system corresponds to the frontal direction of the head, as shown by the solid arrows in Figure 1C. As the HMD is affixed to the head, the head-centered coordinate system is fixed relatively to the HMD and is unaffected by the directions of the eye. The solid line in Figure 1C-left represents the eccentricity threshold, beyond which the blur is unnoticeable even with eye movements. We determined the size of the minimum intact region, defined as twice the estimated eccentricity threshold, as shown in Figure 1C-right. With the minimum intact region, the display does not induce any perceptual quality degradation in terms of pixel density or FoV.

1.2 Eye movements during measurements

(Harasawa et al., 2021; Harasawa et al., 2023 measured visual characteristics in the head-centered coordinate system with maximal eye movements. In their experiments, participants were asked whether they perceived blurred images or blinking LEDs in their periphery with their eyes deviated as far laterally as possible. Results showed that the region beyond a horizontal FoV of approximately 150° width can filter the spatial frequency components above two cycles per degree (cpd) without being noticed (Harasawa et al., 2023). In addition, the participants were able to identify the LED blinking up to a maximum horizontal FoV of approximately 250° width (Harasawa et al., 2021).

Actual measurements with eye movements are crucial for estimating the specifications required for ideal HMDs because predicted values obtained by compensating the size of the visual field obtained in the absence of eye movements with values for the possible range of eye movements would result in an overestimation. Specifically, the obtained FoV of 250° in the head-centered coordinate system (Harasawa et al., 2021) was approximately 60° smaller than the naively estimated FoV of 310°, which was derived by the linear summation of the previous observations: 220° [the horizontal FoV in the retina-centered coordinate system (Traquair, 1938)] and 90° [the horizontally movable range of the eye (Hanif et al., 2009)]. Plausible factors contributing to a reduction in FoV size may include anatomical structures of the face, such as the eyelids (Harasawa et al., 2021).

1.3 Hypotheses

We hypothesized that smaller intact regions would be required under practical conditions. The maximal eye movements in prior studies (Harasawa et al., 2021; 2023) are not common in the typical usage of HMDs. In natural viewing conditions, we usually move our heads to look at an object in the periphery in coordination with eye movements. This eye–head coordination contributes to suppressing the size of eye movements (Freedman and Sparks, 1997; Sidenmark and Gellersen, 2019). Consequently, we hypothesized that the required intact regions would be smaller for natural gaze shift in coordination with head rotation than for maximal eye movements.

In addition, spatially abrupt changes of the image at the boundary of the intact region may affect the noticeability of the quality degradations, as shown in Figure 4A. The human visual system is more sensitive to spatially steep variations than to gradual ones (Campbell and Robson, 1968). Therefore, we hypothesized that users would be more likely to notice a degradation in pixel density or a change in FoV size if the image at the boundary of the intact region transitioned abruptly into a blurred or black image.

1.4 Study design

In this study, we investigated the minimum horizontal specifications for ideal HMDs based on a head-centered coordinate system under practical conditions of HMD use. We were unable to use off-the-shelf HMDs for the investigation because they cannot present images in the far periphery. Instead, we used a cylindrical display capable of showing images horizontally in 360° with high pixel density. Participants naturally moved their eyes in coordination with their head movements. They observed computer-generated images of natural scenes likely to be displayed during actual usage. At the boundary between the intact and modified regions, there were two types of variations: steep and gradual.

We conducted two experiments. In Experiment 1, we estimated the eccentricity of the boundary above which users cannot identify the cutoff of spatial frequency components above a specific level. In Experiment 2, we estimated the size of FoV above which users cannot identify the absence of image, i.e., a region filled in with black. Additionally, in Experiment 2, we investigated the effect of maximal eye movements on FoV size by asking the participants to move their eyes as far as possible while resting their head on a chinrest.

The remainder of the paper is organized into five sections. Section 2 reviews related works, focusing on the design and development of HMDs. Section 3 describes the methods for Experiments 1 and 2, presented in Sections 3.1 and 3.2, respectively. Section 4 presents the results of each experiment in the same order (Sections 4.1 and 4.2). For quick reference to the results, readers can refer directly to the corresponding subsection. Section 5 provides a discussion, including limitations and directions for future work. Finally, Section 6 concludes our research.

2 Related work

2.1 Specifications of head-mounted displays

Previous studies have investigated the specifications of HMDs, such as pixel density, refresh rate, and latency, necessary to fulfill various objectives. Newman et al. (1998) introduced specifications of HMDs such as pixel density and refresh rate, focusing on performance in specialized tasks such as aggressive visual tracking, assuming actual use in flight. Jerald and Whitton (2009) investigated the imperceptible level of motion-to-photon latency, which is defined as the delay between head movement and its application to the visualization in an HMD. In this study, we focused on the specifications of the pixel density and FoV size such that users do not detect any degradation of these specifications.

Cuervo et al. (2018) introduced the necessary specifications to achieve “life-like virtual reality (VR),” a concept similar to “ideal HMDs” in terms of the criterion that users cannot detect any degradation in quality. However, their study employed retinal coordinates and did not account for eye movements. Our study employed head-centered coordinate system to estimate the specifications required for users not to perceive quality degradation.

2.2 Reduction of rendering resolution in the periphery of head-mounted displays

Foveated rendering saves computational resources by reducing the rendering resolution in the periphery (Guenter et al., 2012; Tan et al., 2018; Singh et al., 2023). Additionally, reducing resolution in the periphery helps alleviate discomfort. Hussain et al. (2021) found that peripheral blur, as a depth-of-field effect, significantly reduces cybersickness. Foveated rendering methods are typically classified into two categories: fixed foveated rendering (FFR) and dynamic foveated rendering (DFR). Without eye tracking, FFR presents an image with high resolution in the frontal direction of the head and gradually decreases the resolution as eccentricity increases in the head-centered coordinate system (Singh et al., 2023). By contrast, DFR leverages eye tracking to generate an image with high resolution in the region being gazed at while gradually decreasing the resolution of pixels away from that region. Using either method, ideal HMDs must have sufficient pixel density to display foveated images without the user perceiving a decrease in pixel density.

Previous quality assessment methods for foveated images are insufficient in revealing the specifications necessary for ideal HMDs. Wang et al. (2001) and Lee et al. (2002) proposed objective metrics to estimate the subjective quality of foveated images. However, these methods cannot predict whether the foveated images are indistinguishable from their original images. To achieve perceptually lossless foveated rendering, Swafford et al. (2016) proposed an objective metric called HDR-VDP2 (Mantiuk et al., 2011), which is based on the cortical magnification factor (Rovamo and Virsu, 1979; Virsu and Rovamo, 1979). Using a similar concept, Mantiuk et al. (2021) proposed an objective metric called FovVideoVDP, which models the perceptual characteristics of contrast sensitivity, cortical magnification, and contrast masking (Legge and Foley, 1980). These objective metrics were constructed based on subjective evaluations using a retina-centered coordinate system with the participants fixating on a specific point on a display placed in front of them. In this study, we used head-centered coordinate system with eye movements to estimate the pixel density required for ideal HMDs.

2.3 Expanding the field of view of head-mounted displays

Various methods have been proposed to expand the FoV of HMDs. Rakkolainen et al. (2017) proposed a method to expand the FoV by including modules consisting of a Fresnel lens and screens on the sides of a frontal display. Rakkolainen et al. (2016) and Ratcliff et al. (2020) achieved FoV expansion utilizing a curved Fresnel lens and curved microlens array, respectively. Xiao and Benko (2016) introduced SparseLightVR, which expands the FoV of off-the-shelf HMDs by arranging LEDs and diffusers at the periphery of the displays.

Although these methods can significantly expand the FoV of HMDs, the extent to which we need to expand the FoV for ideal HMDs remains unclear. A large horizontal FoV of 270° can be achieved for HMDs using curved Fresnel lenses (Rakkolainen et al., 2016). However, the cost of developing HMDs that surpass the human visual function does not commensurate with their benefits. Reducing the FoV to a size where users cannot perceive the edge of the display will result in smaller, lighter, and more power-efficient devices. This study aimed to determine the minimum horizontal FoV for ideal HMDs, considering the natural eye movements.

3 Methods

3.1 Experiment 1

In Experiment 1, we investigated the eccentricity of the boundary above which the participants could not distinguish between high- and low-resolution areas. Specifically, we estimated the size of the minimum intact region above which the cutoff of spatial frequency components above a certain frequency was imperceptible. The participants naturally moved their eyes in coordination with their heads while observing images of natural scenes.

3.1.1 Participants

We recruited 20 participants with visual acuity better than 20/20 without any correction. We assessed their visual acuity using a the vision inspection device based on a tumbling E chart (Screenoscope SS-3, TOPCON, Japan). Their mean age was 25.5 years old, with a standard deviation (SD) of 5.5. The female-to-male ratio was 1:1. The number of participants was determined based on the findings of a previous study (Miyashita et al., 2023). This study was reviewed and approved by the Ethics Committee of NHK Science and Technology Research Laboratories. All participants provided written informed consent to participate in the study.

3.1.2 Apparatus

As shown in Figure 2A, we used a cylindrical display that simulated an HMD with high pixel density and wide horizontal FoV. Currently available HMDs are inadequate for investigating the visual characteristics in the periphery because they lack sufficient pixel resolution and FoV. The cylindrical display consisted of twenty 4K panels (LG, 65EV5E) that were rotated vertically and curved along a cylindrical shape with a radius of 2.6 m. An observer placed at the center of the cylindrical display can view images with a spatial frequency of up to 60 cpd. This frequency exceeds the threshold of 30 cpd at which individuals with normal vision (20/20) can resolve the spatial details in an image. The cylindrical display inevitably showed vertical lines between the panels because each panel had 1-cm-wide bezels. By placing the panels such that the bezels of adjacent panels overlapped to minimize the gaps where no images were displayed, the gaps were suppressed to approximately 0.22° of visual angle. To ensure a continuous image across multiple panels, we created display images with bezel compensation to give viewers a seamless view even with the gaps. The experiment was conducted in a dimly lit room such that the light reflected from the displays was not obvious.

Figure 2
www.frontiersin.org

Figure 2. Experimental setup. (A) A cylindrical display that covers the participant’s 360° horizontal FoV with high-resolution images. Each participant sat at the center of the display. (B) A marker-attached helmet and motion-tracking cameras. The participants’ head positions and postures were captured using nine motion-tracking cameras mounted on the upper edge of the displays. (C) The intact region rotated in synchronization with the rotation of the head.

As shown in Figure 2B, we used nine motion trackers (OptiTrack, Prime X13) to capture the participants’ head positions and postures. The participants wore bicycle helmets attached with retroreflective markers. We simulated the displaying manner of HMDs when users rotated their heads, as shown in Figure 2C. We divided the display areas into two regions: the intact and the modified regions, and examined the position of the boundary. In the intact regions, original images with a spatial frequency of 60 cpd were presented, whereas in the modified region, blurred images without spatial frequency components above a certain threshold were presented. The horizontal center of the intact region was always aligned with the forward direction of the head by following the yaw angle of the participant’s head. The participants’ head positions and postures were recorded while they observed the stimuli, allowing us to investigate their head movement patterns.

3.1.3 Stimuli

Figure 3A shows the stimulus images of three virtual scenes with a 360° horizontal FoV. These scenes included both inside and outside locations as well as artificial and natural environments. Figure 3B shows the spatial frequency spectra of these images. The spectral power decreased with frequency, similar to that for images of real scenes (van der Schaaf and van Hateren, 1996). The area of the image that the participants observed in front of them at the beginning of each trial was randomized to avoid any bias caused by the choice of areas. Specifically, the areas were determined by horizontally rotating the 360° image in integer multiples of 72°, and the participants were presented with one of the areas, ensuring equal distribution of frequency (i.e., 20%) across each angular phase.

Figure 3
www.frontiersin.org

Figure 3. Computer-generated stimuli. (A) Stimulus images. Three images had a horizontal FoV of 360°. (B) Horizontal spatial frequency spectra of the three stimulus images. The vertical axis shows the spectral power, computed as the logarithm of the amplitude in the spatial frequency domain.

Prior to the experiment, we rendered images of these scenes offline and saved them as files using Unity 3D software (Unity Technologies, United States). For each scene, we generated 20 images, each image corresponding to a single panel. The shape of the panels was compensated for in the rendering process. That is, an image rendered for a flat panel appears to be distorted when displayed on a curved panel. To address this issue, we applied pixel warping. Specifically, we filled each pixel of the 4K (2,160 × 3,840) curved panel by selecting the corresponding pixels from an 8K (4,320 × 7,680) virtual flat panel, which was placed tangent to the curved panel.

We created blurred images to be displayed in the modified region by applying low-pass filters (LPFs) of spatial frequency to the original images. We created two levels of blurred images using LPFs with cutoff frequencies of 0.5 and two cpd. The images produced with the 0.5-cpd cutoff LPF were more blurred than those produced with the 2-cpd cutoff LPF. We generated the LPFs of the finite impulse response (FIR) using Filter Designer of MATLAB (MathWorks, United States). We designed the filters such that the amplitude at the cutoff frequency decreased by half of the input, i.e., −6 dB, and the stopband attenuation reached −80 dB. Because a steep cutoff of spectra in the frequency domain can lead to ringing artifacts in the spatial domain (Wang and Healy, 2023), we designed smoother transitions to ensure that low-pass filtering was performed without ring-back artifacts. The kernel sizes were 459 × 459 and 161 × 161 for the LPFs with the 0.5 and 2 cpd cutoffs, respectively. The specific procedure to generate the low-pass filtered images consisted of three steps: 1) increasing the FoV of the original images to compensate for the size of the LPF kernel, 2) generating the filtered images by convolving the LPF kernel with the designed filter parameters onto the original images, and 3) cropping the edges of the filtered images to generate display images that matched the display resolution. This procedure ensured that the images satisfied the designed filter characteristics even at their edges.

At the border between the intact and modified regions, we blended the original and blurred images following two methods: steep and gradual (Figure 4A), wherein the spatial variations in quality degradation occur steeply and gradually, respectively. The pixel values of the original and blurred images, gorg and gmod, respectively, were blended as

g=wgorg+1wgmod,(1)

where the variable w represents a weight ranging from 0 to 1, indicating a fraction of the original image. As shown in Figure 4B, we varied the w values with the horizontal eccentricity (that is, the angular distance from the frontal direction of the head) to blend the two images at the intact-modified border. We used the step and sigmoid functions for steep and gradual blending, respectively. Following a previous study (Miyashita et al., 2023), the slope of the sigmoid function was set to one at the midpoint of the transition. In the gradual blending method, quality degradation starts from the intact region (Figure 4B), and the angle of transition from the original to the blurred image was approximately 13.8°.

Figure 4
www.frontiersin.org

Figure 4. Transition at the border between the intact and modified regions. (A) Schema of the steep and gradual blending methods. The orange triangles indicate the position of the border between the intact and modified regions. (B) Plots of the weight values for image blending versus eccentricity. The stimulus images are generated by blending the original and quality-degraded (blurred) images according to the values of w.

Following the method of constant stimuli, we varied the horizontal angle of the intact region at 11 levels, and the participants were shown the images at each level ten times. When using the 0.5-cpd cutoff LPF, the intact region ranged from 200° to 250° and 188.8°–238.8° at 11 levels with a 5° interval for steep and gradual blending, respectively. When using the 2-cpd cutoff LPF, the intact region ranged from 70° to 170° and 83.8°–183.8° at 11 levels with a 10° interval for steep and gradual blending, respectively. The ranges of the intact region varied depending on the cutoff frequency because the smaller the amount of blurring, the narrower the intact region tends to be. These values were determined based on the preliminary experiments to optimize the number of trials. Each participant observed 330 stimuli (3 scenes × 11 angular levels × 10 repetitions) in a session at a single cutoff frequency with either the steep or gradual blending methods.

3.1.4 Procedure

As shown in Figure 5A, the participants performed a two-interval forced choice (2IFC) task (see the supplementary video for a better understanding). They sequentially observed two images of the same scene, one of which had a modified region and the other had only a 360° intact region, i.e., the original image. They chose an interval in which they perceived the modified region was contained.

Figure 5
www.frontiersin.org

Figure 5. Procedure of the 2IFC task. (A) Head-free condition. (B) Head-fixed condition. The orange triangles indicate the borders of the intact and modified regions. These triangles were drawn for better understanding and were not displayed in the experiments.

Each trial consisted of the following five steps, i) – v), as shown in Figure 5. Before observing each stimulus, the participants adjusted their head positions and postures (Figure 5Ai and Figure 6A). This ensured that the head remained within a 50 mm virtual cube, centered on the cylindrical display. The virtual cube was positioned at a height of approximately 1.2 m, aligning with the height of the center of each curved panel. The adjustment also ensured that the head was oriented within 5° of the frontal direction, as indicated by the red dots in Figure 6A. The participants used two indicators for the adjustment: a white circle and a white cross. The white circle indicated the current position, while the white cross indicated the current orientation of the head. The participants precisely aligned the two indicators with a red dot. Once the adjustment was completed, the color of the dot changed from red to blue.

Figure 6
www.frontiersin.org

Figure 6. Adjustment of the head position and posture. (A) Head adjustment by aligning the white cross and white circle over the red dot. The red dot indicates the frontal direction of the cylindrical display. The white circle’s position on the display indicates both the horizontal and vertical head positions, while its size represents the front-back positioning of the head. The white cross points in the direction of the head. The adjustments are deemed complete when the color of the red dot changes to blue. (B) Definition of the facial landmarks. We calibrated the head-centered coordinate system using four facial landmarks. The relationship between the coordinate system and positions of markers attached to the helmet (Figure 2B) was recorded before each experiment.

The participants observed an image while naturally moving their heads and eyes. As shown in Figure 5Aii), they initiated the observation of each stimulus by pressing the “A” button on a gaming pad at their preferred timing after completing the adjustment of the head position and orientation. Subsequently, a scene image was presented with or without the modified region while the blue dot remained as a fixation point (Figure 5Aiii). After 500 m, a bullseye target with a diameter of 1° appeared either 60° to the right or left of the fixation point (Figure 5Aiv). We placed the target at the 60° position because the saccadic amplitude is near its peak when the target appears at this eccentricity (Guitton and Volle, 1987). The fixation point disappeared simultaneously with the appearance of the target. The appearance of the target was indicated by a 1 kHz beep lasting for 200 m. The participants were instructed to look at the target with natural head movements. The scene image turned black 1,500 m after the target’s appearance (Figure 5Av). When the head was oriented toward the target within 5°, the target disappeared (Figure 5Avi). The second stimulus was observed in the same manner as the first one. After observing the two stimuli, the participants were asked to select one stimulus that they felt the presence of the modified region using the up and down keys on the game pad.

The participants received instructions before the experiment. We clarified that the modified region rotates synchronously with head rotations using a video of a trial captured with a wide-angle camera. This video also demonstrated how fast they rotate their heads during a trial. They were instructed to move their heads slowly, as they would when looking at surrounding objects without rushing. They were instructed not to rotate their chair seats during the trial. We ensured that the participants had a clear understanding of how to execute the task through a practice session of 30 trials.

Following the instructions, we calibrated the head-centered coordinate system based on the facial landmarks of each participant. The head-centered coordinate system was established using four retroreflective adhesives attached to the faces of the participants (Figure 6B). These markers were attached to four facial landmarks: the nasion, left and right preauricular points, and inferior margin of the right orbit. As shown in Figure 6B-right, the head-centered coordinate system was oriented parallel to the Frankfurt plane, approximately corresponding to the true horizontal plane when the visual axis was horizontal (Leitao and Nanda, 2000). The height and depth of the origin of the coordinate system corresponded to those of the nasion and inferior margin of the right orbit, respectively. Finally, we established the relationships between the head-centered coordinate system and helmet makers, the positions of which were streamed to the rendering computers.

Following a between-participant design, the 20 participants were evenly assigned to two groups for the steep and gradual border types. Each participant completed two sessions, each consisting of either 0.5 cpd or two cpd conditions. The two sessions took place on different days. Participants completed the 0.5 cpd session first, followed by the two cpd session.

3.1.5 Analysis

We estimated the minimum horizontal angle of the intact region as the threshold at which the participants could not distinguish the presence or absence of the modified region. For each participant, we obtained a response in each trial. Let i be the trial number and yi1,0 the participant’s response, with one and 0 denoting a correct and incorrect response, respectively. A correct response indicated that the participant correctly chose the stimulus interval with the modified region as modified in the 2IFC trial. We assumed that the responses of each trial followed a Bernoulli distribution with probabilities πi, as described using Equation 2.

yiBernoulliπi.(2)

We introduce μ to be the angular threshold above which a participant cannot detect any quality degradation even in the presence of the modified region. Let xi be the angular position of the border between the intact and modified regions of the stimulus presented in the ith trial, we expected that a participant would not notice stimulus modification if the following condition was fulfilled,

xiμ+εi,(3)

where εi is a random number following the normal distribution N0,σ2, representing the variability in the participants’ perceptions.

Equation 3 indicates that the probability of not noticing stimulus modification, qi, is described as the cumulative distribution function (CDF) for normal distributions. The probability qi corresponding to xi is calculated as,

qi=xiNz|μ,σ2dz.(4)

An incorrect response indicated that the participant failed to notice the modified region and randomly chose the incorrect stimulus interval (that is, the stimulus interval that was not modified). This probability is described as qi×1/2. Therefore, the probability of obtaining the correct response, πi, is

πixi,μ,σ=1qi2.(5)

We estimated the parameters θ=μ,σ for each participant, given the stimulus and response data for all trials, D=x1,x2,,x330;y1,y2,,y330. Following Bayes’ theorem, the posterior distribution of θ is proportional to the product of the likelihood function and the prior distribution,

fθ|DfD|θfθ.(6)

The likelihood function fDθ can be expressed using Equations 2, 4 and 5.

fDθ=i=1330πixi,μ,σyi1πixi,μ,σ1yi.(7)

We adopted a uniform distribution as the prior distribution fθ and obtained the posterior distribution of θ (Equation 6) for each participant using Bayesian inference with the Markov chain Monte Carlo (MCMC) method (Hoffman and Gelman, 2014) implemented in RStan (Stan Development Team, 2023). To assess the reliability of the parameters, we calculated the 95% Bayesian credible interval, which corresponds to the range between the 2.5th and 97.fifth percentiles of the posterior distribution.

When estimating the parameters θ=μ,σ, we corrected the size of the intact region, described by xi, based on the motion-to-photon latency of our cylindrical display system; the system exhibited latency between head rotation and its visualization response. When the latency was present, the substantive size of the intact region was equivalent to that with a narrower size. Once the amount of latency is known, the intact region size correction can be determined. Figure 7A-left shows the ideal case without latency, wherein the size of the intact region corresponds to the specified value throughout the head rotation. By contrast, in Figure 7A-right with latency, the modified region intrudes into the intact region in the same direction as head rotation, resulting in a narrower substantive intact region.

Figure 7
www.frontiersin.org

Figure 7. Compensation for motion-to-photon latency in the display system. (A) Substantive intact region corrected by the latency. The intact region was inherently narrowed if the latency existed. (B) Relationship between the actual head direction, the recorded head direction, and the visualization. The plot shows an example of the recorded yaw angles of a participant’s head during a stimulus observation with a target shown on the right. The latency can be estimated based on the maximum difference between the actual head angle and the visualization.

Before the experiment, we measured the motion-to-photon latency using a video camera (SONY, DSC-RX0) operating at 960 Hz. We captured the motion of retroreflective markers and the response images displayed on the cylindrical display. By analyzing the captured video, we obtained the motion-to-photon latency by counting frames from when the retroreflective markers began to move to when the images on the cylindrical display began to move. We iterated this procedure 100 times to estimate the latency as 136.4 ± 6.1 m (mean ± SD).

We corrected the size of the intact region, xi, by using the estimated latency and the head rotation data recorded during the trials. We recorded the yaw angle of the head while the participants observed the stimuli. The bold line in Figure 7B shows the temporal changes of the yaw angle. Because the latency was constant during the experiment, the curves for the actual head rotation angle and the corresponding visualization timing should both be parallel to the recorded curve, as shown by the blue solid, the red dashed, and the bold lines in Figure 7B. The amount of motion-to-photon latency can then be identified as the temporal gap between the actual transitions of the head and the intact region. By plotting the actual and visualization curves with the previously estimated latency, we obtained the maximum intrusion angle Imax as the maximum difference in yaw angles between the head and the intact region during stimulus observation. The size of the substantive intact region is minimum at Imax. We corrected the size of the intact region using Equation 8.

xi=xi2Imax.(8)

We subtracted twice the maximum intrusion angle from the intact region for both the left and right sides, assuming symmetry of the intact region, as shown by gray lines in Figure 7A-right.

In addition, we investigated the effect of border type (steep and gradual blending of the intact and modified images, shown in Figure 4A) on threshold (μ) of the size of the intact region. We estimated μ for each participant and tested the difference in μ between the two groups: steep and gradual, using one-sided Welch’s t-tests. The significance level was set at p=0.05. In the statistical test, we excluded μ for some participants that was larger or smaller than the size range of the intact regions with latency compensation, xi. This exclusion was necessary because the lack of data around the thresholds results in unreliable estimation (Levitt, 1971). Consequently, in the t-test for each condition, we did not use data from participants who showed that μ was outside the range of latency-corrected intact region.

3.2 Experiment 2

In Experiment 2, we investigated the size of FoV for ideal HMDs. The FoV was represented by the horizontal angle of the intact region above which the pixels were filled with black. There was no blur region in Experiment 2. Following the same scheme as in Experiment 1, we examined the effects of border type (steep and gradual) on the horizontal FoV. This experiment was conducted in head-free and head-fixed conditions. In the head-free condition, the participants performed the task with natural gaze shifts in coordination with the head movements. In the head-fixed condition, they performed the same task but without head movements, while shifting their gaze maximally to the temporal side.

3.2.1 Participants

We recruited 20 participants, 18 of whom participated in Experiment 1. We assessed their visual acuity using a vision inspection device a tumbling E chart (Screenoscope SS-3, TOPCON, Japan). The visual acuity of all participants was better than 20/20 without any correction. Their mean age was 25.3 years old, with SD = 5.7. The female-to-male ratio was 1:1. All participants provided written informed consent to participate in the study.

3.2.2 Apparatus

For Experiment 2, we used the same cylindrical display and motion-capture system as those in Experiment 1. We captured the participants’ head positions and postures in head-free and head-fixed conditions. In both conditions, the orientation of the intact region was synchronized with that of the head.

In the head-fixed condition, the participants performed the task on a chinrest. Although the head may have moved slightly during stimulus observation, we assumed that this would not affect the response because the intact region rotated synchronously with the head. To ensure that the head movements were small, we recorded the participants’ head positions and postures while they observed the stimuli. After the experiment, we examined how much the participants’ heads were fixed.

3.2.3 Stimuli

We used the same three scenes as those in Experiment 1. We filled the pixels in the modified region with black. At the border between the intact and modified regions, we blended the original and black images based on Equation 1. For Experiment 2, gorg and gmod in Equation 2 indicate the pixel values of the original and black images, respectively. We used the same weight transition functions as those in Experiment 1, namely, the step and sigmoid functions for the steep and gradual methods, respectively.

Following the method of constant stimuli, we varied the horizontal angle of the intact region at 11 levels (5° intervals), and the participants were shown the images at each level ten times. In the head-free condition, the intact region ranged from 225° to 275° and 233.8°–283.8° for stimuli generated using the steep and gradual blending methods, respectively. In the head-fixed condition, we used only the steep method, with the intact region ranging from 235° to 285°. Each participant observed 330 stimuli (3 scenes × 11 angular levels × 10 repetitions) in each condition.

3.2.4 Procedure

For the head-free condition, the procedure was identical to that in Experiment 1. As shown in Figure 5A, the participants sequentially observed two images, one of which had a modified region filled with black. The participants were instructed to naturally shift their gaze toward the bullseye target with head movements. They selected one of the two stimuli for which they felt the presence of a black mask in the modified region.

Figure 5B shows the procedure for the head-fixed condition, wherein the participants observed two images sequentially by shifting their gaze maximally. Each trial consisted of the following five steps. Before stimulus observation, i) the participants adjusted their head positions and postures on the chinrest in the same manner as in Experiment 1. ii) After the completion of head and posture adjustment, the color of the fixation dot changed from red to blue. Simultaneously, an arrow indicating the direction of maximal eye movement was presented adjacent to the dot. The direction of the arrow was randomly selected to be either left or right for each stimulus. After completing the head and posture adjustment, the participants initiated stimulus observation by pressing the “A” button on a gaming pad at their preferred timing. iii) Subsequently, a scene image was presented with or without the modified region while the blue dot and arrow stayed in the same position. The participants maintained their gaze on the blue dot. iv) After 500 m, a 1 kHz beep that lasted for 200 m signaled the start of eye movements. Simultaneously, the blue dot and arrow disappeared. The participants then shifted their gaze maximally in the direction indicated by an arrow. v) The scene image turned black 1,500 m after the beeping sound was emitted. The second stimulus was observed in the same manner as the first one. After the two observations, the participants selected a stimulus that they felt contained the modified region.

The 20 participants were evenly assigned with an equal female-to-male ratio to two groups: one containing stimuli with steep blending and the other with gradual blending. Those assigned to the steep group in Experiment 1 were assigned to the same group in Experiment 2, and vice versa. The participants assigned to the steep group performed the experiments under both head-free and head-fixed conditions. These experiments were conducted on two separate days to minimize participant fatigue.

3.2.5 Analysis

We estimated the threshold of the required size of FoV for each participant under each condition using Bayesian inference using the likelihood function given in Equation 7. Motion-to-photon latency was corrected in both head-free and head-fixed conditions using the method presented in Section 3.1.5. As in Experiment 1, we tested the difference in the thresholds between the border types (steep and gradual blending) using one-sided Welch’s t-test. In the same manner, we tested the difference in the thresholds between the head movement conditions (free and fixed) using one-sided Welch’s t-test.

4 Results

4.1 Experiment 1

We collected the responses yi for the stimulus images with the size of the latency-corrected intact regions xi and estimated the threshold μ with the maximization of the posterior probability described in Equation 6 for each participant. Based on the predetermined criteria, wherein μ outside the range of the corrected intact regions was regarded as unreliable, those obtained for some participants were excluded in the subsequent analysis. Specifically, two and eight out of ten participants were excluded in the LPF with a 0.5-cpd cutoff in the steep and gradual blending methods, respectively. In addition, one out of ten participants was excluded in the LPF with a 2-cpd cutoff in the steep blending method. The uncertainties in μ for the excluded participants were large; the median of the 95% Bayesian credible intervals estimated using the posterior distribution of μ for the excluded participants was approximately 100.0°, which was significantly larger than 18.2° for the included participants. The details of the thresholds μ and range of the latency-corrected intact regions for each participant are provided in Supplementary Table S1.

4.1.1 Required horizontal angles of intact region

Figure 8A shows the estimated μ for each participant to not notice the cutoff of the spatial frequency components in the modified region. The medians of μ for the steep and gradual blending methods were approximately 213° and 201°, respectively, for the 0.5-cpd cutoff LPF, and 122° and 116° for the 2-cpd cutoff LPF. We did not find any significant differences between the steep and gradual methods in both 0.5- and 2-cpd cutoff conditions, (t (2.6) = 1.66, p = 0.10, Hedges’ g = 0.87) and (t (15.1) = 0.05, p = 0.48, Hedges’ g = 0.02), respectively.

Figure 8
www.frontiersin.org

Figure 8. Horizontal angles of intact regions required for participants to not notice quality degradations. Each point indicates the threshold μ of each participant. (A) Relationship between the required sizes of the intact region and the cutoff frequencies of 0.5 and two cpd used to generate the blurred images for the modified region. Box plots represent the estimated thresholds for the participants to not notice the cutoff of the spatial frequency components. The horizontal line at approximately 152° indicates the median of the thresholds with a 2-cpd cutoff LPF from a previous study (Harasawa et al., 2023) in which the participants maximally moved their eyes. (B) The required size of the intact region when the modified region was filled with a black image. Box plots represent the estimated thresholds for the participants to not notice the narrowed FoV. The horizontal solid line at approximately 253° indicates the median of the thresholds of FoV with maximal eye movements from a previous study (Harasawa et al., 2021). The dashed line at 310° is the FoV calculated by linear summation of the horizontal FoV based on retina-centered coordinate system and the horizontally movable range of the eyes.

The medians of μ obtained in this study were approximately 30° lower than the median reported in a previous study with maximal eye movement (Harasawa et al., 2023). The horizontal line at approximately 150° in Figure 8A indicates the median of μ for the participants to not notice the cutoff of spatial frequency components above two cpd with their maximal eye movements in Harasawa et al. (2023). All the estimated μ in the 2-cpd cutoff condition were lower than the median of the thresholds in Harasawa et al. (2023), except for one excluded μ.

4.1.2 Head motion

The recorded data of head motion indicated that the participants started to move their heads after the appearance of the target. Figure 9A shows the yaw angle transitions for each participant’s head during the observation of stimuli with the modified region in Experiment 1. Each curve represents the regression of the yaw angles per participant, obtained by the geom_smooth function in R (Wickham et al., 2024.) with the generalized additive model (GAM) (Hastie and Tibshirani, 1986). The transition curves indicated that head rotation started approximately 0.5 s after the appearance of the bullseye target, which was approximately 1 s after the scene image appeared. This suggests that the participants were following the task procedures correctly.

Figure 9
www.frontiersin.org

Figure 9. Yaw angle transitions of each participant’s head during stimulus observation. (A) Transitions observed in Experiment 1. (B) Transitions observed in Experiment 2. The left and right panels in A and B compare the observations with respect to the blending methods (border types). The rising and falling curves correspond to head rotations in the right and left directions, respectively. The dashed lines represent the head rotations of the excluded participants in the analysis. The vertical gray lines indicate the time of appearance of the bullseye target. The horizontal gray lines indicate the yaw angles of the bullseye target: 60° for the right and −60° for the left.

The velocity of head rotation varied among the participants. We measured vmax, the maximum absolute velocity for yaw angle, in each stimulus observation. The median vmax for each participant ranged from approximately 61°/s–233°/s. The mean and SD of these velocities were approximately 111°/s and 37°/s, respectively.

Faster head movement tended to reduce the size of the intact region, depending on the blending method. The Pearson’s correlation coefficients between the medians of vmax and the estimated μ (the required intact region) were negative, as shown in Figure 10A. The correlation coefficient ρ in the 0.5 cpd-filtering condition was −0.64 for the steep blending method. In the two cpd-filtering condition, the coefficient values were −0.89 and −0.33 for the steep and gradual blending methods, respectively. The correlation in the two cpd-filtering condition with the steep blending method was significantly negative (p=0.001). We were unable to calculate the correlation in the 0.5 cpd-filtering condition with the gradual method due to the lack of μ samples.

Figure 10
www.frontiersin.org

Figure 10. Correlations between the velocity of head rotation and the required intact region. (A) Correlation observed in Experiment 1. (B) Correlation observed in Experiment 2. The shaded areas indicate the 95% confidence intervals.

4.2 Experiment 2

In the analysis, we excluded μ that fell outside the range of the latency-corrected intact regions xi. In the head-free condition, we excluded μ for two and three out of ten participants for the steep and gradual blending methods, respectively. In the head-fixed condition, we excluded μ for five participants. The uncertainties in μ for the excluded participants were large; the median of the 95% Bayesian credible intervals estimated using the posterior distribution of μ for the excluded participants was approximately 190.6°, which was significantly larger than 11.4° for the included participants. The details of the thresholds μ and range of the latency-corrected intact regions for each participant are provided in Supplementary Table S2.

4.2.1 Required horizontal angles of intact region

Figure 8B shows the estimated μ in the head-free and head-fixed conditions. Each dot plotted in Figure 8B represents μ for each participant and indicates the required horizontal angles at which they could not notice the modified region where pixels were filled in black. Under the head-free condition, the medians of μ for the steep and gradual blending methods were approximately 240° and 241°, respectively. The mean and SD were 240° (SD: 9°) for the steep blending method and 238° (SD: 11°) for the gradual blending method. Under the head-fixed condition, the median of μ was approximately 246°, with the mean of 247° and SD of 11°. We did not find any significant differences between the steep and gradual blending methods in the head-free condition (t (11.8) = 0.42, p = 0.34, Hedges’ g = 0.21), nor between the head-free and head-fixed conditions in the steep blending method (t (7.6) = 1.40, p = 0.10, Hedges’ g = 0.77).

In the head-free and head-fixed conditions, the medians of μ were approximately 13° and 7° lower, respectively, than the median reported in Harasawa et al. (2021), which investigated the FoV thresholds with maximal eye movements. The horizontal line in Figure 8B corresponds to the result of Harasawa et al. (2021). All the estimated μ, except for two participants, were lower than the median reported in Harasawa et al. (2021). We conducted one-sided Welch’s t-tests using thresholds μ of nine participants estimated in Harasawa et al. (2021), with a mean and SD of approximately 248° and 14°, respectively. The results showed no significant difference between the head-free condition of this study and head-fixed condition of Harasawa et al. (2021) (t (13.1) = 1.45, p = 0.09, Hedges’s g = 0.65). Additionally, we found no significant difference in μ under the head-fixed condition between this study and the previous study (Harasawa et al., 2021) (t (11.3) = 0.13, p = 0.45, Hedges’ g = 0.06). Although there were no significant differences, the larger effect size in the head-free condition compared to the head-fixed condition implies that the head-free condition tends to relax the size of FoV required for ideal HMDs more than the head-fixed condition.

The estimated μ was much smaller than its naive estimate using the sum of the previous findings on the size of the visual field without eye movement and the typical range of eye movements. The dashed line at 310° in Figure 8B represents the naive estimate of FoV, which is much higher than the medians of μ obtained in this study. The naive estimate was obtained by linear summation of the horizontal FoV (220°) based on the retina-centered coordinate system (Traquair, 1938; Lestak et al., 2021) and the eye’s horizontally movable range (90°) (Lee et al., 2019). The medians of μ obtained in this study were approximately 70° and 64° lower, respectively, than the naively estimated threshold.

4.2.2 Head motion

Figure 9B shows transitions of head rotations in the head-free condition in Experiment 2, with results similar to those obtained in Experiment 1 (Figure 9A). The head rotation started approximately 0.5 s after the appearance of the bullseye target, which was approximately 1 s after the scene image appeared.

In the head-free condition, the velocity of head rotation varied among the participants. The median of vmax for each participant ranged from approximately 55°/s–164°/s. The mean and SD of the medians of vmax were approximately 105°/s and 28°/s, respectively.

In the head-free condition, we found no correlations between the velocity of the head rotation and μ, although the Pearson’s correlation coefficients between the medians of vmax and thresholds μ were negative, as shown in Figure 10B. The correlation coefficients ρ were −0.47 and −0.17 for the steep and gradual methods, respectively. However, these correlations were not statistically significant (p=0.24 and p=0.69, respectively).

In the head-fixed condition, the head did not rotate or translate significantly and remained close to the center of the cylindrical display. Figure 11A shows the magnitude of the head rotation for each stimulus observation. The medians of rotations in the yaw, pitch, and roll axes were approximately 0.7°, 0.3°, and 0.3°, respectively. Figure 11B shows the extent of head translation during each stimulus observation. The medians of the extent of head translation were approximately 0.08, 0.04, and 0.04 cm along the three axes: left-right, down-up, and back-front, respectively. Figure 11C shows the medians of head positions along each axis for each stimulus observation. We found that the head positions were sufficiently close to the center such that 95% of the medians were within 1.3 cm of the center.

Figure 11
www.frontiersin.org

Figure 11. Amount of head movements relative to the display in the head-fixed condition. (A) Amount of rotation angles observed during stimulus viewing. (B) Amount of translation observed during stimulus viewing. These values indicate the difference between the maximum and minimum values along each axis in each stimulus observation. (C) Box plots of the head positions in each stimulus observation. The origin indicates the center of the cylindrical display.

5 Discussion

5.1 Interpretation and significance

We investigated the horizontal specifications of pixel resolution and FoV required for ideal HMDs under practical conditions, wherein the participants naturally shifted their gaze and observed natural scenes. We found that ideal HMDs do not have to provide spatial frequency components higher than 0.5 and two cpd in the periphery outside of approximately 210° and 120°, respectively. We also found that the horizontal FoV required for ideal HMDs was approximately 240°, which is much smaller than the naive estimate of 310° obtained by previous studies (Traquair, 1938; Lee et al., 2019; Lestak et al., 2021). The required sizes of the intact region estimated in this study tended to be lower than those estimated in the previous studies (Harasawa et al., 2021; 2023) in which the participants maximally shifted their gaze. Specifically, specifications lower than the previous estimates were sufficient for all participants except two, when applying the LPF at the 2-cpd cutoff and narrowing FoV. These findings suggest that the specifications required for ideal HMDs can be relaxed in practical conditions.

We observed no significant differences in the required size of the intact region between the head-fixed and head-free conditions, as shown in Figure 8B. This result was unexpected, as the participants should have observed the larger visual field with maximal gaze shift in the head-fixed condition than with moderated gaze shift in the head-free condition. The median of the required FoV in the head-fixed condition is 5° larger than that in the head-free condition. This difference was not statistically significant, possibly due to the small number of participants, resulting in a lack of statistical power. A post hoc analysis indicated that the test had a power of approximately 0.4.

We observed no significant differences in the required specifications between the steep and gradual blending methods in both Experiments 1 and 2. As shown in Figure 4B, the width of the transition from the original to the completely quality-degraded images was approximately 13.8° in the gradual blending method. This means that the definition of the intact region somewhat includes the area with gradual quality degradation. This transition width might be too large and affect μ in the gradual blending method. By narrowing the width of the transition, the area with quality degradations in the intact region becomes smaller. Thus, optimizing the transition width may reduce the size of the required intact region. Further investigation is necessary in this regard.

There were much smaller values of μ than those predicted by the results of previous studies. In the head-fixed condition, the μ values for five participants were even smaller than the presented size of the intact region, as highlighted in ‘i)’ in Figure 8B. This may have been due to low concentration levels among the participants while performing the experiments. During the practice session, we confirmed that each participant could detect quality degradation with the small intact regions that were used in the experiments. Therefore, the actual μ for each participant should have been larger than the smallest intact regions presented in the experiments. We suppose that task difficulty could have lowered the concentration levels among the participants, which resulted in reducing the size of the required intact region. Increasing the number of trials in which quality degradation is easy to detect could assist in maintaining the participant concentration levels.

A rapid head rotation might contribute to reducing the size of the required intact region. A faster head rotation induces smaller saccade amplitudes (Guitton and Volle, 1987), which probably reduces the size of the required intact region. As shown in Figure 10, the correlation coefficients between median of vmax and estimated μ were all negative. In particular, the correlation was significantly negative in the 2-cpd cutoff LPF with the steep blending method. In this study, we did not exclude data with rapid head rotation as the head rotation velocities in daily life vary among individuals.

5.2 Limitations and future work

In the condition of 0.5-cpd cutoff LPF with the gradual method, the estimated μ values for the seven participants were larger than the upper limit of the latency-corrected intact regions presented in the experiment, as indicated in ‘ii)’ in Figure 8A. We suppose that the intact regions presented in the experiment were too small, making the task excessively easy. A post hoc analysis showed that the accuracy of the responses from the seven excluded participants was approximately 96%. Revealing the threshold under this condition will be a part of future work.

In this study, we used only static images. We suppose that when observing dynamic images, the intact region required for not noticing the cutoff of spatial frequency components does not surpass that with static images. In dynamic images, the stimulus velocity on the retina could be greater than that in static images. Dynamic visual acuity (DVA), the ability to identify the details of a moving stimulus, becomes worse as the stimulus velocity on the retina increases (Brown, 1972; Antonio Aznar-Casanova et al., 2005; Lewis et al., 2011; Patrick et al., 2019). This DVA degradation might make the cutoff of spatial frequency components less noticeable. By contrast, a narrowed FoV might become more noticeable owing to the greater temporal luminance variation in the periphery in dynamic images. Therefore, an investigation using dynamic images is expected to provide a more conservative estimate of the required FoV.

In this study, we investigated the size of required intact region with cutoff frequencies of 0.5 and two cpd. Harasawa et al. (2023) showed that the required intact regions with 4- and 8-cpd LPFs were approximately 34° and 45° smaller, respectively, than that with 2-cpd LPF under the condition of maximal eye movements. It remains unclear whether differences between 2-cpd and higher cut-off frequencies during natural eye movements are comparable to those during maximal eye movements. If they are comparable, then the required intact regions during the natural eye movements could be predicted based on the results obtained under the condition of maximal eye movements, and vice versa. Future studies will include estimations involving higher cut-off frequencies in practical conditions.

It is known that there are sex-based differences in human visual system (Vanston and Strother, 2017). Investigating these differences in the specifications required for ideal HMDs is valuable for developing more personalized HMDs tailored to each user’s gender.

6 Conclusion

In this study, we estimated the horizontal specifications required for developing ideal HMDs under practical conditions wherein the participants moved their eyes in coordination with head rotations while observing natural scene images. The study revealed that ideal HMDs do not have to provide images with spatial frequencies above 0.5 and two cpd in the periphery horizontally outside of approximately 210° and 120°, respectively. In addition, we found that ideal HMDs need to provide a FoV of at least 240°. This is much smaller than a naively estimated FoV of 310° obtained by summing the FoVs in retina-centered coordinate system and the eye’s movable range. We believe that our findings will aid the design of ideal HMDs.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Ethics Committee of NHK Science and Technology Research Laboratories. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

YM: Data curation, Formal Analysis, Investigation, Methodology, Software, Visualization, Writing–original draft. MH: Conceptualization, Methodology, Supervision, Writing–review and editing. KH: Resources, Software, Writing–review and editing. YS: Methodology, Writing–review and editing. KK: Project administration, Writing–review and editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

Authors YM, MH, KH, YS, and KK were employed by Japan Broadcasting Corporation.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/frvir.2024.1485243/full#supplementary-material

References

Anstis, S. M. (1974). A chart demonstrating variations in acuity with retinal position. Vis. Res. 14, 589–592. doi:10.1016/0042-6989(74)90049-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Antonio Aznar-Casanova, J., Quevedo, L., and Sinnett, S. (2005). The effects of drift and displacement motion on Dynamic Visual Acuity. Psicológica 26, 105–119.

Google Scholar

Brown, B. (1972). Resolution thresholds for moving targets at the fovea and in the peripheral retina. Vis. Res. 12, 293–304. doi:10.1016/0042-6989(72)90119-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Campbell, F. W., and Robson, J. G. (1968). Application of Fourier analysis to the visibility of gratings. J. Physiol. 197, 551–566. doi:10.1113/jphysiol.1968.sp008574

PubMed Abstract | CrossRef Full Text | Google Scholar

Cuervo, E., Chintalapudi, K., and Kotaru, M. (2018). “Creating the perfect illusion: what will it take to create life-like virtual reality headsets?,” in HotMobile 2018 - proceedings of the 19th international workshop on mobile computing systems and applications (New York, NY, USA: Association for Computing Machinery), 7–12. doi:10.1145/3177102.3177115

CrossRef Full Text | Google Scholar

Freedman, E. G., and Sparks, D. L. (1997). Eye-head coordination during head-unrestrained gaze shifts in rhesus monkeys. J. Neurophysiol. 77, 2328–2348. doi:10.1152/jn.1997.77.5.2328

PubMed Abstract | CrossRef Full Text | Google Scholar

Guenter, B., Finch, M., Drucker, S., Tan, D., and Snyder, J. (2012). “Foveated 3D graphics,” in ACM transactions on graphics. doi:10.1145/2366145.2366183

CrossRef Full Text | Google Scholar

Guitton, D., and Volle, M. (1987). Gaze control in humans: eye-head coordination during orienting movements to targets within and beyond the oculomotor range. J. Neurophysiol. 58, 427–459. doi:10.1152/jn.1987.58.3.427

PubMed Abstract | CrossRef Full Text | Google Scholar

Hanif, S., Rowe, F. J., and O’connor, A. R. (2009). A comparative review of methods to record ocular rotations. Br. Ir. Orthopt. J. 6, 47. doi:10.22599/bioj.8

CrossRef Full Text | Google Scholar

Harasawa, M., Miyashita, Y., and Komine, K. (2021). Perimetry on head-centered coordinate system for requirements of head-mounted display. Proc. Int. Disp. Work. 28, 587–589. doi:10.36463/idw.2021.0587

CrossRef Full Text | Google Scholar

Harasawa, M., Miyashita, Y., and Komine, K. (2023). Required specification on spatial aspects for ideal HMD estimated on head-centered coordinate system. Proc. Int. Disp. Work. 30, 748–750. doi:10.36463/idw.2023.0748

CrossRef Full Text | Google Scholar

Hastie, T., and Tibshirani, R. (1986). Generalized additive models. Stat. Sci. 1. doi:10.1214/ss/1177013604

CrossRef Full Text | Google Scholar

Hoffman, M. D., and Gelman, A. (2014). The No-U-turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo. J. Mach. Learn. Res. 15, 1593–1623. Available at: http://jmlr.org/papers/v15/hoffman14a.html (Accessed July 2, 2024).

Google Scholar

Hussain, R., Chessa, M., and Solari, F. (2021). Mitigating cybersickness in virtual reality systems through foveated depth-of-field blur. Sensors 21, 4006. doi:10.3390/s21124006

PubMed Abstract | CrossRef Full Text | Google Scholar

Jerald, J., and Whitton, M. (2009). “Relating scene-motion thresholds to latency thresholds for head-mounted displays,” in 2009 IEEE virtual reality conference, Lafayette, LA, USA, 14-18 March 2009 (IEEE), 211–218. doi:10.1109/VR.2009.4811025

CrossRef Full Text | Google Scholar

Lee, S., Pattichis, M. S., and Bovik, A. C. (2002). Foveated video quality assessment. IEEE Trans. Multimed. 4, 129–132. doi:10.1109/6046.985561

CrossRef Full Text | Google Scholar

Lee, W. J., Kim, J. H., Shin, Y. U., Hwang, S., and Lim, H. W. (2019). Differences in eye movement range based on age and gaze direction. Eye 33, 1145–1151. doi:10.1038/s41433-019-0376-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Legge, G. E., and Foley, J. M. (1980). Contrast masking in human vision. J. Opt. Soc. Am. 70, 1458–1471. doi:10.1364/JOSA.70.001458

PubMed Abstract | CrossRef Full Text | Google Scholar

Leitao, P., and Nanda, R. S. (2000). Relationship of natural head position to craniofacial morphology. Am. J. Orthod. Dentofac. Orthop. 117, 406–417. doi:10.1016/S0889-5406(00)70160-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Lestak, J., Lestak, T., Fus, M., and Klimesova, I. (2021). Temporal visual field border. Clin. Ophthalmol. 15, 3241–3246. doi:10.2147/OPTH.S321110

PubMed Abstract | CrossRef Full Text | Google Scholar

Levitt, H. (1971). Transformed up-down methods in psychoacoustics. J. Acoust. Soc. Am. 49, 467–477. doi:10.1121/1.1912375

PubMed Abstract | CrossRef Full Text | Google Scholar

Lewis, P., Rosén, R., Unsbo, P., and Gustafsson, J. (2011). Resolution of static and dynamic stimuli in the peripheral visual field. Vis. Res. 51, 1829–1834. doi:10.1016/j.visres.2011.06.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Mantiuk, R., Kim, K. J., Rempel, A. G., and Heidrich, W. (2011). “HDR-VDP-2: a calibrated visual metric for visibility and quality predictions in all luminance conditions,” in ACM SIGGRAPH 2011 papers (New York, NY, USA: ACM), 1–14. doi:10.1145/1964921.1964935

CrossRef Full Text | Google Scholar

Mantiuk, R. K., Denes, G., Chapiro, A., Kaplanyan, A., Rufo, G., Bachy, R., et al. (2021). FovVideoVDP: a visible difference predictor for wide field-of-view video. ACM Trans. Graph 40, 1–19. doi:10.1145/3450626.3459831

CrossRef Full Text | Google Scholar

Miyashita, Y., Harasawa, M., Hara, K., Sawahata, Y., and Komine, K. (2023). “Estimation of required horizontal FoV for ideal HMD utilizing vignetting under practical range of eye displacement,” in Proceedings - 2023 IEEE conference on virtual reality and 3D user interfaces abstracts and workshops, VRW 2023. Shanghai, China, 25-29 March 2023 (IEEE), 691–692. doi:10.1109/VRW58643.2023.00188

CrossRef Full Text | Google Scholar

Newman, R. L., Greeley, K. W., and Turpin, T. S. (1998). in HMDs: a standard and a design guide. Editors R. J. Lewandowski, L. A. Haworth, and H. J. Girolamo, 103–109. doi:10.1117/12.317423

CrossRef Full Text | Google Scholar

Patrick, J. A., Roach, N. W., and McGraw, P. V. (2019). Temporal modulation improves dynamic peripheral acuity. J. Vis. 19, 12. doi:10.1167/19.13.12

PubMed Abstract | CrossRef Full Text | Google Scholar

Rakkolainen, I., Raisamo, R., Turk, M., and Höllerer, T. (2017). “Field-of-view extension for VR viewers,” in Proceedings of the 21st international academic mindtrek conference, AcademicMindtrek 2017 (Association for Computing Machinery, Inc). doi:10.1145/3131085.3131088

CrossRef Full Text | Google Scholar

Rakkolainen, I., Turk, M., and Höllerer, T. (2016). “A compact, wide-FOV optical design for head-mounted displays,” in Proceedings of the ACM symposium on virtual reality software and Technology (VRST: Association for Computing Machinery), 293–294. doi:10.1145/2993369.2996322

CrossRef Full Text | Google Scholar

Ratcliff, J., Supikov, A., Alfaro, S., and Azuma, R. (2020). ThinVR: heterogeneous microlens arrays for compact, 180 degree FOV VR near-eye displays. IEEE Trans. Vis. Comput. Graph 26, 1981–1990. doi:10.1109/TVCG.2020.2973064

PubMed Abstract | CrossRef Full Text | Google Scholar

Rovamo, J., and Virsu, V. (1979). An estimation and application of the human cortical magnification factor. Exp. Brain Res. 37, 495–510. doi:10.1007/BF00236819

PubMed Abstract | CrossRef Full Text | Google Scholar

Sidenmark, L., and Gellersen, H. (2019). Eye, head and torso coordination during gaze shifts in virtual reality. ACM Trans. Computer-Human Interact. 27, 1–40. doi:10.1145/3361218

CrossRef Full Text | Google Scholar

Singh, R., Huzaifa, M., Liu, J., Patney, A., Sharif, H., Zhao, Y., et al. (2023). “Power, performance, and image quality tradeoffs in foveated rendering,” in 2023 IEEE conference virtual reality and 3D user interfaces (VR), Shanghai, China, 25-29 March 2023 (IEEE), 205–214. doi:10.1109/VR55154.2023.00036

CrossRef Full Text | Google Scholar

Wickham, H., Chang, W., Henry, L., Pedersen, T., Takahashi, K., Wilke, C., et al. (2024). Smoothed conditional means. Available at: https://ggplot2.tidyverse.org/reference/geom_smooth.html. (Accessed June 10, 2024)

Google Scholar

Stan Development Team (2023). RStan: the R interface to Stan. R. package version 2.26.1. Available at: https://mc-stan.org/.

Google Scholar

Sutherland, I. (1968). A head-mounted three dimensional display. Proc. Fall Jt. Comput. Conf. 33, 757–764. doi:10.1145/1476589.1476686

CrossRef Full Text | Google Scholar

Swafford, N. T., Iglesias-Guitian, J. A., Cosker, D., Koniaris, C., Mitchell, K., and Moon, B. (2016). “User, metric, and computational evaluation of foveated rendering methods,” in Proceedings of the ACM symposium on applied perception, SAP 2016 (Association for Computing Machinery, Inc), 7–14. doi:10.1145/2931002.2931011

CrossRef Full Text | Google Scholar

Tan, G., Lee, Y.-H., Zhan, T., Yang, J., Liu, S., Zhao, D., et al. (2018). Foveated imaging for near-eye displays. Opt. Express 26, 25076. doi:10.1364/oe.26.025076

PubMed Abstract | CrossRef Full Text | Google Scholar

Traquair, H. M. (1938). An introduction to clinical perimetry.

Google Scholar

van der Schaaf, A., and van Hateren, J. H. (1996). Modelling the power spectra of natural images: statistics and information. Vis. Res. 36, 2759–2770. doi:10.1016/0042-6989(96)00002-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Vanston, J. E., and Strother, L. (2017). Sex differences in the human visual system. J. Neurosci. Res. 95, 617–625. doi:10.1002/jnr.23895

PubMed Abstract | CrossRef Full Text | Google Scholar

Virsu, V., and Rovamo, J. (1979). Visual resolution, contrast sensitivity, and the cortical magnification factor. Exp. Brain Res. 37, 475–494. doi:10.1007/bf00236818

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Y., and Healy, J. J. (2023). Image quality assessment for gibbs ringing reduction. Algorithms 16, 96. doi:10.3390/a16020096

CrossRef Full Text | Google Scholar

Wang, Z., Bovik, A. C., Lu, L., and Kouloheris, J. L. (2001). in Foveated wavelet image quality index. Editor A. G. Tescher, 42–52. doi:10.1117/12.449797

CrossRef Full Text | Google Scholar

Xiao, R., and Benko, H. (2016). “Augmenting the field-of-view of head-mounted displays with sparse peripheral displays,” in Conference on human factors in computing systems - proceedings (Association for Computing Machinery), 1221–1232. doi:10.1145/2858036.2858212

CrossRef Full Text | Google Scholar

Keywords: head-mounted display, spatial frequency, field of view, eye-head coordination, perimetry, foveated rendering, specification, human visual system

Citation: Miyashita Y, Harasawa M, Hara K, Sawahata Y and Komine K (2024) Estimation of horizontal spatial specifications for ideal head-mounted displays in practical conditions. Front. Virtual Real. 5:1485243. doi: 10.3389/frvir.2024.1485243

Received: 23 August 2024; Accepted: 23 October 2024;
Published: 05 November 2024.

Edited by:

Xueni Pan, Goldsmiths University of London, United Kingdom

Reviewed by:

David Lindlbauer, Carnegie Mellon University, United States
Rohith Venkatakrishnan, University of Florida, United States

Copyright © 2024 Miyashita, Harasawa, Hara, Sawahata and Komine. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yamato Miyashita, miyashita.y-fc@nhk.or.jp

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.