- 1Department of Biology, University of Konstanz, Konstanz, Germany
- 2Centre for the Advanced Study of Collective Behaviour, University of Konstanz, Konstanz, Germany
- 3Zukunftskolleg, Universität Konstanz, Konstanz, Germany
A large proportion of animal species enjoy the benefits of being active at night, and have evolved the corresponding optical and neural adaptations to cope with the challenges of low light intensities. However, over the past century electric lighting has introduced direct and indirect light pollution into the full range of terrestrial habitats, changing nocturnal animals’ visual worlds dramatically. To understand how these changes affect nocturnal behavior, we here propose an animal-centered analysis method based on environmental imaging. This approach incorporates the sensitivity and acuity limits of individual species, arriving at predictions of photon catch relative to noise thresholds, contrast distributions, and the orientation cues nocturnal species can extract from visual scenes. This analysis relies on just a limited number of visual system parameters known for each species. By accounting for light-adaptation in our analysis, we are able to make more realistic predictions of the information animals can extract from nocturnal visual scenes under different levels of light pollution. With this analysis method, we aim to provide context for the interpretation of behavioral findings, and to allow researchers to generate specific hypotheses for the behavior of nocturnal animals in observed light-polluted scenes.
Introduction
Many of the world’s animals, from tiny insects to large mammals, are active during the night. These species face a dramatic change in their sensory environment, as the available light varies more than one million-fold between day and night (Johnsen et al., 2006). This critically limits the information available to the sense of vision, which is fundamental to locomotion control, orientation and navigation, and also heavily relied upon for foraging, predator avoidance, mate detection and communication. Yet, colonizing this challenging diel niche comes with great benefits in avoiding both competitors and predators (Wcislo et al., 2004), thus making evolutionary adaptations to the low light conditions worthwhile. Over millions of years, intricate nocturnal ecosystems have thus evolved, with unique predator-prey interactions (Park, 1940), specialized pollinators (Macgregor and Scott-Brown, 2020; Walton et al., 2020) and even further subdivision of the night into temporally distinct niches (Greiner et al., 2007; Narendra et al., 2011; Stöckl A. L. et al., 2017; Tocco et al., 2019). Counterintuitively, these fragile ecological communities, defined by the scarcity of light, are now threatened by what might be considered a benefit for other species that rely on vision: increasing levels of light at night (Longcore and Rich, 2004; Gaston et al., 2012; Falchi et al., 2016). This artificial light at night (ALAN) can occur either as isolated concentrations of light (i.e., street lights), or an increase in the brightness of the entire sky (i.e., through sky glow) (Longcore and Rich, 2004). However, such ALAN is extremely challenging for visual systems that are so well-adapted to using even the last available photon, as they suddenly encounter an overabundance of light in nocturnal visual scenes.
To date, the dominant paradigm in the study of light-pollution’s effects on nocturnal animals has been to observe individual behavioral and physiological effects in an attempt to infer larger-scale network effects, in which light introduced to the environment can be demonstrated to alter the ecosystem as a whole (Longcore and Rich, 2004; Gaston et al., 2012; Grubisic et al., 2018; Giavi et al., 2021; Gaston and de Miguel, 2022). By comparison, the visual mechanisms themselves have received relatively little attention. This provides those working on mitigating ALAN’s damaging effects with evidence to bolster their case, but limited tools with which to explain these effects or, better yet, to make policy proposals–a “what,” but not a “how” or a “why.” In this article, we make the case for combining the ecological approach to the study of light pollution, with the paradigms, methods and tools developed in the fields of visual ecology and dim-light vision outlined below. We first provide an overview of nocturnal visual environments, and the typical forms of light pollution observed, and then go on to describe the visual systems of nocturnal animals and how they are adapted to low light conditions. Based on this, we present estimates of how light pollution alters the visual percepts of different nocturnal animals, based on realistic ocular modeling of light information from natural terrestrial visual scenes.
Light and visual habitats
The vast majority of photons animals observe are not produced here on Earth. For millions of years the visual worlds of animals were lit (with the notable exception of deep-sea environments) by light from the sun; direct, scattered by the atmosphere or reflected by the moon; or from the stars. In the terrestrial world, any earthbound light sources are generally swamped by extra-terrestrial light sources, with exceptions such as lightning flashes or volcanic eruptions being either fleeting, ephemeral, or both. So it is that the visual systems of most animal species evolved to suit the available sunlight (Figures 1C,D), moonlight (Figure 1B), and starlight (Figure 1A), and their typical spectra, intensities, and circadian schedule.
Figure 1. Light intensity distributions in visual scenes across times of day. (A–D) The same visual scene near Vryburg (26°23′55.82″ S, 24°19′37.380″E) in South Africa, photographed Southward under different celestial illumination conditions as indicated by the icons (starlight on a moonless night, full moon light with moon above 60° elevation, sunset, noon). (E) The relative proportion of pixel intensities and the (F) average pixel intensies across scene elevation. Panel (E) shows the average brightness histograms for starlight, moonlight, sunset/runrise, and sun at noon of a number of individual scenes with elevation intensity profiles depicted in gray in panel (F). The profiles of the example scenes in panels (A–D) are highlighted in black. The dashed lines indicate average light intensities under typical celestial conditions (max. starlight, max. moonlight, end of twilight, sunset/sunrise and max. sunlight). The relative scaling of these were taken from Warrant (2008), while the scale was anchored in absolute terms to the intensity for sunlight from Nilsson and Smolka (2021). The visual thresholds of the three indicated model species [mouse Mus musculus (Umino et al., 2008), hawkmoth Deilephila elenor (Stöckl et al., 2016a), fruitfly Drosophila melanogaster (Currea et al., 2022)] were based on behavioral and physiological measurements of the response thresholds to moving contrast in optmotor-type tasks, both for comparability and because of the high ecological relevance of these tasks for animal behavior.
While natural scenes range in brightness across more than six log units (Figure 1E; Johnsen et al., 2006), the range within an individual scene is typically limited to 2 log units (Figure 1F; Nilsson and Smolka, 2021). Around sunset, ambient illumination at visible wavelengths (400–700 nm) is around 1018 photons m−2s−1, dropping to 1017 photons m−2s−1 as the sun reaches the horizon (Johnsen et al., 2006). After sunset, a moonlit night may be as bright as 1015 photons m−2s−1, while a starlit night is around 1013 photons m−2s−1 (Johnsen et al., 2006). When clouds or vegetation then block moonlight and starlight from view, ambient illumination can reach a low of 1012 photons m−2s−1 (Foster et al., 2021). Celestial cycles therefore provide a benchmark for total light available, but these changes need not be uniform across all regions within a scene. Above the horizon, the brightness distribution is dominated by visible celestial bodies (sun, moon, and stars) and blue skylight (solar or lunar) (Nilsson et al., 2022) as well as the dim background of airglow (produced by spontaneous emissions in the ozone layer) visible only on dark nights (Johnsen, 2012). Below the horizon, the observer views mainly reflected light, diffuse reflections from matte objects and specular reflections from shiny ones. On a clear day, the scene may be as bright below the horizon as above it (compare Figures 1B,D and respective profiles), thanks to direct sunlight and moonlight reflected from earthbound surfaces (Nilsson and Smolka, 2021; Nilsson et al., 2022). When the sun and moon are below the horizon, or clouds cover the sky, skylight, starlight, and cloudlight produce diffuse illumination that typically results in a much dimmer scene below the horizon (Figures 1A,C and respective profiles). Subtle changes in spectrum across different solar elevations and between scenes illuminated by the sun, moon and stars can also contribute to changes in diffuse reflections within a scene (Johnsen et al., 2006). It is the detection of these contrasts, between the sky and horizon, objects and their backgrounds, and celestial bodies and empty sky, to which animal vision is adapted. Achieving this, however, requires vastly different functionality within a visual system under brighter and dimmer conditions.
The industrial age threw these reliable features of day and night into disarray. Electric lighting introduced an unparalleled source of earthbound light into the environment, both directly, when these lights are in the animal’s field of view (Figures 2A,C), and indirectly, when this stray light projects up into the atmosphere and is returned to earth in the form of skyglow (Figures 2B,D). This ALAN dramatically alters the light environment, raising the ambient light intensity by 1–4 log units (Foster et al., 2021), obscuring the stars from view for human and animal observers alike, and introducing luminous objects into the light field that can outshine the full moon (Figures 2E,F). Increasing awareness of the impacts of human activity on the environment have led to a great interest in the impact of excess light on the structures of ecosystems (Bennie et al., 2015; Grubisic et al., 2018), and number of specific behaviors, such as flight to light (Baker and Sadovy, 1978; Doren et al., 2017; Wakefield et al., 2017), shore migration (Torres et al., 2020), and predation by insectivores on moths attracted to streetlights (Wakefield et al., 2015).
Figure 2. Nocturnal visual scenes with light pollution. (A–D) Nocturnal visual scenes on moonless nights with different types of light pollution. (A) Starry sky in rural Limpopo (24°43′46″S, 27°57′06″E), South Africa with sky glow visible at the horizon. (B) Nocturnal scene in Sodankylä, Finland, showing a weak aurora borealis in the sky, and direct light pollution relfected by the snowy ground below the horizon. (C) Sky at the University of the Witwatersrand, South Africa, with direct light pollution from surrounding buildings, and (D) sky in Östra Odarslöv, Sweden, showing clouds lit up by sky glow. Lines in panels (C,D) indicate 60 and 30° elevation. (E) The relative proportion of pixel intensities [colors correspond to those of the star icons for scenes depicted in panels (A–D), with the average intensity profiles for starlight and moonlight from Figure 1E in gray] and the (F) average pixel intensies across scene elevation [the elevation intensity profiles from Figure 1F in gray, and the elevation profiles of the scenes in panels (A,B) in color].
Here we introduce methods that help us understand what information animals can extract from a visual scene when their eyes are adapted to the intensity of ambient illumination.
Nocturnal vision
Even though the signal amplitude available to the visual system at night is only a minute fraction of that during the day (Figures 1E,F), many animals possess exquisite night vision. The elephant hawkmoth Deilephila elpenor, for example, can distinguish flower colors even under starlight intensities (Kelber, 2002), the tropical sweat bee Megalopta genalis skillfully approaches and securely lands on its nest stick in the dead of night in the rainforest (Warrant et al., 2004), and the night-migrating indigo buntings are guided by patterns of stars on their journey from North- to Central-America (Emlen, 1967a,b).
These are representative of typical nocturnal animals that invest in specializations of their eyes and nervous systems to make the most of the sparse visual information available. But why evolve such a system, rather than relying on other senses for the same tasks? The answer may be that it is not straightforward to replace the unrivaled depth of information the visual sense provides (Warrant and Johnsen, 2013; Warrant et al., 2020): high spatial- and high-temporal resolution information in several separate wavelength channels, which reaches from distances of only millimeters to many light-years (when viewing the starry sky) (Foster et al., 2018). While some of the information vision provides can be substituted by other senses at night, for example by olfaction for food localization via the odors the food source emits (Healy and Guilford, 1990; Stöckl et al., 2016c), visual information remains crucial for control tasks in relation to remote objects. Such objects, including predators or prey that sparsely emit sound or odors, are almost impossible to localize without vision–except with active sensing (Nelson and MacIver, 2006). And even species with active sensing alternatives, such as bats, use vision for obstacle control (Jones and Moss, 2021). One behavior in particular that cannot be executed with other passive senses, is flight (Davies and Green, 1994). Flying animals need to obtain rapid and remote information about their three-dimensional environment, to detect obstacles and changes in their own body position (Srinivasan et al., 1999; Bhagavatula et al., 2011). Thus, for flying animals in particular, but also for fast moving walkers, negotiating their environment on the search for food, mates or to avoid predators, vision is essential. Thus, many animals exploit even the sparsest light sources to control a range of behavioral tasks, using visual specializations that let them detect intensities down to single photons (Reingruber et al., 2013; Honkanen et al., 2014; Stöckl et al., 2016a; Tinsley et al., 2016; Stöckl A. L. et al., 2017).
A separate reason to rely on vision at night is not despite but because of the low availability of visual signals: glow worms and fireflies generate their own light signals for communication, and thus benefit from the low background levels of light at night, to produce singular and far reaching communication signals. Their visual challenge is similar to that of other nocturnal insects, in particular since the animals need to safely navigate toward their visual beacons (Kahlke and Umbers, 2016).
What restricts vision at night?
To understand the specializations for vision at night, a closer look at the challenges generated by the low intensities is required. The most fundamental limit of any light detection arises from the physical nature of light. Light particles, or photons, interact in a quantal fashion with the photopigments in the receptors that detect light in animal eyes (Yeandle, 1958; Lillywhite, 1977), or camera sensors (Yadid-Pecht et al., 1997; Tian et al., 2001). However, photons reach any possible surface in a stochastic manner (Rose, 1942; De Vries, 1943), so that the number of photons that arrive within a given area fluctuates over time. This uncertainty in photon arrival–and consequently absorption–is termed photon “shot noise” (Schottky, 1918). It sets the absolute noise limit for any visual reconstruction. Irrespective of how sensitive a light detector is, if the photon shot noise is too high, the original signal received by the detectors cannot be reliably reconstructed. The variance in photon arrival is governed by Poisson statistics (Rose, 1942; De Vries, 1943), so that the signal-to-noise ratio (SNR), the “detection criterion” required to reliably infer the original image, equates to the square root of the signal (N) (Snyder, 1979; Land, 1981):
Thus, the SNR is much smaller for lower photon catches–as measured at night–than for the high ones during the day. It is important to note that the SNR sets not only the limit for the absolute visual threshold, or absolute visual sensitivity, but also the limit for an animals’ contrast sensitivity: the ability to discriminate between different light levels. Contrast sensitivity forms the basis of spatial vision, for which the inputs of neighboring visual units, representing separate “pixels” of the image, need to be discriminated (Snyder, 1977; Snyder et al., 1977), as well as temporal vision, where the signal at consecutive time points needs to be resolved (Hornstein et al., 1999; Donner, 2021). As light levels decrease, the SNR decreases at higher spatial (Stöckl A. et al., 2017) and temporal frequencies, thus reducing the maximum achievable resolution of the visual system. Thus, the SNR at the photoreceptors and subsequent visual processing sets the limit for absolute visual detection and contrast vision, which is crucial for most visual behaviors.
However, this photon-based description of the SNR applies to a perfect sensor–with no loss in photon signal upon arrival at the eye, and no further sources of noise. Eyes are far from perfect sensors, though. Only a fraction of the photons arriving at the eye are absorbed by photopigments in the receptors (Warrant and Nilsson, 1998). Moreover, the absolute detection and accuracy of the photoreceptors is limited by “false positive” activations of the biochemical phototransduction cascade by thermal energy. The resulting electrical responses are indiscriminable from real photon responses, thus constituting the physiological limit for visual detection (Barlow, 1956; Aho et al., 1988, 1993). This so-called “dark variance” (σ2D) is relatively low in insects [ca. 10 false alarms per hour at 25°C in locusts (Lillywhite and Laughlin, 1979)], but is distinctly higher in vertebrates [ca. 360 false alarms per hour at 20°C in toads (Baylor et al., 1980)].
In addition, the photoreceptors’ responses to single photons vary in amplitude, latency and shape because of variations in the transduction cascade (Lillywhite and Laughlin, 1979; Howard et al., 1984). In dim light, this so-called “transducer noise” (σ2T) has an equal contribution to the total noise variance as photon shot noise, and at higher light intensities by far exceeds it in insects (Lillywhite and Laughlin, 1979; Laughlin and Lillywhite, 1982; Howard and Snyder, 1983). Thus, noise sources intrinsic to the sensory and nervous system decrease the sparse signal reliability in very dim light further, giving an estimate of the total noise as (Snyder et al., 1977; Cohn, 1983; Warrant, 1999):
Taken together, at only a few photons per sampling time, the variance of the visual signal is much higher than at higher average photon catches (compare Figures 3E,F). This high uncertainty in the available signal challenges nocturnal vision and requires specializations to improve the low signal to noise ratios in dim light.
Figure 3. The three most common eye designs in animals. (A) In a camera eye, light is focused by the cornea and lens on to the photoreceptors in the retina. This eye design is found in all vertebrates and is also common in the main eyes of molluscs and arachnids. A, f, d, l correspond to Eq. 3. (B) Apposition compound eyes are made of repeating units, the ommatidia, which are comprised of a facet lens, crystalline cone and the a group of photoreceptors, the retinula. The retinula cells of all ommatidia form the retina. Light is focused by the cornea and crystalline cone of each ommatidium onto the underlying rhabdom, the photosensitive tissue of the retinula cells. In a classical apposition eye, each ommatidium thus constitutes a sampling unit that views a “pixel” of the image. This eye design is commonly found in diurnal insects. (C) In superposition compound eyes, light is focused by the corneas and crystalline cones of a large number of ommatidia across the clear zone (cz) onto one rhabdom. Therefore, this eye design is typical for nocturnal insects, which require improvements in relative photon catch. (D) Nocturnal scene at moonlight recorded by the camera. Photon count of this scene simulated for (E) an apposition, and (F) a superposition compound eye [photon count is similar to the camera eye, compare panels (G,I)]. All eye types were modeled with photoreceptors of the same width and length, the same spatial resolution (for both the visual angle covered by sampling units, and for the photoreceptor acceptance angle), the same effective apertures for camera and superposition eye, as well as the same facet aperture in both compound eyes (see Table 1 for parameters). The noise levels include photon shot and transducer noise of the same magnitude. (G–I) Pixels were resampled to represent the spatial sampling units of a perfectly spherical eye. (J–L) The same scene viewed through the eye of a mouse, fruit fly, and elephant hawkmoth (see Table 1 for parameters).
Anatomical modifications to increase visual sensitivity
To use vision under such adverse conditions, animals need to overcome the challenges that high noise and low signal levels pose to obtaining reliable information for the behavioral tasks at hand–or wing. Improving the SNR can be achieved through directly increasing the number of photons collected by the eye. To understand how eyes can increase sensitivity, it is worth considering the optical sensitivity, a measure for the light capture ability of a visual system. The higher it is, the more light is captured. For extended scenes, this can be expressed as (Land, 1981; Warrant and Nilsson, 1998; Warrant, 1999):
In this equation, A is the diameter of the aperture, d is the photoreceptor diameter, f is the focal length of the eye, and kl/(2.3 + kl) is the fraction of light absorbed, where k is the absorption coefficient of the photoreceptor and l its length. As Eq. 3 shows, the number of photons each visual unit detects at a given integration time can be maximized by modifications that increase the aperture diameter (A) or decrease the focal length (f), and that increase the photoreceptors’ diameter (d) or length (l).
Increasing the cross-section area d2 of the photoreceptor and its length l (or more precisely the volume of tissue presenting photopigment within the receptor) increases the potential photon catch. This is especially important, because the absorption efficiency k of every single photopigment is relatively low [ranging between ca. 0.005–0.01 μm–1 in arthropods and 0.025–0.05 μm–1 in vertebrates (Warrant and Nilsson, 1998)]. Efficient absorption requires stacking membranes with photopigments. The fraction of absorbed white light γ is therefore directly proportional to the length of the photoreceptors l: γ∝kl/(2.3 + kl) (Warrant and Nilsson, 1998).
In addition to optimizing the probability of photon absorption, nocturnal eye designs are optimized to focus a wider cone of light onto the photoreceptors than typical diurnal eyes (Warrant and McIntyre, 1991). This is achieved via a large aperture combined with a shorter focal length. This relationship between aperture and focal length is expressed in an eye’s F-number: F = f/A. In eyes and photographic cameras, low F-numbers denote high sensitivity and vice versa.
Camera eyes
Nocturnal animals with camera type eyes (Figure 3A), as possessed by all eyed vertebrates, generally have large lenses with wide pupils (the effective aperture of the eye) and shorter focal lengths than their diurnal relatives. For comparison, a dark-adapted human eye has an F-number of 2.1 (Westheimer, 1970), while the F-numbers of nocturnal primates [Lesser Bushbaby: 0.71 (Kirk, 2006)] or nocturnal birds [barn owl: 0.83 (Pettigrew, 1983)] are distinctly lower. Invertebrates also follow this trend: the nocturnal spider Dinopis subrufus achieves its remarkable sensitivity with the largest lens of any terrestrial arthropod (Blest and Land, 1977), which together with a short focal length results in the very low F-number of 0.6.
Apposition compound eyes
Compound eyes, the most common eye type among insects and crustaceans, are composed of many “little eyes” or ommatidia, which constitute the visual units of the eye. In apposition compound eyes (Figure 3B), each ommatidial facet lens focuses light onto the photoreceptors comprising a visual unit, thus acting as the effective aperture of the eye. Because of the small size of individual facets, which constitute the eyes’ aperture, apposition compound eyes are limited in sensitivity (compare the photon distributions in Figures 3G,H) and are mostly found in diurnal arthropods. There are a few notable exceptions including nocturnal mosquitoes (Land, 1997), the tropical halictid bee M. genalis (Greiner et al., 2004b; Warrant et al., 2004), and nocturnal ants (Menzi, 1987; Moser et al., 2004; Narendra et al., 2011). As in camera eyes, these nocturnal apposition eyes generally have larger apertures, in their case facets, than their diurnal relatives, but even these modifications do not increase sensitivity to the level of nocturnal camera eyes [i.e., sensitivity of 2.69 in the nocturnal sweat bee (Greiner et al., 2004b)]. Moreover, the increase in sensitivity through larger apertures is achieved at the expense of spatial resolution, since an eye with larger facets can “fit” fewer of them than the same eye with narrower facets, and thus resolve fewer “pixels.” The only way to escape this trade-off is to increase overall eye size to retain a comparable spatial resolution with wider facet apertures (Jander and Jander, 2002; Greiner et al., 2004b).
Superposition compound eyes
An optical solution to the trade-off between sensitivity and spatial acuity is provided by the most prominent eye type of nocturnal insects: superposition compound eyes (Figure 3C). In this eye type, the pigment that separates the ommatidia in apposition compound eyes can withdraw, which leaves a wide clear zone between the crystalline cones underlying the facet lenses and the retina. The crystalline cones focus light from one point in space through multiple facets onto a single rhabdom (Exner, 1891; Nilsson, 1989). Thus, the effective aperture of this type of eye comprises all facets with a shared optical axis for this point in space. We can therefore extend Eq. 3 by nf, the number of facets contributing to the superposition aperture, and hence obtain the overall aperture area of the superposition eye as π/4D2nf where D is the facet diameter.
This increased aperture greatly increases the optical sensitivity of the eye, as noticeable by the low F-numbers of the nocturnal elephant hawkmoth at 0.72 (Warrant, 1999) or the dung beetle Onitis aygulus at 0.60 (McIntyre and Caveney, 1998). Thus, in superposition compound eyes, as in camera eyes, sensitivity can be increased by increasing the aperture without affecting the spatial sampling base (nominal spatial resolution) of the eye, resulting in comparable visual sensitivities between these types of eyes, given comparable eye features (compare Figures 3G,I). Nevertheless, due to the intricate constraints on their optics, superposition compound eyes are often limited by spherical aberration, which results in a blurring of the image focussed on the retina and thus reduces spatial acuity below the theoretical maximum (Caveney and McIntyre, 1981; McIntyre and Caveney, 1985; Warrant and McIntyre, 1990). This can be seen by comparing Figure 3I, which shows a visual scene through a model superposition eye with 1° spatial resolution, and Figure 3L, which shows this scene through the eye of the elephant hawkmoth D. elpenor. This hawkmoth possesses an anatomical sampling base (angular separation of ommatidia or interommatidial angle △φ) of 1.12°, but the effective spatial resolution of the photoreceptors (photoreceptor acceptance angle △ρ) is only 4°, resulting in considerable loss of resolution compared to Figure 3I.
Modeling photon catches and spatial resolution
As discussed in the previous paragraph, the combination of anatomical eye parameters, including aperture, focal length, photoreceptor diameter and length determines the visual sensitivity, while also setting the limits for spatial resolution.
To calculate the photon catch each sampling channel of an eye receives from a natural visual scene, the visual sensitivity S (Eq. 3) is multiplied with the measured radiance in a given time interval △t, generally the channels’ integration time, as well as scaled by the quantum capture efficiency of the transduction cascade κ, and the fraction of incident light transmitted by the optics τ of the eye (Warrant, 1999):
The term (πd2)/(4f2) can also be replaced by the solid angular subtense of the sampling channel’s (Gaussian) receptive field 1.13 △ρ2. △ρ is the half-width (full-width at half maximum) of the channel’s receptive field. It is often assumed that the sampling base of the eye is a single photoreceptor–as this is the smallest sensory unit that makes up the retinal mosaic. However, functionally, the sampling base of the visual percept does not depend on the resolution of the sensory sampling, but on the number of separate information channels resulting from this mosaic. Take the case of a classic apposition or superposition compound eye (Figures 3B,C): several photoreceptors are grouped together to form the retinula of a single ommatidium, which view the same visual space. The axons of these photoreceptors (to avoid additional complications arising in color channels, we consider only receptors that terminate in the lamina in our model) transmit information to downstream neurons that pool their information indiscriminately (Rivera-Alba et al., 2011; Matsushita et al., 2022). Thus these photoreceptors in a single ommatidium together form one visual sampling channel. The case in the vertebrate retina becomes somewhat more complicated, since information is first pooled from small groups of photoreceptors by bipolar cells, before these in turn are pooled by retinal ganglion cells (RGCs) (Masland, 2012) (Figure 4A). Individual photoreceptors only constitute separate sampling units in rare cases, and even in those cases they are subsequently sampled via bipolar cells, which constitute the functional sampling units. A number of different bipolar (Behrens et al., 2016) and RGC (Baden et al., 2016) types form parallel visual channels with different spatial sampling bases. Since the RGCs ultimately provide the information channel from the eye to the rest of the brain, one might argue that they constitute the functional sampling base of the retina. However, recent work provides strong evidence for non-linear integration of bipolar cell information by RGCs, suggesting that these function as separate sub-units at least to some degree (Zapp et al., 2022). For the mouse example in our study (Figure 3J), we focused only on rod pathways that are active in dim light, which strongly reduces the number of bipolar cells, and assumed these to be the elementary sampling units. Thus, in Eq. 4 we would have to adjust d for the number of photoreceptors integrated by one rod bipolar cell [see for example (Behrens et al., 2016)].
Figure 4. Different types of spatial summation. (A) Spatial convergence in the vertebrate retina occurs in two stages: photoreceptors signals are pooled by bipolar cells (sample unit), which in turn are integrated by retina ganglion cells (RGC), which transmit visual information to the brain. Note that the tiling of RGCs differs from the of bipolar cells, thus resulting in an altered sampling base after spatial summation. (B) In most insects, photoreceptors from one ommatidium of the eye (Figures 2B,C) project as a combined sample unit to a single cartridge in the lamina, where lamina monopolar cells (LMC) integrate their information and relay it to the next brain area. Particularly in nocturnal insects, LMCs with lateral processes that extend into neighboring cartridges can thus pool information across sample units, and perform spatial summation as a result. Note that the tiling of these LMCs corresponds to that of the photoreceptors, so that the spatial sampling base does not change when LMCs sum information in space. (C) The neural superposition eyes of Diptera possess a special organization to integrate information spatially: the photoreceptors in each of their ommatidia view different points in space, as represented here by different colors for photoreceptors in each ommatidium. In neighboring ommatidia, one receptor each views the same point in space (same color code). These receptors project to the same cartridge in the lamina, so that the lamina units represent on “pixel” of the image. This way, light representing this “pixel” is focused by six neighboring facets. Neural summation of information across neighboring sample units takes place by integrating information from the photoreceptors in one ommatidium. (D) The same starlight scene as in Figure 1A viewed through a mouse eye without and (E) with spatial summation. (F) Same moonlit scene as in Figure 3A viewed through a fruitfly eye without and (G) with spatial summation.
As a consequence, when attempting to obtain a realistic impression of an animals’ view of a natural scene, it is vital to consider the functional visual sampling base relevant to the question (and thus the visual pathway) at hand.
Dynamic adjustments of visual sensitivity
In addition to the anatomical adjustments that improve visual sensitivity at night, the dynamic range of the photoreceptors and subsequent neural processing are often adjusted to the ambient light intensity. This is important, because for any given neuron the dynamic range is much smaller than the range of light intensities experienced between day and night (Walcott, 1975). The contrast range of single visual scenes at a given light intensity, however, is generally constrained to one, maximum two log units (Figure 1F), and thus more closely matches the dynamic ranges of visual neurons (at about two orders of magnitude response variation) (Laughlin S., 1981; Laughlin S. B., 1981; Srinivasan et al., 1982; Laughlin, 1989; van Hateren and van der Schaaf, 1996). By dynamically matching the visual responses to the current contrast range, the visual system can encode scenes over the entire intensity range an animal experiences. This so-called light adaptation takes place both at the level of the eyes’ optics, and at several processing stages of the visual pathway.
Dynamic adjustments of the optics
In all three eye types, the amount of light entering the photoreceptors can be regulated dynamically. This is important, because photoreceptors have a limited dynamic range (Walcott, 1975). Like an overexposed camera image, allowing too much light into the eye saturates the photoreceptors, thereby severely compromising visual perception. Thus, a pupil restricts the incoming visual signal to a manageable dynamic range. In vertebrate camera eyes, the iris works as a pupil and contracts to restrict the amount of light entering through the lens (Walls, 1942). In arthropod compound eyes, the pupil is composed of migrating pigment granules. In apposition eyes, pigment typically moves radially toward (light adaptation) and away from the rhabdom (dark adaptation) (Ribi, 1978). In the light-adapted state the pigment absorbs stray photons and additionally sharpens spatial acuity (Land and Osorio, 1990). In superposition eyes, pigment moves longitudinally to separate (light adaptation) or combine the light rays (dark adaptation) from neighboring optical units (Warrant and McIntyre, 1996). The opening or closing of the pupil is controlled by light detected within the eye itself (Nilsson et al., 1992).
For all estimates of animal visual-scene perception in this study, we assumed open camera and superposition pupils. This is realistic for hawkmoths, whose pupil opens at twilight intensities (Stöckl et al., 2016a), while in mice the pupil is only fully open for starlight intensities, and closes gradually with increasing light intensity over a range of at least five orders of magnitude (Bushnell et al., 2016). One important aspect concerning ALAN is to what degree pupil dynamics are affected by light pollution, in particular direct light pollution that appears suddenly, such as passing cars, or streetlights animals encounter while on the move. Do these alter the pupil dynamics, and as a result render the animals subsequently less sensitive when the light has passed, while the pupil still requires some time to open again? Careful measurements of the responses of pupils in camera and superposition compound eyes to dynamic ALAN encounters will be required to address these questions.
Temporal dynamics of photoreceptors and downstream neurons
As depicted in Eq. 4, the photoreceptors’ response times directly scale the absorbed photons during a single processing interval or integration time. Consequently, in most animal species dark-adapted photoreceptors respond more slowly than light-adapted ones (Laughlin and Weckström, 1993; Juusola and Hardie, 2001; Reber et al., 2015; Stöckl et al., 2016a), thus effectively increasing their integration time, the time during which photons are integrated to generate a response. Just as when we increase the exposure time for a camera, more photons are thus collected, and a higher signal to noise ratio is achieved (Snyder, 1977, 1979; Lythgoe, 1979; Dubs, 1981; Warrant, 1999; Laughlin S. B., 1981; Warrant and McIntyre, 1996). Moreover, integration times tend to be longer in the dark-adapted than the light-adapted state (Laughlin and Weckström, 1993; Juusola and Hardie, 2001; Reber et al., 2015; Stöckl et al., 2016a) and also longer in nocturnal than diurnal species (Laughlin and Weckström, 1993; Frederiksen et al., 2008; Stöckl A. L. et al., 2017; Frolov and Ignatova, 2019; Donner, 2021). Extremely long integration times have been measured in nocturnal toads [1.5 s (Donner, 1989)] and in a deep-sea crustacean [160 ms (Moeller and Case, 1995)]. Such long integration times lead to severe blurring of moving objects (such as predators or prey) or the visual surroundings of animals that move themselves, which makes them challenging for flying animals, or animals that need to chase fast moving prey.
Very similar to the dynamic adjustments in photoreceptors, temporal integration can also occur at other stages of visual processing to improve the SNR, albeit not by directly affecting the photon signals, but by summing visual responses correlated in time, while averaging out uncorrelated noise (Laughlin S. B., 1981; Snyder, 1977, 1979; Lythgoe, 1979; Warrant and McIntyre, 1996; Warrant, 1999; Stöckl et al., 2016a).
For this study, we chose not to address the effects of temporal response dynamics in animal visual systems in detail, because to assess these in relation to natural visual scenes, we would need to process time-varying visual inputs. Obtaining such data is considerably more challenging than obtaining still-image radiance maps–for technical reasons regarding both the camera and animal tracking. The imaging methods applied here [and in Hänel et al. (2018), Jechow et al. (2018)] generally require very long exposure times (e.g., darker than full moon: sky 2–3 min, below the horizon 8 min) for sampling the full intensity range using HDR bracketing (Hänel et al., 2018; Foster et al., 2021; Nilsson and Smolka, 2021), so filming a moving sequence in real time is not possible with this method. While it may be feasible to record an image set at regular intervals along the desired route (e.g., Wystrach et al., 2012) at the necessary exposure times, to simulate the changes in visual information as the animals follows this route, there is potential for significant changes to the visual environment during the time interval required that are far greater than those genuinely experienced by an animal following the route in real time. Moreover, to assess the impact on natural movement input, in particular that generated the animal’s own activity (the majority of movement perceived by terrestrial animals), movement sequences ought to realistically simulate the resulting image velocities and rotations experienced by the animals in question. However, published information on the fine-scale trajectories of nocturnal animals, in particular those of insects, are extremely sparse, even more so within natural environments. This is due to the methodological challenges of detecting small animals under low light conditions, outside of highly constrained and controlled lab environments. However, new tracking methods are continuously developed (Vo-Doan and Straw, 2020), including with potential for night-time applications (Haalck et al., 2020; Walter et al., 2021).
Spatial summation in the visual system
Just as for temporal integration, spatial summation of visual information, that is integration of information from neighboring sampling units (or “pixels” of the image), can increase the SNR. Since most visual signals are correlated in space (resulting from objects viewed by several neighboring sampling units), integrating information at the same point in time from neural elements processing neighboring “pixels” thus sums this correlated signal and averages out uncorrelated noise (Laughlin S. B., 1981; Snyder, 1977, 1979; Lythgoe, 1979; Warrant and McIntyre, 1996; Warrant, 1999). While spatial summation does not directly act on the photon catch, summing the visual signals generated by photons in the neighboring sampling units thus effectively increases the SNR by a virtual increase in photon numbers per summation sampling unit. This increase is calculated by multiplying the photon estimate in Eq. 4 by the effective number of summed sampling units nsum. nsum can be obtained from anatomical analyses, or by measuring the (half-width) acceptance angle △ρsum of the receptive field of the summing neuron and calculating the number of sampling units within it using the anatomical sampling base of the original visual units △φ (Warrant, 1999):
Unlike the signatures of temporal summation, the signatures of spatial summation manifest in the morphology of neurons: to integrate from neighboring visual units, neurons with extended processes that contact neurons are required, collecting local visual information (Figures 4A,B). In the vertebrate retina, the photoreceptor signals that are conveyed by bipolar cells are, in turn, integrated by RGCs (Figure 4A). While in the high resolution areas of the eye, such as fovea in many primates, only very few photoreceptors connect to a RGC center, in the visual periphery thousands of photoreceptors can converge onto a single ganglion cell (Roska and Werblin, 2001; Chichilnisky, 2005; Helmstaedter et al., 2013). With such convergence, the spatial resolution of the corresponding visual percept decreases (as a function of the number of photoreceptors connected via bipolar cells to one ganglion cell) to the receptive field half-width of the RGCs, while the sensitivity greatly increases with the number of pooled sampling units (compare Figures 4D,E). Since the RGCs with large receptive fields tile the retina at lower density than the bipolar cells they integrate, this means a conversion to the sampling base of the integrating neurons (compare number of pixels in Figures 4D,E) not only for the spatial resolution, but also for the spatial sampling base of the image.
In non-dipteran insects, spatial summation occurs in the lamina (Stöckl et al., 2020), the first visual processing layer of the insect brain. This neuropil’s main relay neurons, lamina monopolar cells, receive information directly from photoreceptors, relaying it further downstream (Figure 4B). In nocturnal insects, the extent of the lateral processes can be extensive, and reach dozens of neighboring visual units (Strausfeld and Blest, 1970; Ohly, 1975; Ribi, 1977; Wolburg-Buchholz, 1979; Greiner et al., 2004a; Stöckl et al., 2016b). The lamina is divided into retinotopic cartridges, so that each cartridge receives information from one ommatidium, and possesses a set of spatially summing neurons. As a result, the sampling base does not change with spatial summation. Even though the spatial resolution of the visual percept changes, due to the lateral integration of visual information by lamina neurons.
Flies possess an exceptional ocular and laminar anatomical organization, that inherently incorporates spatial summation, the so-called neural superposition compound eye. The photoreceptors of each retinal unit that underly a facet of the compound eye receive light from different neighboring locations in a visual scene (Braitenberg, 1967; Kirschfeld, 1967). All photoreceptors that view the same location, originating in different sampling units of the eye, project to the same unit in the lamina (Figure 4C). Thus, the photoreceptors contributing to viewing one “pixel” of the image receive light through six neighboring facets, as they are located in six neighboring ommatidia. Spatial summation, that is integration of information over neighboring points in space, is achieved by pooling information from the photoreceptors of each ommatidium–since these view different points in space. Practically, this is accomplished by electrically coupling photoreceptors within one ommatidium (Dubs et al., 1981; Dubs, 1982). Because each ommatidium in the fly eye contains six photoreceptors that project to the lamina sample neighboring points in space (Braitenberg, 1967; Kirschfeld, 1967), spatial summation by this neural superposition mechanism is limited in extent–yet it is still effective to make the visual percept more reliable at the limits of the fly eye’s visual sensitivity (Figures 3K, 4F,G).
Light-dependent dynamics of spatial processing
In both vertebrates and insects, spatial processing in the visual periphery is adjusted according to the ambient light intensity, integrating information in space in dim light and thereby sacrificing spatial resolution to gain sensitivity (van Hateren, 1992; Stöckl A. et al., 2017; Wienbar and Schwartz, 2018). In vertebrates, RGCs linearly integrate the responses of bipolar cells in their receptive fields at low light intensities, while they integrate thresholded individual responses, thus resolving smaller spatial patterns, at higher light intensities (Farrow et al., 2013; Grimes et al., 2014; Mazade and Eggers, 2016). Although the light intensity at which this change occurs has been reported to be at the threshold where vision switches from relying on the less sensitive cone cells to the more sensitive rod cells, the exact luminance can differ for individual ganglion cells, as well as between ganglion cell types, and the ultimate effect on the spatial properties can also differ between RGC types (Tikidji-Hamburyan et al., 2015). In insects, under dim conditions visual information is spatially integrated by lamina monopolar cells, the first visual processing stage downstream from the photoreceptors, (van Hateren, 1992; Stöckl et al., 2020). In bright light, their receptive fields narrow to a single visual sampling base, and often show evidence of lateral inhibition, similar to the receptive fields of many vertebrate RGCs.
While these light-dependent changes in spatial processing are well described under constant illumination conditions, it is not clear how rapidly this dynamic spatial tuning can switch: experiments in both vertebrates and invertebrates were conducted with a minimum of 5 min light adaptation to a background intensity. However, animals moving within natural light environments, particularly those that include artificial light, experience rapid changes in local and average illumination that can occur within seconds–as animals turn toward or away from light sources. To understand how their visual systems adjust to these changes, further investigations into the speed of these light-dependent dynamics of spatial and temporal tuning are required.
Since visual scenes can greatly differ in local light intensity (Figures 1, 2), both the temporal dynamics of the luminance-dependent processing, and, crucially, their spatial dynamics, are relevant for the visual percept. Studies of motion adaptation in wide-field neurons in insects (Li et al., 2021), and contrast adaptation in vertebrate RGCs found that both local and global processes play a role (Garvert and Gollisch, 2013). In vertebrate RGCs, local and global adaptation contribute with different proportions for different RGC types (Khani and Gollisch, 2017). The stimulus protocol used in Farrow et al. (2013) suggests that the light-dependent changes in spatial processing are also triggered by illumination local to the RGC, rather than the average background intensity. On the other hand, effects of neural modulation on light-dependent adaptations have been described that ranged well beyond the receptive fields of individual ganglion cells (Eggers and Lukasiewicz, 2010). These also feature very prominently in the circadian system (Ribelayga et al., 2008; Allen et al., 2014). Since the circadian system modulates light adaptation in insect eyes (Nilsson et al., 1992), and prominent circadian changes have been observed in the insect lamina (Meinertzhagen and Pyza, 1996), where dynamic spatial processing takes place, it is very possible that dynamic spatial tuning in insects is also modulated by the circadian system. It would therefore be highly informative to determine how local fluctuations in light intensity (such as a street-light in a nocturnal visual scene) alter luminance-dependent spatial and temporal tuning in individual RGCs in vertebrates, and individual lamina neurons in insects–to understand how such spatially dynamic light environments affects the visual percept of nocturnal animals.
Impact of artificial light at night on an animal’s visual percept
To estimate how visual responses play out for natural scenes, it is crucial to understand both light stimulation and the internal states that cause changes in the spatial (and temporal tuning of visual neurons), as well as the dynamics with which these changes occur. This is all the more important for light polluted scenes, since, almost by definition, these have a broader dynamic range than natural scenes.
Real-world environments
A noticeable feature of night-time scenes with light pollution is that both direct and indirect light pollution enhance across-scene extremes in light intensity. Skyglow (Figures 2A,D), particularly at the horizon, increases the brightness of the upper half of the scene (Figures 2E,F), while illuminating the ground diffusely, unlike direct illumination from the sun or moon (compare Figures 1E,F). This results in an average intensity-difference between above and below the horizon of more than tenfold. Similarly, direct light pollution on the ground (Figure 2B) greatly increases the brightness in the lower half of the visual field, leaving the sky distinctly dimmer (N.B. the Aurora borealis in this scene alone raised the sky brightness by two orders of magnitude compared to clear starlight). Even though one might assume that the added light could benefit the visual systems of nocturnal animals, which are brought to their limits precisely because not enough light is available (Figures 3E,F, 4D,F), the distribution of the artificial night, both spatially and temporally, can lead to detrimental effects for animal vision.
This is because, as discussed earlier, visual neurons only have limited dynamic ranges. If the brightness distribution of a visual scene spans many orders of magnitude, the instantaneous mapping to the dynamic range of a neuron results in poor resolution of potentially crucial differences in intensity. For our modeling, we assumed a typical neural response mapping onto two orders of magnitude (corresponding to 95% of the model neuron’s response range) of light intensity (Laughlin S. B., 1981; Srinivasan et al., 1982; Laughlin, 1989; Figures 5A,D). Of the different options for anchoring the mapping reference points we chose to map the center of the response range (inflection point of the response curve) to the mean intensity of the scene (Figures 5B,E,H), and to the brightness range that occurs most frequently [median, (Figures 5C,F,I)].
Figure 5. Modeled neural responses to visual scenes with and without light pollution. (A,D,G) Natural scenes on moonless nights with different levels of light pollution: (A) no light pollution, (B) same location with indirect light pollution (sky glow) and (C) direct lighting of a high albedo ground. Overlaid are the distributions of photon numbers per sample unit with spatial summation applied for the hawkmoth D. elpenor. The photon distributions were mapped to a sigmoidal response curve spanning 2.5 log10-units of brightness, which were anchored either to the mean (dark green) or median (light green) of the distribution. Panels (B,E,H) show the responses for mean mapping, and panels (C,F,I) for median mapping, resampled to the sampling base of the animal.
We applied this mapping to different visual scenes all viewed through the eyes of D. elpenor for comparability. Even though a non-polluted starlit scene (Figure 5A, same as Figure 1A) also has large brightness differences between the sky and the ground, these still map well onto the 2-log-unit range of the neuronal response, so that both contrasts within the sky and within the below-horizon environment are strong. As, a result, if the animal moved its gaze across the scene, the neuron would be instantaneously capable of encoding all the brightness differences, without losing information to saturation or noise at the brightness distribution’s upper and lower ends, respectively. While the ground was of similar brightness in a comparable environment with indirect light pollution (sky glow on the horizon), the distribution was skewed toward higher light intensities (Figure 5D). Even though this effect was quite subtle, it ultimately resulted in the compression of either the dimmer (mean mapping, Figure 5E) or brighter (median mapping, Figure 5F) intensities, and thus, in a neuron adapted to the average (mean or median) intensity of the scene, a loss of contrast discriminability either above or below the horizon. An extreme version of this unnaturally skewed brightness histogram is created by light pollution combined with a high ground albedo (Figure 5G). Thus, in visual scenes where light pollution generates extreme brightness differences, the neurons’ dynamic range might not be sufficient to resolve contrasts across the entire scene.
Our assumption that neurons adapt only to the average scene brightness is certainly a simplification, since visual neurons also adapt to the local luminance level in combination with global adaptation effects (see Light-dependent dynamics of spatial processing). Information about the balance of these two effects in specific neuron types would be crucial to more accurately predict a visual system’s responses to nocturnal light polluted scenes. Likewise, the speed at which these light-dependent adjustments occur would be needed to describe responses to dynamic scenes and shifting gaze. The latter aspect is important, given that similar effects of overextending the dynamic ranges of visual neurons can occur across time, if the light levels the animals experience change rapidly–for example because the light source moves (such as the headlights of a car), or because the animal moves (for example flies past a row of streetlights). While the light intensity animals experience over time also fluctuates with them moving in their environment, these fluctuations mostly remain within the 2.5 log10-unit dynamic range of visual neurons (van Hateren and van der Schaaf, 1996). However, flood lights or car headlights viewed on a starry night can exceed this range distinctly (for some examples see Figure 2; Foster et al., 2021), and thus cause similar temporal contrast coding problems as discussed for the spatial domain (Figure 5).
Moreover, in addition to light-dependent neural processing to the sensitivity of visual neurons, it is important to account for changes to their spatial and temporal tuning. Where visual neuron adapt locally to vastly different intensities in different parts of the visual field, how does the visual system form a consistent percept from units that vary locally in both their spatial and temporal resolution–and how does that affect the animal’s behavior? Similarly, if there are rapid changes in spatial or temporal tuning that correspond to rapid changes in light intensity, how does that affect animals’ visually guided behaviors, such as flight control or the detection of foraging targets? If the adjustments occur on longer time-scales than the fluctuation in light intensity, the visual responses would not be optimally adjusted to the current visual environment. For example, an animal might be exposed to bright light sources, resulting in a more constricted pupil and a low level of spatial summation. If it then flies away from the light source, its visual sensitivity would not be sufficient to resolve an unpolluted visual scene (i.e., compare Figures 4D,E,F,G). How impactful these effects would be, would depend on the dynamics of these optical and neural adjustments. For example, in terms of pupil dynamics, there are distinct differences between vertebrates and insects: while a mouse pupil can change its aperture over the course of seconds (Grozdanic et al., 2003), the pupil of a hawkmoth takes an order of magnitude longer to close and more than 10 min to reach its fully open state (Nilsson et al., 1992; Stöckl et al., 2016a).
Thus, to fully understand the impact of ALAN on the visual system of nocturnal animals, we require a better understanding of the spatial response dynamics of visual neurons across a visual scene, as well as the temporal dynamics of dim light-dependent adjustments of spatial and temporal tuning and contrast sensitivity.
Outlook
As we have seen, ALAN, both direct and indirect, can increase the range of intensities across a visual scene drastically, making it challenging to match this wide range of inputs to the limited dynamic ranges of photoreceptors and downstream neurons (Figure 5). However, often the bright and dim areas cluster in particular regions of the visual field (sky glow in the dorsal hemisphere, direct light pollution often in ventral hemisphere for flying animals). Thus, if animals maintain a stable eye orientation with respect to the horizon, local adaptation (neurons adapting to the regional rather than overall brightness levels) could resolve this dynamic range problem. Since many visual pathways show both local and global adaptation, as well as global state-dependent control, we need to understand how these processes interact to predict how nocturnal animals view scenes with great differences in spatial and temporal intensity.
Artificial illumination not only introduces more light to an animal’s visual environment, but also alters its spectrum at the wavelengths most prevalent in the artificial light source. By increasing the relative intensity in specific regions of the illumination spectrum, this can alter chromatic contrasts between objects and background, which is of particular importance for camouflaged animals (Briolat et al., 2021). Images recorded with RGB cameras, such as those presented in this study, have somewhat limited spectral resolution, but can be used to coarsely estimate chromatic contrast in the human-visible wavelengths, and even into the UV with the correct choice of lens and filters (Troscianko and Stevens, 2015). Recently developed hyperspectral imaging systems (Nevala and Baden, 2019; Tedore and Nilsson, 2019) may make it possible to accurately estimate these changes in chromatic contrast accurately for a range of species with well-studied color vision.
While we focused in this overview on spatial information in images and their representation through animal eyes, temporal changes in the perceived light intensity are of crucial importance for nocturnal animals. Just as they vary over the course of the day, intensities perceived by the eye can also routinely fluctuate by a factor of 100 over shorter time periods, when flying under foliage or in other shaded spaces (Warrant, 2008). These extremes are further enhanced by ALAN, in particular by direct light sources that are many magnitudes brighter than the surrounding natural scenery (compare Figures 1, 2). Emerging from a forest or region of the undergrowth, or turning toward one of these sources after facing a less polluted area, can suddenly change the perceived light intensity on the retina drastically–though how drastically is yet to be reported. How the dynamic adjustments of visual sensitivity, both in the eye (pupil) and the subsequent neural processing (temporal summation, spatial summation), respond to these sudden changes in light intensity, also remains unstudied. While obtaining realistic videos of how a flying insect might perceive a nocturnal visual scene remains technically challenging, studying the dynamic properties of nocturnal visual processing with realistic intensity fluctuations [as has been done for daylight scenes (van Hateren and van der Schaaf, 1996)] is very achievable and would be a crucial first step toward understanding better how sudden changes in light perception as can be caused by direct ALAN affect the visual systems of nocturnal animals.
As described, animals generally do not view static visual scenes, because they themselves are in motion and thus perceive the relative motion of their environment. This also means that, to some degree, they shape the scenes they perceive–they “choose” how long to gaze at certain aspects of a scene, at which distance to pass a street lamp, how fast to walk or fly, and thus shape both the spatial and temporal frequency content, as well as the intensity their eyes perceive. Whether and how nocturnal animals adapt their movements to optimally sample their environment when light conditions change–from day to night, from clear to clouded moon, from a light polluted to unpolluted scene, remains an open and an extremely important question. By actively shaping the visual content they perceive, nocturnal animals might be able to adjust to much more drastic light environments than we might expect–but on the other hand this factor may also make their visual tasks more challenging, if certain movements are required (for example sideways casting for smelling a trail) that deteriorate the visual input (where steady slow forward flying would have been better). As methods for free range tracking improve, we will gain a clearer understanding of the entire cascade of visual perception of natural scenes, from active selection to photon transduction, to subsequent processing. This will greatly enhance our understanding of how animals perceive their night time environments, and where ALAN poses challenges to their visual systems. The methods outlined here can help with this impression–and will be even more powerful once combined with realistic temporal images from animal movement tracks.
Modeling methods
To obtain the eyes’ spatial resolution, equirectangular images (Figure 3D) were filtered with a Gaussian filter with the half-width set to the acceptance angle Δρ of the visual units (photoreceptors for compound eyes and rod bipolar cells for the mouse). Importantly, the Gaussian filter accounted for the variation in angular distance between pixels along the elevation of the equirectangular images (for details see, Nilsson and Smolka, 2021).1 To calculate the photons obtained in each sampling channel per integration time, the light intensity of the imaged scenes was then calculated using Eq. 4, and multiplied by 100 nm to account for the width of a single photoreceptor spectral class (for details see, Stöckl A. et al., 2017; Stöckl A. L. et al., 2017). Spatial summation was implemented by Gaussian filtering with the spatial summation angle, while multiplying the photons per sampling channel by the number of summed units nS. Subsampling the visual scene to the animals’ sampling base was performed by selecting pixels from the equirectangular images at distances equal to the spatial sampling base Δφ (for compound eyes) or the spatial summation angle [for mouse spatial summation, where the sampling base changes from rod bipolar cells to RGCs (see Figure 4A)]. The number of sampling units per elevation was calculated using a sinusoidal projection (for details see, Nilsson and Smolka, 2021).
Data availability statement
The source data of the analysed images (with calibrated radiance per pixel) is available via Figshare: https://doi.org/10.6084/m9.figshare.20188157.
Author contributions
AS and JF: conceptualization, methodology, validation, formal analysis, resources, writing – original draft, review and editing, visualization, and funding acquisition. JF: data curation. Both authors contributed to the article and approved the submitted version.
Funding
The authors acknowledged support to AS by the Bavarian Academy of Sciences and Humanities, the Center for the Advanced Study of Collective Behaviour and the Zukunftskolleg of the University of Konstanz (Funded as part of the Excellence Strategy of the German Federal and State Governments), and to AS and JF by the German Research Council (STO1255/2-1 and FO1308/3-1), and travel support to JF from the Royal Physiographic Society of Lund, the Royal Swedish Academy of Sciences and INTERACT Transnational Access (grant: ANSOC).
Acknowledgments
We would like to thank Marie Dacke and Dan-Eric Nilsson for providing the cameras for image acquisition and Jochen Smolka for valuable advice on the data-analysis.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
References
Aho, A. C., Donner, K., Hydén, C., Larsen, L. O., and Reuter, T. (1988). Low retinal noise in animals with low body temperature allows high visual sensitivity. Nature 334, 348–350. doi: 10.1038/334348a0
Aho, A.-C., Donner, K., Helenius, S., Larsen, L., and Reuter, T. (1993). Visual performance of the toad (Bufo bufo) at low light levels: Retinal ganglion cell responses and prey-catching accuracy. J. Comp. Physiol. A 172, 671–682. doi: 10.1007/BF00195393
Allen, A. E., Storchi, R., Martial, F. P., Petersen, R. S., Montemurro, M. A., Brown, T. M., et al. (2014). Melanopsin-driven light adaptation in mouse vision. Curr. Biol. 24, 2481–2490. doi: 10.1016/j.cub.2014.09.015
Baden, T., Berens, P., Franke, K., Rosón, M. R., Bethge, M., and Euler, T. (2016). The functional diversity of retinal ganglion cells in the mouse.. Nature 529, 345–350. doi: 10.1038/nature16468
Baker, R. R., and Sadovy, Y. (1978). The distance and nature of the light-trap response of moths. Nature 276, 818–821. doi: 10.1038/276818a0
Baylor, D. A., Matthews, G., and Yau, K. W. (1980). Two components of electrical dark noise in toad retinal rod outer segments.. J. Physiol. 309, 591–621. doi: 10.1113/jphysiol.1980.sp013529
Behrens, C., Schubert, T., Haverkamp, S., Euler, T., and Berens, P. (2016). Connectivity map of bipolar cells and photoreceptors in the mouse retina. Elife 5:e20041. doi: 10.7554/eLife.20041
Bennie, J., Davies, T. W., Cruse, D., Inger, R., and Gaston, K. J. (2015). Cascading effects of artificial light at night: Resource-mediated control of herbivores in a grassland ecosystem. Philos. Trans. Royal Soc. B Biol. Sci. 370:20140131. doi: 10.1098/rstb.2014.0131
Bhagavatula, P. S., Claudianos, C., Ibbotson, M. R., and Srinivasan, M. V. (2011). Optic flow cues guide flight in birds. Curr. Biol. 21, 1794–1799. doi: 10.1016/j.cub.2011.09.009
Blest, A. D., and Land, M. F. (1977). The physiological optics of dinopis subrufus l. Koch: A fish-lens in a spider., Proc. R. Soc. Lond. B Biol. Sci. 196, 197–222. doi: 10.1098/rspb.1977.0037
Braitenberg, V. (1967). Patterns of projection in the visual system of the fly. i. retina-lamina projections.. Exp. Brain Res. 3, 271–298. doi: 10.1007/BF00235589
Briolat, E. S., Gaston, K. J., Bennie, J., Rosenfeld, E. J., and Troscianko, J. (2021). Artificial nighttime lighting impacts visual ecology links between flowers, pollinators and predators. Nat. Commun. 12:4163. doi: 10.1038/s41467-021-24394-0
Bushnell, M., Umino, Y., and Solessio, E. (2016). A system to measure the pupil response to steady lights in freely behaving mice. J. Neurosci. Methods 273, 74–85.
Caveney, S., and McIntyre, P. (1981). Design of graded-index lenses in the superposition eyes of scarab beetles. Philos. Trans. R. Soc. Lond. B. Biol. Sci. 294, 589–632.
Chichilnisky, E. J. (2005). Detection sensitivity and temporal resolution of visual signals near absolute threshold in the salamander retina. J. Neurosci. 25, 318–330. doi: 10.1523/JNEUROSCI.2339-04.2005
Cohn, T. E. (1983). Receiver operating characteristic analysis of photoreceptor sensitivity. IEEE Trans. Syst. Man Cybern. 13, 873–881. doi: 10.1109/TSMC.1983.6313082
Currea, J. P., Frazer, R., Wasserman, S. M., and Theobald, J. (2022). Acuity and summation strategies differ in vinegar and desert fruit flies. iScience 25:103637. doi: 10.1016/j.isci.2021.103637
Davies, M. N. O., and Green, P. R. (1994). “Introduction to section III,” in Perception and motor control in birds, eds M. N. O. Davies and P. R. Green (Berlin: Springer), doi: 10.1007/978-3-642-75869-0_13
De Vries, H. (1943). The quantum character of light and its bearing upon threshold of vision. Physica 10, 553–564.
Donner, K. (1989). Visual latency and brightness: An interpretation based on the responses of rods and ganglion cells in the frog retina. Vis. Neurosci. 3, 39–51. doi: 10.1017/s0952523800012499
Donner, K. (2021). Temporal vision: Measures, mechanisms and meaning. J. Exp. Biol. 224:jeb222679. doi: 10.1242/jeb.222679
Doren, B. M. V., Horton, K. G., Dokter, A. M., Klinck, H., Elbin, S. B., and Farnsworth, A. (2017). High-intensity urban light installation dramatically alters nocturnal bird migration. Proc. Natl. Acad. Sci. U.S.A. 114, 11175–11180. doi: 10.1073/pnas.1708574114
Dubs, A. (1981). Non-linearity and light adaptation in the fly photoreceptor. J. Comp. Physiol. 144, 53–59. doi: 10.1007/BF00612797
Dubs, A. (1982). The spatial integration of signals in the retina and lamina of the fly compound eye under different conditions of luminance. J. Comp. Physiol. 146, 321–343. doi: 10.1007/BF00612703
Dubs, A., Laughlin, S. B., and Srinivasan, M. V. (1981). Single photon signals in fly photoreceptors and first order interneurones at behavioral threshold.. J. Physiol. 317, 317–334. doi: 10.1113/jphysiol.1981.sp013827
Eggers, E. D., and Lukasiewicz, P. D. (2010). Interneuron circuits tune inhibition in retinal bipolar cells. Journal of Neurophysiology 103, 25–37.
Emlen, S. T. (1967a). Migratory orientation in the indigo bunting, passerina cyanea: Part i: Evidence for use of celestial cues. Auk 84, 309–342. doi: 10.2307/4083084
Emlen, S. T. (1967b). Migratory orientation in the indigo bunting, passerina cyanea. part II: Mechanism of celestial orientation. Auk 84, 463–489. doi: 10.2307/4083330
Exner, S. (1891). Die Physiologie der facettirten Augen von Krebsen und Insecten. Leipzig: Deuticke. doi: 10.5962/bhl.title.2149
Falchi, F., Cinzano, P., Duriscoe, D., Kyba, C. C. M., Elvidge, C. D., Baugh, K., et al. (2016). The new world atlas of artificial night sky brightness. Sci. Adv. 2:e1600377. doi: 10.1126/sciadv.1600377
Farrow, K., Teixeira, M., Szikra, T., Viney, T. J., Balint, K., Yonehara, K., et al. (2013). Ambient illumination toggles a neuronal circuit switch in the retina and visual perception at cone threshold. Neuron 78, 325–338. doi: 10.1016/j.neuron.2013.02.014
Foster, J. J., Smolka, J., Nilsson, D.-E., and Dacke, M. (2018). How animals follow the stars. Proc. Royal Soc. B 285:20172322. doi: 10.1098/rspb.2017.2322
Foster, J. J., Tocco, C., Smolka, J., Khaldy, L., Baird, E., Byrne, M. J., et al. (2021). Light pollution forces a change in dung beetle orientation behavior. Curr. Biol. 31, 3935.e–3942.e doi: 10.1016/j.cub.2021.06.038
Frederiksen, R., Wcislo, W. T., and Warrant, E. J. (2008). Visual reliability and information rate in the retina of a nocturnal bee.. Curr. Biol. 18, 349–353. doi: 10.1016/j.cub.2008.01.057
Frolov, R. V., and Ignatova, I. I. (2019). Electrophysiological adaptations of insect photoreceptors and their elementary responses to diurnal and nocturnal lifestyles. J. Comp. Physiol. A 206, 55–69. doi: 10.1007/s00359-019-01392-8
Fu, Y., and Yau, K.-W. (2007). Phototransduction in mouse rods and cones. Pflugers Arch. 454, 805–819. doi: 10.1007/s00424-006-0194-y
Garvert, M. M., and Gollisch, T. (2013). Local and global contrast adaptation in retinal ganglion cells. Neuron 77, 915–928. doi: 10.1016/j.neuron.2012.12.030
Gaston, K. J., and de Miguel, A. S. (2022). Environmental impacts of artificial light at night. Annu. Rev. Environ. Resour. 47, doi: 10.1146/annurev-environ-112420-014438
Gaston, K. J., Davies, T. W., Bennie, J., and Hopkins, J. (2012). REVIEW: Reducing the ecological consequences of night-time light pollution: Options and developments. J. Appl. Ecol. 49, 1256–1266. doi: 10.1111/j.1365-2664.2012.02212.x
Geng, Y., Schery, L. A., Sharma, R., Dubra, A., Ahmad, K., Libby, R. T., et al. (2011). Optical properties of the mouse eye. Biomed. Opt. Express 2:717. doi: 10.1364/BOE.2.000717
Giavi, S., Fontaine, C., and Knop, E. (2021). Impact of artificial light at night on diurnal plant-pollinator interactions. Nat. Commun. 12:1690. doi: 10.1038/s41467-021-22011-8
Gonzalez-Bellido, P. T., Wardill, T. J., and Juusola, M. (2011). Compound eyes and retinal information processing in miniature dipteran species match their specific ecological demands.. Proc. Natl. Acad. Sci. U.S.A. 108, 4224–4229. doi: 10.1073/pnas.1014438108
Greiner, B., Narendra, A., Reid, S. F., Dacke, M., Ribi, W. A., and Zeil, J. (2007). Eye structure correlates with distinct foraging-bout timing in primitive ants.. Curr. Biol. 17, R879–R880. doi: 10.1016/j.cub.2007.08.015
Greiner, B., Ribi, W. A., Wcislo, W. T., and Warrant, E. J. (2004b). Neural organisation in the first optic ganglion of the nocturnal bee Megalopta genalis.. Cell Tissue Res. 318, 429–437. doi: 10.1007/s00441-004-0945-z
Greiner, B., Ribi, W. A., and Warrant, E. J. (2004a). Retinal and optical adaptations for nocturnal vision in the halictid bee Megalopta genalis.. Cell Tissue Res. 316, 377–390. doi: 10.1007/s00441-004-0883-9
Grimes, W. N., Schwartz, G. W., and Rieke, F. (2014). The synaptic and circuit mechanisms underlying a change in spatial encoding in the retina.. Neuron 82, 460–473. doi: 10.1016/j.neuron.2014.02.037
Grozdanic, S., Betts, D. M., Allbaugh, R. A., Sakaguchi, D. S., Kwon, Y. H., Kardon, R. H., et al. (2003). Characterization of the pupil light reflex, electroretinogram and tonometric parameters in healthy mouse eyes. Curr. Eye Res. 26, 371–378. doi: 10.1076/ceyr.26.5.371.15439
Grubisic, M., van Grunsven, R., Kyba, C., Manfrin, A., and Hölker, F. (2018). ‘Insect declines and agroecosystems: Does light pollution matter?’. Ann. Appl. Biol. 173, 180–189. doi: 10.1111/aab.12440
Haalck, L., Mangan, M., Webb, B., and Risse, B. (2020). Towards image-based animal tracking in natural environments using a freely moving camera. J. Neurosci. Methods 330, 108455. doi: 10.1016/j.jneumeth.2019.108455
Hänel, A., Posch, T., Ribas, S. J., Aubé, M., Duriscoe, D., Jechow, A., et al. (2018). Measuring night sky brightness: Methods and challenges. J. Quant. Spectrosc. Radiat. Transf. 205, 278–290. doi: 10.1016/j.jqsrt.2017.09.008
Healy, S., and Guilford, T. (1990). Olfactory-bulb size and nocturnality in birds. Evolution 44:339. doi: 10.2307/2409412
Helmstaedter, M., Briggman, K. L., Turaga, S. C., Jain, V., Seung, H. S., and Denk, W. (2013). Connectomic reconstruction of the inner plexiform layer in the mouse retina. Nature 500, 168–174.
Henriksson, J. T., Bergmanson, J. P., and Walsh, J. E. (2010). Ultraviolet radiation transmittance of the mouse eye and its individual media components. Exp. Eye Res. 90, 382–387. doi: 10.1016/j.exer.2009.11.004
Honkanen, A., Takalo, J., Heimonen, K., Vähäsöyrinki, M., and Weckström, M. (2014). Cockroach optomotor responses below single photon level.. J. Exp. Biol. 217(Pt 23), 4262–4268. doi: 10.1242/jeb.112425
Hornstein, E. P., Pope, D. R., and Cohn, T. E. (1999). Noise and its effects on photoreceptor temporal contrast sensitivity at low light levels. J. Opt. Soc. Amer. A 16:705.
Howard, J., and Snyder, A. W. (1983). Transduction as a limitation on compound eye function and design. Proc. R. Soc. B. 217, 287–307. doi: 10.1242/jeb.015396
Howard, J., Dubs, A., and Payne, R. (1984). The dynamics of phototransduction in insects. J. Comp. Physiol. A. 154, 707–718. doi: 10.1007/BF01350224
Jander, U., and Jander, R. (2002). Allometry and resolution of bee eyes (Apoidea).. Arthropod. Struct. Dev. 30, 179–193. doi: 10.1016/s1467-8039(01)00035-4
Jechow, A., Ribas, S. J., Domingo, R. C., Hölker, F., Kolláth, Z., and Kyba, C. C. (2018). Tracking the dynamics of skyglow with differential photometry using a digital camera with fisheye lens. J. Quant. Spectrosc. Radiat. Transf. 209, 212–223.
Johnsen, S., Kelber, A., Warrant, E., Sweeney, A. M., Widder, E. A., and Lee, R. L. Jr., et al. (2006). Crepuscular and nocturnal illumination and its effects on color perception by the nocturnal hawkmoth Deilephila elpenor.. J. Exp. Biol. 209(Pt 5), 789–800. doi: 10.1242/jeb.02053
Jones, T. K., and Moss, C. F. (2021). Visual cues enhance obstacle avoidance in echolocating bats. J. Exp. Biol. 224:jeb241968. doi: 10.1242/jeb.241968
Juusola, M., and Hardie, R. C. (2001). Light adaptation in drosophila photoreceptors: I. response dynamics and signaling efficiency at 25°C. J. Gen. Physiol. 117, 3–25. doi: 10.1085/jgp.117.1.3
Kelber, A. (2002). Pattern discrimination in a hawkmoth: Innate preferences, learning performance and ecology. Proc. R. Soc. B. 269, 2573–2577. doi: 10.1098/rspb.2002.2201
Khani, M. H., and Gollisch, T. (2017). Diversity in spatial scope of contrast adaptation among mouse retinal ganglion cells. J. Neurophysiol. 118, 3024–3043. doi: 10.1152/jn.00529.2017
Kirk, E. C. (2006). Eye morphology in cathemeral lemurids and other mammals. Folia Primatologica 77, 27–49. doi: 10.1159/000089694
Kirschfeld, K. (1967). [the projection of the optical environment on the screen of the rhabdomere in the compound eye of the Musca].. Exp. Brain Res. 3, 248–270. doi: 10.1007/BF00235588
Land, M. (1981). “Optics and vision in invertebrates,” in Handbook of sensory physiology, ed. H. Autrum (Berlin: Springer), 471–592.
Land, M. F., and Osorio, D. C. (1990). Waveguide modes and pupil action in the eyes of butterflies. Proc. R. Soc. Lond. B. 241, 93–100.
Laughlin, S. (1981). A simple coding procedure enhances a neurons information capacity. Zeitschrift für Naturforschung C 36, 910–912.
Laughlin, S. B. (1981). “Neural principles in the peripheral visual systems of invertebrates,” in Handbook of sensory physiology, ed. H. Autrum (Berlin: Springer), 133–280. doi: 10.1016/j.neuron.2013.04.029
Laughlin, S. B., and Lillywhite, P. G. (1982). Intrinsic noise in locust photoreceptors.. J. Physiol. 332, 25–45.
Laughlin, S., and Weckström, M. (1993). Fast and slow photoreceptors - a comparative study of the functional diversity of coding and conductances in the diptera. J. Comp. Physiol. A. 172, 593–609.
Li, J., Niemeier, M., Kern, R., and Egelhaaf, M. (2021). Disentangling of local and wide-field motion adaptation. Front. Neural Circ. 15:713285. doi: 10.3389/fncir.2021.713285
Lillywhite, P. (1977). Single photon signals and transduction in an insect eye. J. Comp. Physiol. 122, 189–200.
Lillywhite, P. G., and Laughlin, S. B. (1979). Transducer noise in a photoreceptor. Nature 277, 569–572. doi: 10.1038/277569a0
Longcore, T., and Rich, C. (2004). Ecological light pollution. Front. Ecol. Environ. 2:191–198. doi: 10.1890/1540-92952004002[0191:ELP]2.0.CO;2
Macgregor, C. J., and Scott-Brown, A. S. (2020). Nocturnal pollination: An overlooked ecosystem service vulnerable to environmental change. Emerg. Topics Life Sci. 4, 19–32. doi: 10.1042/ETLS20190134
Matsushita, A., Stewart, F., Ilic, M., Chen, P.-J., Wakita, D., Miyazaki, N., et al. (2022). Connectome of the lamina reveals the circuit for early colour processing in the visual pathway of a butterfly.. Curr. Biol. 32, 2291–2299. doi: 10.1016/j.cub.2022.03.066
Mazade, R. E., and Eggers, E. D. (2016). Light adaptation alters inner retinal inhibition to shape OFF retinal pathway signaling. J. Neurophysiol. 115, 2761–2778. doi: 10.1152/jn.00948.2015
McIntyre, P., and Caveney, S. (1985). Graded-index optics are matched to optical geometry in the superposition eyes of scarab beetles. Philos. Trans. R. Soc. Lond. B. Biol. Sci. 311, 237–269.
McIntyre, P., and Caveney, S. (1998). Superposition optics and the time of flight in onitine dung beetles. J. Comp. Physiol. A. Sens. Neural. Behav. Physiol. 183, 45–60.
Meinertzhagen, I., and Pyza, E. (1996). Daily rhythms in cells of the flys optic lobe: Taking time out from the circadian clock. Trends Neurosci. 19, 285–291. doi: 10.1016/S0166-2236(96)10033-3
Moeller, J., and Case, J. (1995). Temporal adaptations in visual systems of deep-sea crustaceans. Mar. Biol. 123, 47–54.
Moser, J. C., Reeve, J. D., Bento, J. M. S., Della Lucia, T. M. C., Cameron, R. S., and Heck, N. M. (2004). Eye size and behaviour of day- and night-flying leafcutting ant Alates. J. Zool. 264, 69–75.
Narendra, A., Reid, S. F., Greiner, B., Peters, R. A., Hemmi, J. M., Ribi, W. A., et al. (2011). Caste-specific visual adaptations to distinct daily activity schedules in Australian Myrmecia ants.. Proc. R. Soc. B. 278, 1141–1149. doi: 10.1098/rspb.2010.1378
Nelson, M. E., and MacIver, M. A. (2006). Sensory acquisition in active sensing systems. J. Comp. Physiol. A 192, 573–586.
Nevala, N. E., and Baden, T. (2019). A low-cost hyperspectral scanner for natural imaging and the study of animal colour vision above and under water. Sci. Rep. 9:10799. doi: 10.1038/s41598-019-47220-6
Nilsson, D.-E. (1989). “Optics and evolution of the compound eye,” in Facets of vision, eds D. G. Stavenga and R. C. Hardie (Berlin: Springer), 30–73.
Nilsson, D.-E., and Smolka, J. (2021). Quantifying biologically essential aspects of environmental light. J. Royal Soc. Interface 18:20210184.
Nilsson, D.-E., Hamdorf, K., and Höglund, G. (1992). Localization of the pupil trigger in insect superposition eyes. J. Comp. Physiol. A 170, 217–226.
Nilsson, D.-E., Smolka, J., and Bok, M. (2022). The vertical light-gradient and its potential impact on animal distribution and behaviour. Front. Ecol. Evol. 10:951328. doi: 10.3389/fevo.2022.951328
Ohly, K. P. (1975). The neurons of the first synaptic region of the optic neuropil of the firefly, Phausis splendidula l. (Coleoptera).. Cell Tissue Res. 158, 89–109. doi: 10.1007/BF00219953
Pettigrew, J. (1983). A note on the eyes of the letter-winged kite elanus scriptus. Emu Austral Ornithol. 82(sup1), 305–308.
Reber, T., Vähäkainu, A., Baird, E., Weckström, M., Warrant, E., and Dacke, M. (2015). Effect of light intensity on flight control and temporal properties of photoreceptors in bumblebees.. J. Exp. Biol. 218(Pt 9), 1339–1346. doi: 10.1242/jeb.113886
Reingruber, J., Pahlberg, J., Woodruff, M. L., Sampath, A. P., Fain, G. L., and Holcman, D. (2013). Detection of single photons by toad and mouse rods. Proc. Natl. Acad. Sci. U.S.A. 110, 19378–19383.
Ribelayga, C., Cao, Y., and Mangel, S. C. (2008). The circadian clock in the retina controls rod-cone coupling. Neuron 59, 790–801.
Ribi, W. A. (1977). Fine structure of the first optic ganglion (lamina) of the cockroach. Periplaneta americana., Tissue Cell 9, 57–72. doi: 10.1016/0040-8166(77)90049-0
Ribi, W. A. (1978). Ultrastructure and migration of screening pigments in the retina of Pieris rapae L. (Lepidoptera. Pieridae)., Cell Tissue Res. 191, 57–73. doi: 10.1007/BF00223215
Rivera-Alba, M., Vitaladevuni, S. N., Mishchenko, Y., Lu, Z., Takemura, S.-y, Scheffer, L., et al. (2011). Wiring economy and volume exclusion determine neuronal placement in the drosophila brain. Curr. Biol. 21, 2000–2005. doi: 10.1016/j.cub.2011.10.022
Rose, A. (1942). The relative sensitivities of television pickup tubes, photographic film, and the human eye. Proc. IRE 30, 293–300.
Roska, B., and Werblin, F. (2001). Vertical interactions across ten parallel, stacked representations in the mammalian retina. Nature 410, 583–587. doi: 10.1038/35069068
Schottky, W. (1918). Über spontane Stromschwankungen in verschiedenen Elektrizitätsleitern. Annalen der Physik 362, 541–567.
Snyder, A. (1979). “The physics of compound eyes,” in Handbook of senory physiology, ed. Autrum (Berlin: Springer), 225–213.
Snyder, A. W. (1977). Acuity of compound eyes: Physical limitations and design. J. Comp. Physiol. A. 116, 161–182. doi: 10.1007/BF00605401
Snyder, A. W., Stavenga, D. G., and Laughlin, S. B. (1977). Spatial information capacity of compound eyes. J. Comp. Physiol. 116, 183–207.
Srinivasan, M. V., Laughlin, S. B., and Dubs, A. (1982). Predictive coding: A fresh view of inhibition in the retina.. Proc. R. Soc. Lond. B Biol. Sci. 216, 427–459. doi: 10.1098/rspb.1982.0085
Srinivasan, M. V., Poteser, M., and Kral, K. (1999). Motion detection in insect orientation and navigation.. Vis. Res. 39, 2749–2766.
Stöckl, A. L., OCarroll, D. C., and Warrant, E. (2017). Higher-order neural processing tunes motion neurons to visual ecology in three species of hawkmoths. Proc. R. Soc. B. 284:20170880. doi: 10.1098/rspb.2017.0880
Stöckl, A. L., OCarroll, D. C., and Warrant, E. J. (2016a). Neural summation in the hawkmoth visual system extends the limits of vision in dim light.. Curr. Biol. 26, 821–826.
Stöckl, A., Heinze, S., Charalabidis, A., El Jundi, B., Warrant, E., and Kelber, A. (2016c). Differential investment in visual and olfactory brain areas reflects behavioural choices in hawk moths.. Sci. Rep. 6:26041. doi: 10.1038/srep26041
Stöckl, A. L., Ribi, W. A., and Warrant, E. J. (2016b). Adaptations for nocturnal and diurnal vision in the hawkmoth lamina.. J. Comp. Neurol. 524, 160–175. doi: 10.1002/cne.23832
Stöckl, A. L., OCarroll, D. C., and Warrant, E. J. (2020). Hawkmoth lamina monopolar cells act as dynamic spatial filters to optimize vision at different light levels. Sci. Adv. 6:eaaz8645. doi: 10.1126/sciadv.aaz8645
Stöckl, A., Smolka, J., OCarroll, D., and Warrant, E. (2017). Resolving the trade-off between visual sensitivity and spatial acuity - lessons from hawkmoths. Integr. Comp. Biol. 57, 1093–1103. doi: 10.1093/icb/icx058
Strausfeld, N. J., and Blest, A. D. (1970). The optic lobes of Lepidoptera.. Philos. Trans. R. Soc. Lond. B. 258, 81–134.
Tedore, C., and Nilsson, D.-E. (2019). Avian UV vision enhances leaf surface contrasts in forest environments. Nat. Commun. 10:238. doi: 10.1038/s41467-018-08142-5
Tian, H., Fowler, B., and Gamal, A. (2001). Analysis of temporal noise in CMOS photodiode active pixel sensor. IEEE J. Solid State Circ. 36, 92–101.
Tikidji-Hamburyan, A., Reinhard, K., Seitter, H., Hovhannisyan, A., Procyk, C. A., Allen, A. E., et al. (2015). Retinal output changes qualitatively with every change in ambient illuminance.. Nat. Neurosci. 18, 66–74. doi: 10.1038/nn.3891
Tinsley, J. N., Molodtsov, M. I., Prevedel, R., Wartmann, D., Espigulé-Pons, J., Lauwers, M., et al. (2016). Direct detection of a single photon by humans. Nat. Commun. 7:12172.
Tocco, C., Dacke, M., and Byrne, M. (2019). Eye and wing structure closely reflects the visual ecology of dung beetles. J. Comp. Physiol. A 205, 211–221. doi: 10.1007/s00359-019-01324-6
Torres, D., Tidau, S., Jenkins, S., and Davies, T. (2020). Artificial skyglow disrupts celestial migration at night. Curr. Biol. 30, R696–R697.
Troscianko, J., and Stevens, M. (2015). Image calibration and analysis toolbox a free software suite for objectively measuring reflectance, colour and pattern. Methods Ecol. Evol.S 6, 1320–1331. doi: 10.1111/2041-210X.12439
Umino, Y., Solessio, E., and Barlow, R. B. (2008). Speed, spatial, and temporal tuning of rod and cone vision in mouse.. J. Neurosci. 28, 189–198. doi: 10.1523/JNEUROSCI.3551-07.2008
van Hateren, J. H., and van der Schaaf, A. (1996). “Temporal properties of natural scenes,” in Proceedings of the SPIE, eds B. E. Rogowitz and J. P. Allebach (Bellingham: SPIE).
Vo-Doan, T. T., and Straw, A. D. (2020). Millisecond insect tracking system. arXiv [Preprint]. doi: 10.48550/arXiv.2002.12100
Wakefield, A., Broyles, M., Stone, E. L., Harris, S., and Jones, G. (2017). Quantifying the attractiveness of broad-spectrum street lights to aerial nocturnal insects. J. Appl. Ecol. 55, 714–722.
Wakefield, A., Stone, E. L., Jones, G., and Harris, S. (2015). Light-emitting diode street lights reduce last-ditch evasive manoeuvres by moths to bat echolocation calls. R. Soc. Open Sci. 2:150291. doi: 10.1098/rsos.150291
Walcott, B. (1975). “Anatomical changes during light-adaptation in insect compound eyes,” in The compound eye and vision of insects, ed. G. A. Horridge (Oxford: Clarendon Press), 20–33.
Walter, T., Degen, J., Pfeiffer, K., Stöckl, A., Montenegro, S., and Degen, T. (2021). A new innovative real-time tracking method for flying insects applicable under natural conditions. BMC Zool. 6:35. doi: 10.1186/s40850-021-00097-3
Walton, R. E., Sayer, C. D., Bennion, H., and Axmacher, J. C. (2020). Nocturnal pollinators strongly contribute to pollen transport of wild flowers in an agricultural landscape. Biol. Lett. 16:20190877. doi: 10.1098/rsbl.2019.0877
Warrant, E. J. (1999). Seeing better at night: Life style, eye design and the optimum strategy of spatial and temporal summation.. Vis. Res. 39, 1611–1630. doi: 10.1016/s0042-6989(98)00262-4
Warrant, E. J. (2008). Seeing in the dark: Vision and visual behaviour in nocturnal bees and wasps.. J. Exp. Biol. 211(Pt 11), 1737–1746.
Warrant, E. J., and Johnsen, S. (2013). Vision and the light environment. Curr. Biol. 23, R990–R994.
Warrant, E. J., and McIntyre, P. D. (1990). Limitations to resolution in superposition eyes. J. Comp. Physiol. A. 167, 785–803. doi: 10.1007/BF00189768
Warrant, E. J., and McIntyre, P. D. (1991). ‘Strategies for retinal design in arthropod eyes of low f-number’. J. Comp. Physiol. A 168, 499–512. doi: 10.1007/BF00199610
Warrant, E. J., and McIntyre, P. D. (1996). The visual ecology of pupillary action in superposition eyes. J. Comp. Physiol. A. 178, 75–90. doi: 10.1007/BF00189592
Warrant, E. J., and Nilsson, D. E. (1998). Absorption of white light in photoreceptors.. Vis. Res. 38, 195–207.
Warrant, E., Johnsen, S., and Nilsson, D.-E. (2020). “1.02 - light and visual environments,” in The senses: A comprehensive reference, 2nd Edn, ed. B. Fritzsch (Amsterdam: Elsevier), 4–30. doi: 10.1016/B978-0-12-805408-6.00002-6
Warrant, E. J., Kelber, A., Gislén, A., Greiner, B., Ribi, W., and Wcislo, W. T. (2004). Nocturnal vision and landmark orientation in a tropical halictid bee.. Curr. Biol. 14, 1309–1318. doi: 10.1016/j.cub.2004.07.057
Wcislo, W. T., Arneson, L., Roesch, K., Gonzalez, V., Smith, A., and Fernández, H. (2004). The evolution of nocturnal behaviour in sweat bees, Megalopta genalis and M. ecuadoria (Hymenoptera: Halictidae): An escape from competitors and enemies? Biol. J. Linn. Soc. 83, 377–387. doi: 10.1111/j.1095-8312.2004.00399.x
Wienbar, S., and Schwartz, G. W. (2018). The dynamic receptive fields of retinal ganglion cells. Prog. Retinal Eye Res. 67, 102–117.
Wolburg-Buchholz, K. (1979). The organization of the lamina ganglionaris of the hemipteran insects, Notonecta glauca, Corixa punctata and Gerris lacustris. Cell Tissue Res. 197, 39–59. doi: 10.1007/BF00233552
Wystrach, A., Beugnon, G., and Cheng, K. (2012). Ants might use different view-matching strategies on and off the route. J. Exp. Biol. 215, 44–55.
Yadid-Pecht, O., Mansoorian, K., Fossum, E. R., and Pain, B. (1997). “Optimization of noise and responsivity in CMOS active pixel sensors for detection of ultralow-light levels,” in Solid state sensor arrays: Development and applications, ed. M. M. Blouke (Bellingham: SPIE).
Yeandle, S. (1958). Evidence of quantized slow potentials in the eye of Limulus. Am. J. Ophthalmol. 46, 82–87.
Keywords: natural visual scenes, all-sky imaging, dim light vision, light pollution, ALAN, hawkmoth, mouse, fruitfly
Citation: Stöckl AL and Foster JJ (2022) Night skies through animals’ eyes—Quantifying night-time visual scenes and light pollution as viewed by animals. Front. Cell. Neurosci. 16:984282. doi: 10.3389/fncel.2022.984282
Received: 01 July 2022; Accepted: 09 September 2022;
Published: 06 October 2022.
Edited by:
Marcus H. C. Howlett, Netherlands Institute for Neuroscience (KNAW), NetherlandsReviewed by:
Jonathan Bennie, University of Exeter, United KingdomMichael Manookin, University of Washington, United States
Copyright © 2022 Stöckl and Foster. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Anna Lisa Stöckl, anna.stoeckl@uni-konstanz.de