- 1Research Centre for Language, Cognition and Language Application, Chongqing University, Chongqing, China
- 2School of Foreign Languages and Cultures, Chongqing University, Chongqing, China
- 3School of Education and Languages, Hong Kong Metropolitan University, Hong Kong, Hong Kong SAR, China
The present “visual world” eye-tracking study examined the time-course of how native and non-native speakers keep track of implied object-state representations during real-time language processing. Fifty-two native speakers of English and 46 non-native speakers with advanced English proficiency joined this study. They heard short stories describing a target object (e.g., an onion) either having undergone a substantial change-of-state (e.g., chop the onion) or a minimal change-of-state (e.g., weigh the onion) while their eye movements toward competing object-states (e.g., a chopped onion vs. an intact onion) and two unrelated distractors were tracked. We found that both groups successfully directed their visual attention toward the end-state of the target object that was implied in the linguistic context. However, neither group showed anticipatory eye movements toward the implied object-state when hearing the critical verb (e.g., “weigh/chop”). Only native English speakers but not non-native speakers showed a bias in visual attention during the determiner (“the”) before the noun (e.g., “onion”). Our results suggested that although native and non-native speakers of English largely overlapped in their time-courses of keeping track of object-state representations during real-time language comprehension, non-native speakers showed a short delay in updating the implied object-state representations.
Introduction
There is extensive evidence that native speakers anticipate what comes next in language comprehension (Altmann and Mirković, 2009; Kuperberg and Jaeger, 2015). For example, Altmann and Kamide (1999) found that visual attention was directed to the target object before it was explicitly mentioned in the language. However, as most of the available studies focused on native speakers, it remains debated whether non-native speakers anticipate upcoming information in language comprehension to the same extent as native speakers.
Existing studies on non-native speakers have primarily focused on the use of morphosyntactic features and grammatical knowledge during language comprehension, such as gender (Lew-Williams and Fernald, 2010; Dussias et al., 2013; Hopp, 2013; Bañón and Martin, 2021), syntactic or semantic ambiguity (Frenck-Mestre and Pynte, 1997; Wilson and Garnsey, 2009; Dussias et al., 2010), and phonological forms (DeLong et al., 2005; Martin et al., 2013). Several studies have revealed that non-native speakers were not as quick or as accurate as native speakers in making predictions (Kaan et al., 2010; Lew-Williams and Fernald, 2010; Grüter et al., 2012; Martin et al., 2013; Kaan, 2014). But other studies observed native-like predictive processing in non-native speakers (Dahan et al., 2000; Dussias et al., 2013; Hopp, 2013; Foucart et al., 2014; Trenkic et al., 2014). The differences between native and non-native language comprehension are often attributed to factors such as complexity of linguistic subdomains (Clahsen and Felser, 2006) and variability in non-native speakers’ proficiency of and exposure to the target language (Dussias et al., 2013; Kaan, 2014; Hopp and Lemmerth, 2016; Li et al., 2020).
Nonetheless, these studies have not considered the recruitment of non-linguistic information in language comprehension. According to mental/situation models (e.g., Johnson-Laird, 1983; van Dijk and Kintsch, 1983) and perceptual symbol systems (Barsalou, 1999, 2008), language comprehension involves not only the activation of linguistic knowledge but also situations and mental representations grounded in sensorimotor experiences (but see Mahon and Caramazza, 2008). For example, Zwaan and colleagues showed that language comprehenders were faster to verify pictures that matched the implied orientation (Stanfield and Zwaan, 2001), shape (Zwaan et al., 2002), visibility (Yaxley and Zwaan, 2007) than pictures that mismatched (see also Taylor and Zwaan, 2009; Horchak et al., 2014). Marino et al. (2014) revealed that viewing photos and reading nouns of natural graspable objects modulated motor responses. In addition, previous studies demonstrated that toddlers with low reading skills and limited use of language activate mental representations of objects in language comprehension, suggesting that the recruitment of non-linguistic information might not be dependent on the proficiency of language (e.g., Engelen et al., 2011; Johnson and Huettig, 2011; Bobb et al., 2016).
However, compared with the number of studies on the role of non-linguistic information in native language processing, there were fewer studies on the role of non-linguistic information in the case of a non-native language (Kühne and Gianelli, 2019). Some studies support the idea that during the processing of non-native language, non-linguistic information is activated (Kogan et al., 2020). For example, Dudschig et al. (2014) revealed that bilinguals activated motor responses when they processed action and emotion words in their non-native language. Buccino et al. (2017) showed that fluent speakers of a second language processed graspable nouns in a second language like in their native language. Parker Jones et al. (2012) revealed that bilinguals and monolinguals differed in brain activation during picture naming and reading aloud. De Grauwe et al. (2014) found that non-native speakers activated motor and somatosensory brain areas when they were presented motor verbs in the non-native language like native speakers. Nonetheless, there is limited evidence on the timing of activating non-linguistic information in language comprehension by native and non-native speakers.
In the present study, we examined the activation of mental representations of objects in real-time language processing by native and non-native speakers of English. Specifically, we investigated to what extent native and non-native speakers of English overlapped in their time courses of keeping track of object-states as language unfolded. According to the “intersecting object histories” (IOH) hypothesis, dynamic changes in objects across time are used as primitives of event representations (Altmann and Ekves, 2019). Multiple representations of objects are activated and updated during language processing (e.g., Hindy et al., 2012, 2015; Solomon et al., 2015; Kang et al., 2019, 2020; Horchak and Garrido, 2020; Hupp et al., 2020; Lee and Kaiser, 2021; Misersky et al., 2021; Santin et al., 2021). Kang et al. (2020) revealed that native speakers of English shifted their eye movements between two competing object-state representations of the target object in real-time language comprehension. In their study, participants were asked to listen to short stories in 2 × 2 conditions, such as “The chef will chop/weigh the onion. But first/And then, he will smell the onion” while viewing a visual stimulus showing two competing states of the target object (e.g., a chopped onion vs. an intact onion) and two distractors. They found that participants preferred to look at the changed object-state (e.g., a chopped onion) when it matched the implied end-state of the target object compared to when it mismatched the implied end-state in the first sentence (e.g., chop vs. weigh the onion). Interestingly, the bias of visual attention occurred at the end of the second sentence when the target object was explicitly mentioned.
In the present study, we tested two competing hypotheses. One hypothesis is that non-native speakers should be as quick and as accurate in activating and updating object-state representations as native speakers in real-time language processing since the construction of event representation is not subject to how good one is at understanding or using the language (e.g., Bobb et al., 2016). An alternative hypothesis is that non-native speakers and native speakers show differences in keeping track of object-state representations supported by cross-linguistic differences in event categorization and perception (e.g., Papafragou et al., 2006; Brown and Gullberg, 2010; Papafragou and Selimis, 2010; Flecken et al., 2015; Aveledo and Athanasopoulos, 2016).
We opted to test these hypotheses by using the visual world paradigm that has been used in previous studies on real-time language processing (Tanenhaus et al., 1995). In this paradigm, participants are instructed to view or manipulate objects in the “visual world” (either in real-world or on a computer screen) while their eye movements toward these objects are recorded as they listen to short stories that describe events related to these objects. We expect that if native and non-native speakers keep track of event representations to the same extent during real-time language processing, they should have the same time courses of directing their visual attention toward the implied object-state as the language unfolds.
Method
Participants
A 52 native speakers and 46 non-native speakers of English participated in this study. None of them reported impairment in vision or hearing. Non-native speakers of English were native speakers of Cantonese and were studying at a research university in Hong Kong where English was used as the instruction language. All participants signed written informed consent before joining this study and received cash compensation for their participation. Table 1 presents the demographics of participants. The sample size was determined based on a previous study (Kang et al., 2020). Compared to the previous study, the present study has fewer conditions (2 vs. 4) but more trials per condition than the previous study (12 vs. 9). We performed a power simulation using simr package (Green and MacLeod, 2016). Simulation results showed that with 45 participants and 24 trials the statistical power for Degree of Change was 80%.
Materials
We constructed 24 pairs of linguistic stimuli that described either a minimal or a substantial change-of-state event. Each stimulus contained four sentences. The first three sentences set up the context of the story. The fourth sentence was the critical sentence that described either a substantial change-of-state or a minimal change-of-state (“The rabbit is weighing/chopping the onion), followed by a negative clause (e.g., “not smelling the onion”). For example:
(A) Minimal Change-of-State Event (ME): The rabbit has a bowl, a bottle of pills, and an onion. She was going to smell the onion. Then she changed her mind. The rabbit is weighing the onion, not smelling the onion.
(B) Substantial Change-of-State Event (SE): The rabbit has a bowl, a bottle of pills, and an onion. She was going to smell the onion. Then she changed her mind. The rabbit is chopping the onion, not smelling the onion.
The linguistic stimuli were recorded in a soundproof booth by a male native speaker of British English at 44.1k Hz sampling rates with 16 bits resolution. Each stimulus was scaled to 70 dB SPL in mean intensity using Praat (Version 6.0.39; Boersma and Weenink, 2018). Each pair of linguistic stimuli was associated with a visual stimulus that depicted the protagonist with four objects using clipart images (Figure 1). The locations of the objects were counter-balanced across visual stimuli.
Figure 1. Example visual stimulus. Participants heard sentences such as “The rabbit has a bowl, a bottle of pills and an onion. She was going to smell the onion. Then she changed her mind. The rabbit is weighing/chopping the onion, not smelling the onion”.
Procedure
Two counter-balanced lists were created for the experiment. Each list consisted of 24 experimental trials and 20 filler trials. Half of the experimental trials described a minimal change of the object-state (e.g., “weighing the onion”) [ME condition, as in (A)] and the other half a substantial change of the object-state (e.g., “chopping the onion”) [SE condition, as in (B)]. Filler trials followed the same structure as the experimental trials. The trials were presented in a pseudorandomized order. In each trial, participants viewed the visual display and heard the auditory stimuli simultaneously. Eye movements on the visual stimulus were tracked during the experiment. The total time of the experiment was about 20 min.
Tobii TX300 was used to collect eye movement data. The sampling rate was 300 Hz from both eyes. Freedom of movement was 37 × 17 cm at a 65 cm distance and gaze accuracy was 0.47 degrees. Tobii Studio was used to display the stimuli and collect the data. Experimental trials in which eye movements could not reliably be tracked were excluded from the analyses. This resulted in the exclusion of 9.4% of all trials (3.6% for native English speakers, 5.8% for non-native English speakers).
Data Processing
All the participants achieved above 90% of gaze samples (calculated by dividing the number of eye-tracking samples that were correctly identified by the number of attempts) with a mean percentage of 95.26%, indicating that they were consistently looking at the visual stimuli during the experiment. For each participant, we exported the raw eye gaze data (timestamp and gaze tracking data) using Tobii Pro Studio software.
Our analyses focused on language-mediated visual attention. Raw eye-tracking data were aggregated into proportions of fixations first by-subjects and then by-items for nine critical time windows, on a trial-by-trial basis, in the linguistic stimuli (see an example in Table 2). We conducted statistical analyses during the time window spanning from the onset of a critical time window in the linguistic stimulus (e.g., onion) +200 ms until its offset +200 ms. We selected the time window of a critical word +200 ms since previous studies have demonstrated that the competition effects of related objects were observed around 200–300 ms after the onset of the target word (e.g., Huettig and Altmann, 2005; Yee and Sedivy, 2006).
We transformed the proportion of fixations for each time window using the arcsine square root transformation to account for the bounded nature of binomial responses (e.g., Williams et al., 2019). We then fit linear mixed models for data of each time window using the lmer function in the lme4 package (Bates et al., 2015) of R (R Core Team, 2020). We assigned sum-coded contrasts to Degree of Change (minimal change = −1; substantial change = 1) and Object-state (intact-state = −1; changed-state = 1).
In the 1st time window, we included Object-state as a fixed effect, participants as random effects in the by-subject model, and items as random effects in the by-item model. In the 2nd–9th time windows, we included Degree of Change, Object-state, their interaction as fixed effects, participants as random effects in by-subject models, and items as random effects in by-item models.1 See example models below:
By-subject<-lmer(Trans_Prop∼Degree-of-Change*Object -state + (1| Subject), data = T2)
By-item<-lmer(Trans_Prop ∼ Degree-of-Change*Object-state + (1| Item), data = T2)
We did not fit maximal models due to convergence problems across more complex models in later time windows. To assess the goodness of fit, we compared the models using the χ2-distributed likelihood ratio and its associated p-value. The model with a smaller Akaike Information Criterion (AIC) and the Bayesian Information Criterion (BIC) was considered as a better fit (Baayen et al., 2008). Only effects that were significant in both by-subject and by-item analyses were accepted as significant. Significant interaction effects between fixed effects were followed by pairwise comparisons with “tukey” adjustment for multiple comparisons using emmeans package (Lenth, 2018).
Results
Figure 2 presents the percentage of trials with fixations on the competing object-states as language unfolded. Table 2 presents results of statistical analyses during 9 critical time windows.
Figure 2. Percentage of trials with fixations launched on the interest areas (AOIs) across sentential conditions. (A) Fixations of native speakers. (B) Fixations of non-native speakers. The x-axis shows the elapsed time increments from the onset of linguistic stimuli (e.g., “The rabbit is weighing/chopping the onion, not smelling the onion”). The y-axis shows percentage of trials with at least one fixation on the AOIs. Standard errors above and below the mean were shown as shaded areas. The dashed lines indicate the offset of critical time windows.
Native Speakers of English
During the 1st time window, native speakers of English showed no differences in their proportions of fixations to the intact state and the changed state. During the 2nd time window when the critical verbs (e.g., weighing/chopping) were mentioned, there was not yet an interaction effect between Object-state and Degree of Change. The first significant interaction effect was found in the 3rd (“the”) and in all following time windows. Pairwise comparisons suggested that after hearing a substantial change (e.g., “chopping”) (SE condition) than a minimal change (e.g., “weighing”) (ME condition) participants initiated higher proportions of fixations toward the changed-state (e.g., a chopped onion) from the 3rd time window to the 9th time window [By-Subjects: p < 0.001 (3rd–7th), p = 0.006 (8th), p = 0.046 (9th); By-Items: p = 0.001 (3rd), p < 0.001 (4th, 5th), p = 0.002 (6th), p = 0.036 (7th), p = 0.003 (8th), p = 0.039 (9th)]. No such differences were found in the proportions of fixations toward the intact state. Thus, despite directing their visual attention toward the changed state of the target object, native English speakers did not show anticipatory eye movements when they just heard the critical verb (e.g., chop vs. weigh). The earliest time window revealing such differences in visual attention was during the determiner (“the”) right after the critical verb.
Non-native Speakers of English
Similar to native speakers, non-native speakers did not show any differences in eye movements between the intact-state and the changed-state in the 1st time window. There was no interaction effect between Object-state and Degree of Change in the 2nd time window either. However, unlike native speakers, non-native speakers showed no interaction effect between Object-state and Degree of Change in the 3rd time window (“the”). The first interaction effect was found during the 4th (e.g., “onion”) and in all following time windows. Pairwise comparisons suggested that there were higher proportions of fixations to the changed-state after a substantial change (e.g., “chopping”) (SE condition) than a minimal change (e.g., “weighing”) (ME condition) was described [By-Subjects: p < 0.001 (3rd–9th); By-Items: p = 0.016 (3rd), p = 0.007 (4th), p < 0.001 (5th, 6th), p = 0.003 (7th, 8th), p = 0.002 (9th)]. By contrast, no such differences in visual attention were found on the intact state. Thus, compared with native speakers non-native speakers showed a short delay in linguistically mediated visual attention toward the implied end-state of the target object.
Discussion
The present study investigated how language comprehenders keep track of implied object-states during real-time language processing. We revealed that both native and non-native speakers of English speakers activated and updated object-states in real-time language comprehension. Both groups did not show any anticipatory eye movements at the verb region (e.g., “chopping/weighing”), but directed visual attention to the end-state of the target object when they heard the object name (e.g., “onion”). In principle, participants could have moved their eye movements toward the expected end-state of the target object as soon as the critical verb was heard. One possibility for the lack of anticipatory eye movement during the verb region is that the competing object-states of the target object on the visual display cannot be integrated with the linguistic context until the specific cue (e.g., “the onion”) is provided. Anticipatory eye movements on the visual scene may reflect the integration of linguistic, visual, and world knowledge (Smith and Levy, 2013; Nieuwland et al., 2020). Participants may not be motivated to look to one or the other depiction of the target object as a specific token of the target object until the object name was directly referred to.
It is also possible that the intact-state and the changed-state may be two discrete episodic tokens of the target object on the continuum of trajectories in event representations. Preferential looks to object-states may reflect a featural overlap between the visual depiction and mental representations of the target object. In this process, participants may have to go through a multi-step process, in which they first activate the initial state of the target object that affords for the action before activating its intermediate states and the end-state. Only after the verb is specified, they are then able to update mental models of the change-of-state event and thus direct their visual attention to the end-state. A similar hypothesis, known as the two-step hypothesis, was proposed to understand “negation” (e.g., “The door is not closed”). According to the two-step hypothesis, we have to first activate the state of affairs before the negation (a closed door) and then the negated state (an open door) (e.g., Kaup and Zwaan, 2003; Kaup et al., 2006; Lüdtke et al., 2008). We postulate that something analogous might be going on when we keep track of object-states in language processing.
However, despite these similarities, non-native speakers showed differences from native speakers in the time course of activating the implied object-state. Only native speakers but not non-native speakers of English directed their visual attention to the end-state of the target object during the determiner region (“the”). Our results thus support the alternative hypothesis that native and non-native speakers showed differences in activating mental representations of objects during real-time language processing.
This short delay of visual attention toward the implied object-state among non-native speakers could be accounted for by the Reduced Ability to Generate Expectations (RAGE) account (Grüter et al., 2017). According to the RAGE account, even advanced non-native speakers were less likely to rely on predictive mechanisms at the discourse level to the same extent as native speakers (see also Kaan et al., 2010; Kaan, 2014). Therefore, non-native speakers may not be able to show the pre-nominal prediction effect (Fleur et al., 2020; Bañón and Martin, 2021), thus they have to launch anticipatory eye movements toward the implied object-state when the object was explicitly mentioned.
However, we could not exclude the possibility that morphosyntactic differences between the L1 (Cantonese) and the L2 (English) of non-native speakers might lead to this delay. Cantonese and English are typologically divergent and genetically unrelated languages (Matthews and Yip, 2011). The change-of-state events were coded differently in Cantonese and English. For example, in English, the verb “break” indicates both the action and the consequences, but in Cantonese, they have to be specified separately using the serial verb construction (Francis and Matthews, 2006). Another difference between Cantonese and English is that there is no determiner such as “the” in Cantonese, but classifiers are used before nouns (Chow and Chen, 2020). Thus, further studies may examine whether these morphosyntax differences in L1 and L2 will slow down non-native speakers’ activation of mental representations of event knowledge in real-time language comprehension.
In conclusion, our study demonstrated that both native and non-native speakers of English kept track of object-state representations in real-time language comprehension. They all directed their visual attention toward the end-state of the target object when the object name was directly referred to, but no anticipatory eye movements were found during the verb region. Nonetheless, native speakers but not non-native speakers showed anticipatory eye movements during the determiner (“the”). Such similarities between native and non-native speakers in real-time language processing indicate that non-native speakers do not differ significantly from native speakers in how predictive mechanisms are employed for event representations in real-time processing. Our study provides empirical evidence that native-like processing of event knowledge is possible among non-native speakers during real-time language comprehension but a short delay can be observed.
Data Availability Statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://osf.io/9vb7s/.
Ethics Statement
The studies involving human participants were reviewed and approved by the Ethics Committee of the Chinese University of Hong Kong. The patients/participants provided their written informed consent to participate in this study.
Author Contributions
XK: conceptualization, design of sentence stimuli, design of picture stimuli, analysis and interpretation of data, and draft and revision of the manuscript. HG: design of picture stimuli, data acquisition, and draft and revision of the manuscript. Both authors contributed to the article and approved the submitted version.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We would like to thank James Britton, Hannah Lam, and Christy Yuen for their assistance with various aspects of the study. We also gratefully acknowledge Patrick C. M. Wong and Virginia Yip for providing experimental facilities for data collection.
Footnotes
- ^ We also analyzed the data using linear mixed models by including Object-state, Degree of Change, and their interactions as fixed effects, and both participants and items as random effects by following Kang et al. (2020). During the 3rd time window, there was a significant interaction among native speakers (χ2 = 5.70, p = 0.017), but not among non-native speakers (χ2 = 0.037, p = 0.848).
References
Altmann, G. T. M., and Ekves, Z. (2019). Events as intersecting object histories: a new theory of event representation. Psychol. Rev. 126, 817–840. doi: 10.1037/rev0000154
Altmann, G. T. M., and Kamide, Y. (1999). Incremental interpretation at verbs: restricting the domain of subsequent reference. Cognition 73, 247–264. doi: 10.1016/s0010-0277(99)00059-1
Altmann, G. T. M., and Mirković, J. (2009). Incrementality and prediction in human sentence processing. Cogn. Sci. 33, 583–609. doi: 10.1111/j.1551-6709.2009.01022.x
Aveledo, F., and Athanasopoulos, P. (2016). Second language influence on first language motion event encoding and categorization in Spanish-speaking children learning L2 English. Int. J. Biling. 20, 403–420.
Baayen, R. H., Davidson, D. J., and Bates, D. M. (2008). Mixed-effects modeling with crossed random effects for subjects and items. J. Mem. Lang. 59, 390–412. doi: 10.1080/00273171.2021.1889946
Bañón, J. A., and Martin, C. (2021). The role of crosslinguistic differences in second language anticipatory processing: an event-related potentials study. Neuropsychologia 155:107797. doi: 10.1016/j.neuropsychologia.2021.107797
Bates, D., Mächler, M., Bolker, B., and Walker, S. (2015). Fitting linear mixed-effects models using lme4. J. Stat. Softw. 67, 1–48.
Bobb, S. C., Huettig, F., and Mani, N. (2016). Predicting visual information during sentence processing: toddlers activate an object’s shape before it is mentioned. J. Exp. Child Psychol. 151, 51–64. doi: 10.1016/j.jecp.2015.11.002
Boersma, P., and Weenink, D. (2018). Praat: Doing Phonetics by Computer [Computer software]. Version 6.0.39.
Brown, A., and Gullberg, M. (2010). Changes in encoding of path of motion in a first language during acquisition of a second language. Cong. Linguist. 21, 263–286.
Buccino, G., Marino, B. F., Bulgarelli, C., and Mezzadri, M. (2017). Fluent speakers of a second language process graspable nouns expressed in L2 Like in their native language. Front. Psychol. 8:1306. doi: 10.3389/fpsyg.2017.01306
Chow, W. Y., and Chen, D. (2020). Predicting (in)correctly: listeners rapidly use unexpected information to revise their predictions. Lang. Cogn. Neurosci. 35, 1149–1161.
Clahsen, H., and Felser, C. (2006). How native-like is non-native language processing? Trends Cogn. Sci. 10, 564–570. doi: 10.1016/j.tics.2006.10.002
Dahan, D., Swingley, D., Tanenhaus, M. K., and Magnuson, J. S. (2000). Linguistic gender and spoken-word recognition in French. J. Mem. Lang. 42, 465–480. doi: 10.1006/jmla.1999.2688
De Grauwe, S., Willems, R. M., Rueschemeyer, S.-A., Lemhöfer, K., and Schriefers, H. (2014). Embodied language in first- and second-language speakers: neural correlates of processing motor verbs. Neuropsychologia 56, 334–349. doi: 10.1016/j.neuropsychologia.2014.02.003
DeLong, K. A., Urbach, T. P., and Kutas, M. (2005). Probabilistic word pre-activation during language comprehension inferred from electrical brain activity. Nat. Neurosci. 8, 1117–1121. doi: 10.1038/nn1504
Dudschig, C., de la Vega, I., and Kaup, B. (2014). Embodiment and second-language: automatic activation of motor responses during processing spatially associated L2 words and emotion L2 words in a vertical Stroop paradigm. Brain Lang. 132, 14–21. doi: 10.1016/j.bandl.2014.02.002
Dussias, P. E., Marful, A., Gerfen, C., and Molina, M. T. B. (2010). Usage frequencies of complement-taking verbs in Spanish and english: data from Spanish monolinguals and Spanish—English bilinguals. Behav. Res. Methods 42, 1004–1011. doi: 10.3758/BRM.42.4.1004
Dussias, P. E., Valdés Kroff, J. R., Guzzardo Tamargo, R. E., and Gerfen, C. (2013). When gender and looking go hand in hand: grammatical gender processing in non-native Spanish. Stud. Sec. Lang. Acquisition 35, 353–387.
Engelen, J. A. A., Bouwmeester, S., de Bruin, A. B. H., and Zwaan, R. A. (2011). Perceptual simulation in developing language comprehension. J. Exp. Child Psychol. 110, 659–675. doi: 10.1016/j.jecp.2011.06.009
Flecken, M., Carroll, M., Weimar, K., and Von Stutterheim, C. (2015). Driving along the road or heading for the Village? conceptual differences underlying motion event encoding in French, German, and French-German L2 users. Modern Lang. J. 99, 100–122.
Fleur, D. S., Flecken, M., Rommers, J., and Nieuwland, M. S. (2020). Definitely saw it coming? the dual nature of the pre-nominal prediction effect. Cognition 204:104335. doi: 10.1016/j.cognition.2020.104335
Foucart, A., Martin, C. D., Moreno, E. M., and Costa, A. (2014). Can bilinguals see it coming? word anticipation in L2 sentence reading. J. Exp. Psychol. Learn. Mem. Cogn. 40, 1461–1469. doi: 10.1037/a0036756
Francis, E. J., and Matthews, S. (2006). Categoriality and object extraction in cantonese serial verb constructions. Nat. Lang. Linguist. Theory 24, 751–801.
Frenck-Mestre, C., and Pynte, J. (1997). Syntactic ambiguity resolution while reading in second and native languages. Quart. J. Exp. Psychol. A 50, 119–148. doi: 10.1016/j.actpsy.2007.09.004
Green, P., and MacLeod, C. J. (2016). SIMR: an R package for power analysis of generalised linear mixed models by simulation. Methods Ecol. Evol. 7, 493–498. doi: 10.1111/2041-210x.12504
Grüter, T., Lew-Williams, C., and Fernald, A. (2012). Grammatical gender in non-native: a production or a real-time processing problem? Sec. Lang. Res. 28, 191–215. doi: 10.1177/0267658312437990
Grüter, T., Rohde, H., and Schafer, A. J. (2017). Coreference and discourse coherence in L2: the roles of grammatical aspect and referential form. Linguist. Approaches Biling. 7, 199–229. doi: 10.1075/lab.15011.gru
Hindy, N. C., Altmann, G. T. M., Kalenik, E., and Thompson-Schill, S. L. (2012). The effect of object state-changes on event processing: do objects compete with themselves? J. Neurosci. 32, 5795–5803. doi: 10.1523/JNEUROSCI.6294-11.2012
Hindy, N. C., Solomon, S. H., Altmann, G. T. M., and Thompson-Schill, S. L. (2015). A cortical network for the encoding of object change. Cereb. Cortex 25, 884–894. doi: 10.1093/cercor/bht275
Hopp, H. (2013). Grammatical gender in adult non-native acquisition: relations between lexical and syntactic variability. Sec. Lang. Res. 29, 33–56. doi: 10.1177/0267658312461803
Hopp, H., and Lemmerth, N. (2016). Lexical and syntactic ongrency in L2 predictive gender processing. Stud. Sec. Lang. Acquisition 40, 171–199 doi: 10.1017/s0272263116000437
Horchak, O. V., and Garrido, M. V. (2020). Dropping bowling balls on tomatoes: representations of object state-changes during sentence processing. J. Exp. Psychol. Learn. Mem. Cogn. 47, 838–857. doi: 10.1037/xlm0000980
Horchak, O. V., Giger, J.-C., Cabral, M., and Pochwatko, G. (2014). From demonstration to theory in embodied language comprehension: a review. Cogn. Syst. Res. 2, 66–85. doi: 10.1016/j.actpsy.2010.09.008
Huettig, F., and Altmann, G. T. M. (2005). Word meaning and the control of eye fixation: semantic competitor effects and the visual world paradigm. Cognition 96, B23–B32. doi: 10.1016/j.cognition.2004.10.003
Hupp, J. M., Jungers, M. K., Porter, B. L., and Plunkett, B. A. (2020). The implied shape of an object in adults’ and children’s visual representations. J. Cogn. Dev. 21, 368–382. doi: 10.1016/s0387-7604(03)00064-0
Johnson, E. K., and Huettig, F. (2011). Eye movements during language-mediated visual search reveal a strong link between overt visual attention and lexical processing in 36-month-olds. Psychol. Res. 75, 35–42. doi: 10.1007/s00426-010-0285-4
Kaan, E. (2014). Predictive sentence processing in non-native and native: what is different? Linguist. Approaches Biling. 4, 257–282.
Kaan, E., Dallas, A. C., and Wijnen, F. (2010). “Syntactic predictions in second-language sentence processing,” in Structure Preserved. Festschrift in the Honor of Jan Koster, eds J.-W. Zwart and M. de Vries (Amsterdam: John Benjamins).
Kang, X., Eerland, A., Joergensen, G. H., Zwaan, R. A., and Altmann, G. T. M. (2019). The influence of state change on object representations in language comprehension. Mem. Cogn. 48, 390–399. doi: 10.3758/s13421-019-00977-7
Kang, X., Joergensen, G. H., and Altmann, G. T. M. (2020). The activation of object-state representations during online language comprehension. Acta Psychol. 210:103162. doi: 10.1016/j.actpsy.2020.103162
Kaup, B., and Zwaan, R. A. (2003). Effects of negation and situational presence on the accessibility of text information. J. Exp. Psychol. Learn. Mem. Cogn. 29, 439–446. doi: 10.1037/0278-7393.29.3.439
Kaup, B., Lüdtke, J., and Zwaan, R. A. (2006). Processing negated sentences with contradictory predicates: is a door that is not open mentally closed? J. Pragmatics 38, 1033–1050.
Kogan, B., Muñoz, E., Ibáñez, A., and García, A. M. (2020). Too late to be grounded? motor resonance for action words acquired after middle childhood. Brain Cogn. 138:105509. doi: 10.1016/j.bandc.2019.105509
Kühne, K., and Gianelli, C. (2019). Is embodied cognition bilingual? current evidence and perspectives of the embodied cognition approach to bilingual language processing. Front. Psychol. 10:108. doi: 10.3389/fpsyg.2019.00108
Kuperberg, G. R., and Jaeger, T. F. (2015). What do we mean by prediction in language comprehension? Lang. Cogn. Neurosci. 31, 32–59. doi: 10.1080/23273798.2015.1102299
Lee, S. H.-Y., and Kaiser, E. (2021). Does hitting the window break it?: investigating effects of discourse-level and verb-level information in guiding object state representations. Lang. Cogn. Neurosci. 36, 921–940.
Lenth, R. (2018). emmeans: Estimated Marginal Means, aka Least-Squares Means. R package version 1.1.3.
Lew-Williams, C., and Fernald, A. (2010). Real-time processing of gender-marked articles by native and non-native Spanish speakers. J. Mem. Lang. 63, 447–464. doi: 10.1016/j.jml.2010.07.003
Li, X., Ren, G., Zheng, Y., and Chen, Y. (2020). How does dialectal experience modulate anticipatory speech processing? J. Mem. Lang. 115:104169. doi: 10.1111/psyp.13196
Lüdtke, J., Friedrich, C. K., De Filippis, M., and Kaup, B. (2008). Event-related potential correlates of negation in a sentence–picture verification paradigm. J. Cogn. Neurosci. 20, 1355–1370. doi: 10.1162/jocn.2008.20093
Mahon, B. Z., and Caramazza, A. (2008). A critical look at the embodied cognition hypothesis and a new proposal for grounding conceptual content. J. Physiol. Paris 102, 59–70. doi: 10.1016/j.jphysparis.2008.03.004
Marino, B. F. M., Sirianni, M., Volta, R. D., Magliocco, F., Silipo, F., Quattrone, A., et al. (2014). Viewing photos and reading nouns of natural graspable objects similarly modulate motor responses. Front. Hum. Neurosci. 8:968.
Martin, C., Thierry, G., Kuipers, J.-R., Boutonnet, B., Foucart, A., and Costa, A. (2013). Bilinguals reading in their second language do not predict upcoming words as native readers do. J. Mem. Lang. 69, 574–588.
Misersky, J., Slivac, K., Hagoort, P., and Flecken, M. (2021). The state of the onion: grammatical aspect modulates object representation during event comprehension. Cognition 214:104744. doi: 10.1016/j.cognition.2021.104744
Nieuwland, M. S., Barr, D. J., Bartolozzi, F., Busch-Moreno, S., Darley, E., Donaldson, D. I., et al. (2020). Dissociable effects of prediction and integration during language comprehension: evidence from a large-scale study using brain potentials. Philos. Trans. R. Soc. Lond. B Biol. Sci. 375, 20180522. doi: 10.1098/rstb.2018.0522
Papafragou, A., and Selimis, S. (2010). Lexical and structural biases in the acquisition of motion verbs. Lang. Learn. Dev. 6, 87–115.
Papafragou, A., Massey, C., and Gleitman, L. (2006). When English proposes what greek presupposes: the cross-linguistic encoding of motion events. Cognition 98, B75–B87. doi: 10.1016/j.cognition.2005.05.005
Parker Jones, O., Green, D. W., Grogan, A., Pliatsikas, C., Filippopolitis, K., Ali, N., et al. (2012). Where, when and why brain activation differs for bilinguals and monolinguals during picture naming and reading aloud. Cereb. Cortex 22, 892–902. doi: 10.1093/cercor/bhr161
R Core Team (2020). R: a Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing.
Santin, M., van Hout, A., and Flecken, M. (2021). Event endings in memory and language. Lang. Cogn. Neurosci. 36, 625–648.
Smith, N. J., and Levy, R. (2013). The effect of word predictability on reading time is logarithmic. Cognition 128, 302–319. doi: 10.1016/j.cognition.2013.02.013
Solomon, S. H., Hindy, N. C., Altmann, G. T. M., and Thompson-Schill, S. L. (2015). Competition between mutually exclusive object states in event comprehension. J. Cogn. Neurosci. 27, 2324–2338. doi: 10.1162/jocn_a_00866
Stanfield, R. A., and Zwaan, R. A. (2001). The effect of implied orientation derived from verbal context on picture recognition. Psychol. Sci. 12, 153–156. doi: 10.1111/1467-9280.00326
Tanenhaus, M. K., Spivey-Knowlton, M. J., Eberhard, K. M., and Sedivy, J. C. (1995). Integration of visual and linguistic information in spoken language comprehension. Science 268, 1632–1634.
Taylor, L. J., and Zwaan, R. A. (2009). Action in cognition: the case of language. Lang. Cogn. 1, 45–58.
Trenkic, D., Mirkovic, J., and Altmann, G. (2014). Real-time grammar processing by native and non-native speakers: constructions unique to the second language. Biling. Lang. Cogn. 17, 237–257.
van Dijk, T. A., and Kintsch, W. (1983). Strategies of Discourse Comprehension. New York: Academic Press.
Williams, G. P., Kukona, A., and Kamide, Y. (2019). Spatial narrative context modulates semantic (but not visual) competition during discourse processing. J. Mem. Lang. 108:104030.
Wilson, M. P., and Garnsey, S. M. (2009). Making simple sentences hard: verb bias effects in simple direct object sentences. J. Mem. Lang. 60, 368–392. doi: 10.1016/j.jml.2008.09.005
Yaxley, R. H., and Zwaan, R. A. (2007). Simulating visibility during language comprehension. Cognition 105, 229–236. doi: 10.1016/j.cognition.2006.09.003
Yee, E., and Sedivy, J. C. (2006). “Eye movements to pictures reveal transient semantic activation during spoken word recognition”: correction to Yee and Sedivy (2006). J. Exp. Psychol. Learn. Mem. Cogn. 32, 1–14. doi: 10.1037/0278-7393.32.1.1
Keywords: object-state, mental representation, visual world paradigm, non-native speakers, native speakers
Citation: Kang X and Ge H (2022) Tracking Object-State Representations During Real-Time Language Comprehension by Native and Non-native Speakers of English. Front. Psychol. 13:819243. doi: 10.3389/fpsyg.2022.819243
Received: 21 November 2021; Accepted: 08 February 2022;
Published: 04 March 2022.
Edited by:
Juhani Järvikivi, University of Alberta, CanadaReviewed by:
Brian Rusk, Da-Yeh University, TaiwanOleksandr Horchak, University Institute of Lisbon (ISCTE), Portugal
Copyright © 2022 Kang and Ge. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Xin Kang, xin.kang@cqu.edu.cn, orcid.org/0000-0002-1126-5771; Haoyan Ge, hge@hkmu.edu.hk, orcid.org/0000-0002-3120-1468