- 1Department of Experimental and Neurocognitive Psychology, Freie Universität Berlin, Berlin, Germany
- 2Centre for Language Studies, Department of Language and Communication, Faculty of Arts, Radboud University, Nijmegen, Netherlands
- 3Donders Centre for Cognition, Department of Artificial Intelligence, Faculty of Social Sciences, Radboud University, Nijmegen, Netherlands
- 4Center for Cognitive Neuroscience Berlin, Department of Education and Psychology, Free University of Berlin, Berlin, Germany
Literary reading is an interactive process between a reader and a text that depends on a balance between cognitive effort and emotional rewards. By studying both the crucial features of the text and of the subjective reader reception, a better understanding of this interactive process can be reached. In the present study, subjects (N=31) read and rated a work of narrative fiction that was written in a poetic style, thereby offering the readers two pathways to cognitive rewards: Aesthetic appreciation and narrative immersion. Using purely text-based quantitative descriptors, we were able to independently and accurately predict the subjective ratings in the dimensions comprehensibility, valence, arousal, and liking across roughly 140 pages of naturalistic text. The specific text features that were most important in predicting each rating dimension are discussed in detail. In addition, the implications of the findings are discussed more generally in the context of existing models of literary processing and future research avenues for empirical literary studies.
Introduction
Reading is a complex process. The meaning of individual words has to be decoded and integrated into a larger understanding of clauses, sentences, or entire paragraphs. In the case of fiction reading, this larger understanding of sentences and paragraphs again has to fit into the understanding of an overarching narrative. On top of that, narrative fiction typically entails considerably more literary language than non-fiction texts. Stylistic deviations from the linguistic norm are used to create non-familiar perspectives during reading, which is also known as foregrounding (Van Peer et al., 2007; Van Peer et al., 2021). For example, suddenly changing the rhyme scheme for dramatic effect would be stylistic foregrounding, and describing a desert as an endless sea of sand would be a sort of semantic foregrounding. Narrative and foregrounding aspects of literary texts both enforce active engagement of the reader. The text does not only convey factual information, but also evokes feelings as a response to both the narrated story and the style of language (Jacobs, 2015, 2023; Mar et al., 2011; Miall and Kuiken, 1994a, 1994b). In other words, literary reading is complex, but can be rewarding, in both cognitive and emotional terms.
Models of literary processing describe reader responses to fiction and verbal art in terms of neuronal activity, subjective experiences, and/or objective (i.e., intersubjectively examinable, directly observable) behavior. In any case, these models have to find a way to account for the complexity of the process (Jacobs, 2015, 2023; Jacobs and Willems, 2018; Willems and Jacobs, 2016). This is true for studies testing such models as well, in the sense that the stimulus material should ideally contain the potential to elicit a wide variety of possible responses from the readers. The longer a literary text is, the more it can be expected to naturally vary in properties that potentially elicit specific reactions in readers. For this paper, we chose a work of narrative poetry. That is, a work of literary fiction that tells a relatively long narrative using highly literary stylistics as stimulus material, thereby containing typical characteristics for narrative as well es poetic texts. Then, we calculated quantitative text features to describe this work of narrative poetry on a level of relatively small text units. For each text unit, we used the resulting features to predict the emotional responses the text elicited in readers, in the form of their subjective ratings. A direct example-wise comparison between the text feature profiles of two pages can be found in Figure 1.
Figure 1. Text profiles of two selected pages (translations in original meter by Bayard Taylor). More details about the abbreviations and all text features in Appendix 1.
Underlying processing mechanisms for literary texts
The Neurocognitive Poetics Model of Literary Reading (Schrott and Jacobs, 2011; Jacobs, 2015; Jacobs, 2023), or NCPM for short, specifies a central distinction between immersive and aesthetic processing. The former is expected to lead to fluent reading with a focus on the narrative. It should occur, for example, when reading suspenseful prose (cf. Hsu et al., 2014; Lehne et al., 2015; Wallentin et al., 2011). The latter is linked to slowed-down reading and appreciation of textual beauty most often associated with reading poetry (cf. Schrott and Jacobs, 2011; Jacobs, 2015; Jacobs, 2023, Miall and Kuiken, 1994a, 1994b).
The NCPM thus suggests two different approaches of readers emotionally processing literary texts, which are associated with differing outcomes on the subjective-experiential, neuro-cognitive, and objective-behavioral levels in the aforementioned studies. One is an immersive trajectory that heavily depends on gathering information to contextualize a story in its world, the other an aesthetic trajectory, in which the reader ponders about the emotional impact of the way in which language is used in the text. Previous studies have investigated either prosaic narratives (e.g., Lehne et al., 2015; Altmann et al., 2012; Eekhof et al., 2021; Hartung et al., 2021) or rather short poems (e.g., Lüdtke et al., 2014; Wassiliwizky et al., 2017). According to the above framework, this would always bias one of the two theoretical trajectories to be dominant. This results in a lack of clarity about how these two trajectories could interact with one another. The present study aims to achieve more clarity about such an interaction on the subjective-experiential level.
With our stimulus material of narrative poetry, we could expect three possible scenarios in the terms of the NCPM: Dominant immersive processing, dominant aesthetic processing, or a hybrid sort of processing that somehow incorporates elements of both immersive and aesthetic processing. In principle, there is also a fourth scenario in which neither immersive nor aesthetic processing takes place in any form. However, this would imply a complete failure of literary processing for narrative poetry. Literary processing can certainly fail for individual readers (not) engaging with individual texts, and the reasons for that are as multifaceted as human language processing is in general (Graf and Landwehr, 2015; Kidd et al., 2018; Tilmatine et al., 2021; Mak et al., 2022). Given the existence of various works of narrative poetry which receive exceptional critical acclaim, it seems highly unlikely that this fourth scenario would be generally true for established works of literature. Nonetheless, textual comprehensibility must be accounted for in studies like this one, especially as comprehensibility is also tied to liking, albeit in complicated ways (Güçlütürk et al., 2016).
The emotional aspect of literary processing
According to the Panksepp-Jakobson hypothesis – verified in several studies (Ziegler et al., 2018) –, humans use the same mechanisms for emotional processing in fiction processing as in real life (Jacobs, 2015). Therefore, the two different trajectories of the NCPM both contributing to emotional reader responses must translate into two different aspects of classic emotional processing. And duality is common in emotion theory: Most dimensional approaches to understanding emotion know some variations of concepts that are similar to the continuous duality of valence versus arousal (Rubin and Talarico, 2009).
While arousal has been related to narrative immersion, for example in the form of suspense (Kaakinen and Simola, 2020), the aesthetic trajectory has been linked to valence (Jacobs and Kinder, 2019). The link between appreciation of literary beauty and valence is not a simple linear one though; for instance, certain studies suggest a correlation between poetry appreciation and both positive and negative valence (Jacobs et al., 2016; Kraxenberger and Menninghaus, 2017). It thus seems like strong valence, positive or negative (as opposed to neutral valence), is one of the better quantitative predictors of aesthetic processing potential. A stronger role than valence, however, May be played by arousal-related social cognition processes like identification and suspense, as well as the reading flow, which is related to comprehensibility (Altmann et al., 2012; Thissen et al., 2018). As mentioned before, comprehensibility in itself is an important element in predicting both NCPM trajectories. This is because a lack of understanding May negatively impact the emotional response elicited by both prosaic narratives (Eekhof et al., 2021; Thissen et al., 2018) and poetic foregrounding (Harash, 2022; Scapin et al., 2023).
In sum, there are reasons to assume that arousal ratings are an indicator of the activation of the immersive processing trajectory, whereas valence ratings should relate more to the aesthetic processing trajectory. We have also established that both of these assumptions are only true if comprehensibility is accounted for, as well.
Text-based prediction of the average reader response to literature
A lot of predictive potential for reader responses is extractable from the text itself, in the textual cues that the author left for the readers to pick up. Authors can be identified based on their usage of language in song poetry (Mendhakar and Tilmatine, 2023) and authorial sentiment for expressively opinionated texts can be extracted fairly successfully with the assistance of deep-learning tools nowadays (Radford et al., 2017; Zhang et al., 2018; Jacobs and Kinder, 2020). Specifically multivariate sentiment analyses have been fairly successful in the endeavor to quantify more generalized emotional loadings in texts (Kim and Klinger, 2019; Jacobs et al., 2020), which can be linked to the concept of a linguistic emotion potential (Winko, 2023; Schwarz-Friesel, 2015). In such multivariate sentiment analyses, a wealth of text features that are computed by tools like Séance (Crossley et al., 2017) or SentiArt (Jacobs and Kinder, 2019) based on either individual words, bi-and trigrams or entire sentences have been shown to serve as fairly accurate predictors for the response of readers to both prose and poetry (for a recent comprehensive review see Jacobs, 2023).
For example, there is evidence that both surface and affective semantic features of the text predict eye movements in short sonnets (Xue et al., 2020; Xue et al., 2023) and that emotionally evocative text features predict affective ratings in short poems (Ullrich et al., 2017; Hugentobler and Lüdtke, 2021), sonnets, song lyrics, or entire books (Jacobs et al., 2020; Jacobs, 2023). These predictions are cross-validated by the fact that standard corpora of valence ratings for single words are similarly well predictable by those tools (e.g., Bestgen, 1994; Jacobs and Kinder, 2019; Lüdtke and Hugentobler, 2022). For short prosaic texts (excerpts and short stories), text features like word valence have been found to predict suspense and immersion ratings, as well as the neural correlates of empathy, mental simulation, predictive inference, and fear (Hsu et al., 2014; Lehne et al., 2015; Mak et al., 2023).
Computational text-analysis tools including multivariate sentiment analysis have thus been shown to predict emotional reader responses pretty well on the level of single words, poems, short prosaic texts, or whole books. They can be considered an excellent way to understand the interaction between readers and literary texts. On the other side of the interaction, we have discussed that four aspects of the reading experience May together cover a reasonable portion of the emotional reader response, namely valence, arousal, comprehensibility, and liking. These four aspects can be linked to the two processing trajectories of the NCPM: The immersive trajectory should lead to fluent reading, which is associated with comprehensibility and arousal. Meanwhile, the aesthetic trajectory should lead to dysfluent reading, which in turn is associated with valence and liking (Schrott and Jacobs, 2011; Jacobs, 2015; Jacobs, 2023). In that light, two main questions for this study emerge. First, do subjective ratings of valence, arousal, comprehensibility and liking show the differences and intercorrelations that we expect on the basis of the emotional reader response predictions of the NCPM? Second, can these ratings be predicted by the text features associated with both processing trajectories? And if so, which specific text properties play which role in the prediction of which rating dimension?
In this respect, we hypothesized that the ratings of comprehensibility would correlate with the ones for liking, because understanding a narrative should be a pre-requisite for enjoying its content. Possibly, the relationship between liking and comprehensibility ratings could also follow the inversed U-shape sometimes argued for narrative texts. If a quadratic fit performs better than a linear fit, this could be an indication that readers prefer narrative texts that are neither too complex nor too simplistic in their writing style (Berlyne, 1970; Errington et al., 2022; Jacobs, 2023). In a similar vein, we expected the arousal ratings to correlate with both the comprehensibility and liking ratings, as they are our chosen window into the immersive processing trajectory. Narrative comprehensibility is probably less important for the aesthetic processing trajectory of the poetic elements in the text. For that reason, we expected valence ratings to be less correlated with comprehensibility ratings, but as an element of appreciation to still be correlated with liking ratings. As arousal and valence are often conceptualized as two independent dimensions in the duality of emotion, we expected their respective ratings to be less correlated (Russell and Barrett, 1999). Experimental evidence has shown the arousal-valence ratings relation follows a U-shape, i.e., both positive and negative valence being related to high arousal (Lang, 1994; Jacobs et al., 2015; Kron et al., 2015).
We then selected a set of 29 text features that we expected to be important predictors for all or some of the four rating dimensions (Appendix 1), most of which were inspired by a previous study on profiling the comprehensibility, emotional tone, and topics of longer texts (Jacobs and Kinder, 2021; Jacobs, 2023). Specifically, we hoped to find four independent sets of features that each uniquely predicts one rating dimension. To account for valence and arousal ratings (cf. Ullrich et al., 2017), we selected vector-based features that measure the average emotional load of the semantic field of content words in a text unit. Specifically, we expected aesthetic-affective potential (AAP) from SentiArt (Jacobs and Kinder, 2019) to be the most important predictor for valence ratings, and the arousal potential from GLEAN (Lüdtke and Hugentobler, 2022) to be the most important predictor for arousal ratings. To specifically predict comprehensibility ratings, we included surface text features like the logarithmic type-token ratio, a measure of the morphological complexity of a text unit (Kettunen, 2014), and higher-level ease-of-processing measures like the average semantic neighborhood density (Fieder et al., 2019; Hameau et al., 2019; Jacobs, 2023), a measure calculated with vector-based deep learning for each content word. The text predictor we expected to be most important for comprehensibility ratings was the page average of logarithmic word frequency (Crossley et al., 2011; Smith and Levy, 2013; Xue et al., 2020).
Finally, we also included text features that we thought to be directly linked to the liking of a continuous text. Basic emotions are the driving forces behind high-level stimulus appreciation (Simonton, 1990; Delmonte, 2016; Jacobs et al., 2016). As we did not know whether narrative poetry processing is dominated by the aesthetic or the immersive trajectory, we included text features related to both. The AAP includes both elements of emotional valence and of verbal beauty (Jacobs and Kinder, 2020; Jacobs, 2023), so it was a strong contender to be the most important predictor of liking ratings, as well as an indicator of the role of the aesthetic processing trajectory for liking ratings. Regarding the immersive trajectory, we expected the strongest emotion-related predictor to be the arousal potential. Next to that, we also included text features more related to the narrative itself, such as measures of the probability of narrative event boundary occurrences (cf. Radvansky and Zacks, 2017; Geerligs et al., 2021) or the presence of the text’s global topics in a text unit, which May contribute to individual liking patterns of narratives (Mak et al., 2022).
We first ran the analyses with the full set of 29 text features predicting the four rating dimensions per page. Based on that, we created reduced models with the seven best-performing text features for each rating dimension. Based on these reduced models, we created bagged predictions for each rating dimension (Breiman, 1996; Hastie et al., 2009). Finally, we also compared the bagged predictions with the actual ratings using moving average windows along the course of the narrative.
Methods
Material
Johann Wolfgang von Goethe’s Faust – Der Tragödie Erster Teil, written in 1808, is a long and well-respected work of rhyming poetry with an intricate narrative. We selected two excerpts of roughly equal length (565 lines for excerpt “HS,” 519 lines for “MG”), both made up of multiple chapters of the play, from a modern transcription (1986). Both excerpts together consisted of 7,077 words (2,397 without duplicate words), arranged in 1084 lines, describing 26 scenes of slightly varying lengths. We subdivided the two excerpts into pages, following three guidelines with the following priorities: (1) There should be a narrative sense to the division into pages; (2) No chain of rhymes should be interrupted; (3) No page should consist of less than 4 or more than 12 lines. This subdivision into s had two reasons. For one, practicality: This format makes it easier to replicate this study with additional online measures like eye-tracking and fMRI. It also roughly follows the segmentation format used for the long narrative in the aforementioned study by Jacobs and Kinder (2019).
The source text is written as a play with stage directions noted along the verses (e.g., “Wagner in a nightcap and a nightgown enters the stage with a lamp in his hand”). For this study, we did not present the subjects any of these, but only numbered verses. However, we did add in-text quotation marks whenever the speaker changed (all verses in Faust I are direct speech by characters), as well as occasional context between the pages – clearly marked as such – with short overviews of the scene and the speaking characters in it. The subjects were told to only rate the original verses, and we only used the original verses for the textual analysis and prediction. The entire text and context can be found in Appendix 1.
We calculated 29 quantitative text properties measuring surface, affective-semantic, syntactic, and narrative-topical text aspects to describe the emotional loadings and writing style of each page. Detailed information about them, including their calculation methods, can be found in Appendix 2.
Participants
From the subject pool of the Freie Universität Berlin, 31 participants, all of which university students, took part in this study. The group that read excerpt HS consisted of 15 of them (13 female, 2 male, mean age: 23.73; SD = 6.11, 12 native speakers, mean completion time: 2703.67; SD = 887.08) and the group that read excerpt MG of the other 16 (12 female, 4 male, mean age: 25.75; SD = 9.54, 15 native speakers, mean completion time: 2728.00; SD = 711.25). The study was run entirely via the internet surveying tool SoSci.1 Each participant received course credits as compensation. The ethics committee of the FU’s Department of Education and Psychology approved the experimental procedures in this study (FU reference nr. 006/2021).
Procedure
The subjects went through a self-paced reading task for one of the two excerpts. They were instructed to carefully read each page and to evaluate it on four different 7-point Likert scales measuring Comprehensibility, Valence, Arousal, and Liking. As visible in Figure 2, each scale was characterized by different labels for the extreme values as follows (freely translated from German to English): incomprehensible vs. comprehensible, negative vs. positive, calming vs. exciting, I do not like it at all vs. I like it a lot. The second and third scale were accompanied by little cartoons giving a visual support for the intensity of Valence and Arousal for each point of the scale, respectively (Bradley and Lang, 1994). Each page text slide and the four corresponding scales were shown together on the same screen, until the subject clicked to continue to the next screen in their own pace (see Figure 2).
Figure 2. Example of a page (left) and the 7-point Likert scales used to survey subjects on their ratings for each page on Comprehensibility, Valence, Arousal, and Liking (right). In the experiment, the scales were placed beneath the text. Experimental flow chart; subjects clicked in their own pace through the pages of their excerpt and rated them one by one on the four dimensions Comprehensibility, Valence, Arousal, and Liking. Context appeared in another font color above the actual text and was clearly marked as such. The subjects were instructed to only rate the original text.
After going through all s, the subjects filled in a survey about their reading behavior, they went through list B of the German-language version of the Author Recognition Test (Grolig et al., 2018), and they rated a number of pre-selected quotes of unused parts of Goethe’s Faust on familiarity. Apart from the ART, these data were mostly collected for analyses that are not part of this paper.
Data analysis and results
First, we checked the plausibility of the ratings for each subject and excluded data accordingly. Then, we averaged each rating category (Liking, Arousal, Valence, and Comprehensibility) across the subjects on the page level. This resulted in four rating variables, each of which consisted of as many observations as there were pages (N = 141, see Table 1 for the descriptive statistics for the four rating variables). Then, we ran the analyses for this study in three steps. In the first step, we analyzed the rating behavior itself, specifically the inter-correlations between the four rating dimensions. As a second step, we tested the predictive accuracy of various text features for the comprehensibility, valence, arousal, and liking ratings, respectively. As a third step, we explored whether we could improve predictive accuracy by smoothing with moving average windows on the supra-page level.
Step 1: analyses of rating behavior
Here, we first checked general assumptions about the subjective ratings of text comprehensibility. Narrative poetry from the 19th century might be challenging to understand for the average contemporary reader. For that reason, we used the comprehensibility ratings to identify individual readers who May not have sufficiently understood the text to provide reliable content ratings for the other three rating dimensions. After that, we took a closer look at the page-level averages across subjects of the valence, arousal, and liking ratings, and how they interacted with each other in our data set.
Assumption checks about text comprehensibility
As visualized in Figure 3, there was quite some variance between subjects in terms of self-reported text comprehensibility. We removed three subjects (IDs: 192, 293, 324) from all further analyses because their ratings were on average lower than 4, the theoretical mean of the comprehensibility scale. Subject 293 also had the lowest score of all subjects on the author recognition test (34.67% correct), which can be regarded as a useful estimate of literary knowledge (Moore and Gordon, 2014). Subjects 192 (42.67% correct) and 324 (52.00% correct) were closer to, but still below the mean value of our sample for author recognition (percentage of correct answers M = 55.66, SD = 15.31). After removing these subjects from the dataset, we averaged the ratings of all remaining subjects per page and used these mean values for the rest of the analyses.
Figure 3. Comprehensibility ratings for each subject, boxplots represent the ratings per page (for excerpt HS in green and MG in violet); the three subjects on the very left of the graph were excluded from further analyses because they gave average Comprehensibility ratings in the lower half of the 7-point scale.
Interrelations between the subjective rating dimensions
Linear regressions (see Table 2) showed that the comprehensibility values were highly correlated with liking values, with a significant positive linear fit [beta = 0.565; t (1,138) = 5.56; R2 = 0.18]. A quadratic fit between these two rating groups did not perform better [betalinear = 0.565, betaquadratic = −0.059; t (2,137) = 5.50; R2 = 0.18], and did not show the hypothesized inverted U-shape. No significant linear fit was found for the relationship of the comprehensibility values with those of valence [beta = −0.016; t (1,138) = −0.22; R2 < 0.01] and arousal [beta = 0.061; t (1,138) = 0.48; R2 < 0.01]. Valence ratings were found to be a significant positive linear predictor for liking ratings [beta = 0.228; t (1,138) = 4.41; R2 = 0.12], which was not the case for the relationship between the ratings of arousal and liking [beta = 0.109; t (1,138) = 1.15; R2 < 0.01].
Step 2: text properties predicting rating behavior
For this step, we ran artificial neural networks (ANN) to predict each rating variable with specifically constructed models of seven different text features. We ran the ANNs over both excerpts together (ANN hyperparameters inspired by Xue et al., 2020: 5-fold cross-validation, one hidden layer with 10 nodes, hyperbolic tan activation function, learning rate = 0.1, number of tours = 10, number of models = 10).
Unlike other statistical approaches, this ANN approach allowed us to assess the importance of each individual predictor even though some of them might correlate or be masked behind non-linear correlations (Borsboom et al., 2021). This approach has proven to be prudent in previous similar studies (Xue et al., 2020; Musso et al., 2020; Eckstein et al., 2023). The effect size estimates were generated by the statistical software JMP 17 (SAS Institute Inc., 1989–2023), which produces both a main effect and a total effect for each predictor in a model. The total effect includes interaction effects with other predictors and thus gives a more complete insight into a predictor’s contribution to a model effect.
We always started with the same full model of 29 features as predictors, then reduced the model to the seven most relevant ones for each rating variable (in order not to surpass a threshold of one predictor per 20 observations, given the total of 140 content pages). Because k-fold cross-validated ANNs tend to vary a lot in their results per run, we let each of these analyses run a hundred times with each iteration having a different random seed and averaged the resulting performance (both training and validation) as well as main and total effects for each predictor. The mean total effects of each predictor across these 100 runs were used to select the seven most relevant predictors for each dimension-specific model. We ran another set of ANNs with the exact same approach as for the full models again, but this time for reduced models with only seven predictors each. We did this to make sure that the predictive accuracies of the reduced models were still high enough, and to better understand the role of each of the seven remaining predictors for each rating dimension.
Based on these reduced models, we created bagged predictions for each rating dimension (Breiman, 1996; Hastie et al., 2009) out of 100 bootstrapped samples for a single ANN run with the set of predictors identified for each reduced model. We used these to directly compare predicted (text features) and observed (subjective ratings) values in a linear regression.
Finally, we also looked into the usage of smoothing to account for effects of the immediate narrative context (cf. Elkins and Chun, 2019; Jacobs and Kinder, 2019). That is, we ran another set of linear regressions, this time between the moving average of each bagged predictions set and the moving average window of the corresponding rating variable on a supra-page level. For all moving average curves, we applied a window of 5 text units (pages).
To ensure validity and avoid the effects of potential auto-correlation introduced through this smoothing procedure, we used control analyses with permutation, that is, randomly scrambled item orders (cf. Crevecoeur et al., 2010; Ali et al., 2022). More precisely, we constructed a null model that the statistical comparisons of these moving average curves could be tested against by averaging the performance of a hundred comparisons between these variables with iteratively randomized page orders for the predictor. In other words, using randomized page orders, we tested if any performance improvements in this step were really due to the smoothing across this specific narrative, and not due to the act of smoothing itself.
As additional sanity check, we also ran additional simple (bivariate) linear regressions for each predictor in each reduced model. These are discussed in detail below, together with the reduced models themselves.
Comprehensibility ratings
The full model for comprehensibility averaged high predictive accuracies for both training (R2 = 0.91) and validation (R2 = 0.65) over 100 runs, the most important predictor being TTR, the logarithmic type-token ratio (Figure 4A). The reduced model, consisting of the seven most important predictors of the full model, averaged a predictive accuracy of R2 > 0.5 for both training (R2 = 0.64) and validation (R2 = 0.57) over 100 runs, the most important predictor again being TTR (Figure 4B).
Figure 4. (A) Average predictive accuracy of all full model predictors (text features) for the comprehensibility ratings across 100 runs; the seven predictors with the highest average total effect were used for the reduced model for comprehensibility rating predictions; (B) Average predictive accuracy of all full model predictors (text properties) for the Comprehensibility ratings across 100 runs; (C) Bagged predictions of the reduced model versus smoothed actual values of the comprehensibility ratings; the lines represent supra-page level smoothing that was done with a centered moving average window of 5 for both curves (all analyses were run across both excerpts together).
In a linear regression, the bagged predictions of the reduced model for the comprehensibility ratings showed a high predictive accuracy in a significant positive linear fit [beta = 1.327; t (1,139) = 19.19; p < 0.0001; R2 = 0.73]. After using a moving window average of 5 to smooth both variables on the supra-page level (Figure 4C), the bagged prediction curve showed an even more accurate significant positive linear fit with the Comprehensibility rating curve [beta = 1.266; t (1,134) = 27.43; p < 0.0001; R2 = 0.85], and performed better than all of the control analyses with scrambled page order (Figure 5).
Figure 5. In comparison with 100 permutations of smoothing with different page orders (gray), the smoothing along the original page orders (pink) performed a lot better in its predictive accuracy for the smoothed ratings in all four dimensions.
In the additional bivariate linear regressions for each of the predictors in the reduced model for comprehensibility ratings, there were multiple significant relationships. Specifically, the mean logarithmic word frequency, the logarithmic type-token ratio, the mean word concreteness, and the page number each had a positive significant relationship with the comprehensibility ratings, whereas the number of probable event boundaries had a significant negative relationship with the comprehensibility ratings. For an overview of all linear regressions we ran, consult Table 3.
Valence ratings
The full model for valence averaged high predictive accuracies for both training (R2 = 0.92) and validation (R2 = 0.75). As expected, emotion-related text properties like AAP, arousal shift, and joy potential all were among the seven most important text features for valence rating prediction in the full model, but the most important one was the number of probable narrative event boundaries (Figure 6A). The reduced model for valence averaged a predictive accuracy of R2 > 0.5 for both training (R2 = 0.62) and validation (R2 = 0.59), with AAP being the most important predictor (Figure 6B).
Figure 6. (A) Average predictive accuracy of all full model predictors (text properties) for the Valence ratings across 100 runs; the seven predictors with the highest average total effect were used for the reduced model for Valence rating predictions; (B) Average predictive accuracy of the seven reduced model predictors (text features) for the valence ratings across 100 runs; (C) Bagged predictions of the reduced model versus smoothed actual values of the valence ratings; the lines represent supra-page level smoothing that was done with a centered moving average window of 5 for both curves (all analyses were run across both excerpts together).
In a linear regression, the bagged predictions of the reduced model for the valence ratings showed a high predictive accuracy in a significant positive linear fit [beta = 1.409; t (1,139) = 17.71; p < 0.0001; R2 = 0.70]. After using a moving window average of 5 to smooth both variables on the supra-page level (Figure 6C), the bagged prediction curve of the reduced model showed an even more accurate significant positive linear fit with the valence rating curve [beta = 1.421; t (1,134) = 19.86; p < 0.0001; R2 = 0.75], and performed better than all of the control analyses with scrambled page order (Figure 5).
In the additional bivariate linear regressions for each of the predictors in the reduced model for valence ratings, there were multiple significant relationships. Specifically, the mean word sonority, the mean word AAP, the mean joy potential, and the number of probable event boundaries each had a positive significant relationship with the valence ratings, whereas the number of higher-frequent orthographic neighbors had a significant negative relationship with the valence ratings (Table 2).
Arousal ratings
The full model for arousal showed high predictive accuracies for both training (R2 = 0.83) and validation (R2 = 0.75) over 100 runs (Figure 7A). The reduced model, consisting of the seven most important predictors of the full model, averaged a predictive accuracy of R2 > 0.5 for both training (R2 = 0.59) and validation (R2 = 0.56) over 100 runs (Figure 7B).
Figure 7. (A) Average predictive accuracy of all full model predictors (text features) for the arousal ratings across 100 runs; the seven predictors with the highest average total effect were used for the reduced model for arousal rating predictions; (B) Average predictive accuracy of the seven reduced model predictors (text properties) for the Arousal ratings across 100 runs; (C) Bagged predictions of the reduced model versus smoothed actual values of the arousal ratings; the lines represent supra-page level smoothing that was done with a centered moving average window of 5 for both curves (all analyses were run across both excerpts together).
Mean content word arousal potential played a central role in the reduced model for arousal rating predictions, with a total effect R2 > 0.3. The other features in the same model with a predictive accuracy above that threshold is related to a specific story topic, i.e., an immersion-related feature.
In a linear regression, the bagged predictions of the reduced model for the arousal ratings showed a high predictive accuracy in a significant positive linear fit [beta = 1.427; t (1,139) = 18.68; p < 0.0001; R2 = 0.72]. After using a moving window average of 5 to smooth both variables on the supra-page level (Figure 7C), the bagged prediction curve showed an even more accurate significant positive linear fit with the arousal rating curve [beta = 1.437; t (1,134) = 25.10; p < 0.0001; R2 = 0.83], and performed better than all of the control analyses with scrambled page order (Figure 5).
In the additional bivariate linear regressions for each of the predictors in the reduced model for arousal ratings, there was only one significant relationship: The mean arousal potential had a significant positive relationship with the subjective arousal ratings (Table 2).
Liking ratings
The full model for liking averaged high predictive accuracies for both training (R2 = 0.82) and validation (R2 = 0.67). The reduced model for liking yielded a predictive accuracy of R2 > 0.5 for both training (R2 = 0.62) and validation (R2 = 0.59). Interestingly, AAP was not among the seven most important features for liking rating prediction in the full model (see Figure 8A).
Figure 8. (A) Average predictive accuracy of all full model predictors (text features) for the liking ratings across 100 runs; the seven predictors with the highest average total effect were used for the reduced model for liking rating predictions; (B) Average predictive accuracy of the 7 reduced model predictors (text properties) for the liking ratings across 100 runs; (C) Bagged predictions of the reduced model versus smoothed actual values of the liking ratings; the lines represent supra-page level smoothing that was done with a centered moving average window of 5 for both curves (all analyses were run across both excerpts together).
The on average most important predictors for the Liking ratings in the reduced model were the page number within the excerpt and the respective contribution to two different narrative topics (all three with R2 > 0.3). That being said, all seven predictors in the model had effects of with a stength of R2 > 0.2, making this reduced model more balanced in its predictor importance distribution than the other three reduced models we discussed (see Figure 8B).
In a linear regression, the bagged predictions of the reduced model for the liking ratings showed a high predictive accuracy in a significant positive linear fit [beta = 1.630; t (1,139) = 18.22; p < 0.0001; R2 = 0.71]. After using a moving window average of 5 to smooth both variables on the supra-page level (the lines in Figure 8C), the bagged prediction curve showed an even more accurate significant positive linear fit with the liking rating curve [beta = 1.740; t (1,134) = 21.51; p < 0.0001; R2 = 0.78], and performed better than all of the control analyses with scrambled page order (Figure 5).
In the additional bivariate linear regressions for each of the predictors in the reduced model for liking ratings, there was only one significant relationship: The contribution of a page to topic 5 had a significant positive relationship with the liking ratings (Table 2).
Discussion
With this study, we investigated the relationship between different subjective rating dimensions of literary reception, as well as the relationship between text features and these rating dimensions over the course of a long work of narrative poetry. More specifically, we asked how the rating dimensions valence, arousal, comprehensibility, and liking relate to each other and if the natural variation of features throughout a long narrative poetry text is useful to predict these dimensions over the course of the narrative. To account for the natural variation between the different features, we ran our analyses on a level of relatively small text units, namely pages of 4 to 12 verse lines. For each text unit, we used the resulting text features to predict the self-reported receptive response the text elicited in readers, in the form of their subjective ratings. Our procedure led to findings in two broader categories, which each need their own discussion in more detail: Findings about the relationships between the four rating dimensions, and findings about the specific text features predicting each rating dimension.
Rating patterns
The rating patterns alone contain plenty of valuable information. Arousal ratings were independent of the three other dimensions. As shown in Table 3, we observed only two significant relationships between our four rating variables. First, a significant positive correlation between comprehensibility and liking, and second, also a positive correlation between valence and liking. In short, the more comprehensible and positive a page was found to be, the higher its likeability. However, valence and comprehensibility ratings were not related to each other. Thus, they did not appear to interact with each other in predicting liking.
As fuel for the ongoing theoretical discourse about the relationship between comprehensibility and liking, these findings are especially interesting in the context of previous findings regarding the IDEST prose database (Kaakinen et al., 2022). For these short emotional stories, an inversed U-shape best described the relationship between these two constructs (Jacobs, 2023). This suggested that readers enjoy texts which are neither too difficult nor too easy to understand yielding a somewhat optimal level of activation or arousal (Berlyne, 1970; Errington et al., 2022). There is the possibility that an 19th century work of poetic literature is by its nature hard enough to read that the rightmost end of the inverse U-shape is not visible, that is, the (too) high comprehensibility leading to low-likeability part of the U-shape does not apply when a text is just not that comprehensible. However, the counter-argument would be that the average comprehensibility of the text excerpts was overall rated rather highly. In addition to that, the U-curve has been linked to individual differences before (Güçlütürk et al., 2016), which of course is an aspect of literary processing that we consciously put aside with our average-based approach to this study.
Alternatively, we May be observing a peculiarity of narrative poetry here, where liking and comprehensibility interact differently than in regular prose. The Pleasure-Interest model of aesthetic liking (Graf and Landwehr, 2015) argues that several processing filters apply during literary reading, depending on a balance between the need for cognitive enrichment (which influences liking) and the cognitive effort needed to achieve reading fluency (which depends on comprehensibility). It could be that the combined complexity of processing both poetry and a long narrative at the same time adds an additional layer of effort on top of the regularly varying comprehensibility in a text. In that case, the balance between effort and enrichment would be tilted toward effort, and any text passages that are higher in comprehensibility would linearly lead to higher liking.
The concept of a reliance of high-dimensional enjoyment on basic emotions would mean that the extreme instances of valence (i.e., both negative and positive) lead to higher liking ratings than flat (close-to-neutral) valence (Jacobs et al., 2016; Kraxenberger and Menninghaus, 2017). However, the relationship between valence and liking ratings was also linear and positive. The correlation between positive valence and high liking was thus expected, but it is remarkable that there was none between high liking and negative valence. A simple possible explanation could be the stimulus material itself: All of it being part of one single piece of work might just not include the full spectrum of valence that makes its varied relationship with liking visible in multiple-piece poetry studies.
Overall, the four chosen rating dimensions thus seem to cover four different aspects of the reader response. A positive Valence potential is better-liked by the reader, whereas high and low Arousal are emotional amplifications on an axis that is independent of Valence and Liking (cf. Russell and Barrett, 1999). The latter can also be regarded as an indication for an extended role of the aesthetic processing trajectory for the subjective Liking of narrative poetry. We could then state that literary appreciation based on Valence can be linked to the processing of poetry (cf. Jacobs and Kinder, 2019), whereas Arousal-based appreciation is typically linked to the processing of narratives (cf. Kaakinen and Simola, 2020).
Text features
All four rating dimensions were predicted very accurately by the full set of 29 features, all of them having training and validation R2s > = 0.65. Just like the rating patterns showed differences between the rating dimensions, there were also differences in which text features were more important for predicting each rating dimension. Reducing the full set to only seven features per rating dimension gave us a lot of additional information about which aspects of the text related to which rating dimension. Fewer predictors obviously lead to a performance drop, but each reduced set of features still performed fairly accurately, all of them having training and validation R2s > 0.5.
The additional simple bivariate linear regressions gave more insights into the directions of effect for some of the individual text features as predictors for each of the four rating dimensions. However, some text features did not have an observable significant relationship with a rating dimension individually, even when they did contribute to the predictive model in the ANN analyses (Table 2; Figures 4B, 6B, 7B, 8B). This showcases one of the main advantages of ANN analyses for literary text data: They allow us to account for non-linear dependencies and non-linear interactions within the data that we get from our multi-dimensional text profiling tools (Figure 1). These same text profiling tools and ANN analyses also showed us that there is a specific set of text features tied to each rating dimension (our reduced models). While these sets partially overlap, there was only one text feature that appeared as a predictor in all four reduced models: The number of probable narrative event boundaries on a page (Table 2). This text feature specifically, and the narrative segmentation aspect it represents in literary processing, certainly proved worthy of more attention in future research.
In sum, the average reader response on the subjective-experiential level was largely predictable with purely text-based predictors. In this study, the four rating dimensions comprehensibility, valence, arousal, and liking could well be predicted with only seven features each. We thus successfully replicated findings of previous studies using text analysis tools to predict reader responses (Crossley et al., 2017; Ullrich et al., 2017; Jacobs and Kinder, 2019; Xue et al., 2020; Lüdtke and Hugentobler, 2022; Xue et al., 2023; Mak et al., 2023; Lei et al., 2023) and extended them to ratings of narrative poetry, as well as to the broadest range of reader response dimensions so far. In addition to that, we used theoretically motivated text features rather than unsupervised deep-learning for our predictive models. As a result, we can further dissect the predictive models for each rating dimension to better understand which textual cues our readers used to evaluate their reading experience and accordingly rate the four dimensions. Such an approach can even explain which particular feature set was likely most influential for a given individual reader (Jacobs, 2023).
Text features predicting comprehensibility ratings
Regarding comprehensibility ratings, we expected surface text features to be the most important ones, that is, aspects of the text that are more related to its immediate readability than its semantic content. And indeed, the single most important predictor of comprehensibility ratings in both the full and the reduced models (Figures 4A,B) was the TTR. This feature indexes lexical diversity within a text unit, i.e., the number of repetitions of words that occur. This was thus an expected result and supports basic concepts on text readability (Graesser et al., 2004; Kettunen, 2014).
Interestingly, out of the six remaining predictors in the reduced model, only one other can be strictly defined as a surface text feature, namely log word frequency, which is also known to contribute to low-level (=surface) readability (Crossley et al., 2011; Smith and Levy, 2013; Xue et al., 2020). Two predictors in the reduced model were clearly related to the narrative, namely the number of probable narrative event boundaries on a page, as well as a page’s contribution to the semantic topic 2. Two other predictors were calculated using semantic networks, but arguably measure text readability, albeit on a higher level, namely the average content word concreteness (Sadoski et al., 2000) and the semantic similarity to the previous page, possibly operating as a measure of cohesion here (cf. Graesser et al., 2004).
Finally, the page number within the excerpt came in as second-most important predictor in the reduced model of comprehensibility ratings, with a total effect that was merely lower than the one of TTR, but with a much lower main effect. The latter suggests that much of the page number’s total effect came from interactions with the other predictors. The page number restarted counting for each excerpt, making it unlikely that this is a pure effect of readers getting tired over the course of the experiment. In that light, it seems possible that the total effect of the page number (also) reflects a narrative dimension of understanding, in the sense that the position of a page within an excerpt always also reflects the position of a page’s content within the narrative structure.
Either way, comprehensibility ratings seem to be linked to both surface text readability and a higher-level narrative understanding. The additional linear regressions underline this, showing that surface text features lead to higher comprehensibility ratings, but a large number of probable event boundaries lead to lower comprehensibility ratings. Surface text readability has been shown to play an important role for pure poetry processing (Xue et al., 2020; Xue et al., 2023), so it is reassuring that related text features also played a role for the comprehensibility of our poetic stimulus material, suggesting an involved aesthetic trajectory. At the same time, the importance of text features related to higher-level narrative understanding indicates that the immersive trajectory was also involved with the comprehensibility ratings. This could mean that there are two separate types of comprehensibility affecting ratings, one of which is related to lower-level surface understanding and aesthetic processing, and the other one to higher-level narrative understanding and immersive processing. Such a hypothesis is best tested with neuroimaging methodology, since the immersive and aesthetic trajectories manifest themselves most clearly on the neural level (Hartung et al., 2021; Jacobs, 2023).
Text features predicting valence ratings
We expected AAP to be the single most important predictor of valence ratings, given that it reflects both the potential for emotional positivity and the usage of stylistic beauty in a text (Jacobs, 2017, 2023; Jacobs and Kinder, 2019). We had associated both of these aspects to the aesthetic processing trajectory and found that AAP was indeed (I) the most important predictor in the reduced model of valence ratings (Figure 6B), (II) among the best few in the full model in terms of the total effect, and (III) had the single highest main effect (Figure 6A). The next best predictor in the reduced model was the average content word sonority, a feature reflecting both euphony and pronounceability (Vennemann, 1988, as cited in Jacobs, 2017; Xue et al., 2020), which thus highlights both text readability and the poetic element of narrative poetry.
Another predictor in the reduced model for the valence ratings that can be related to the aesthetic processing trajectory is SentiArt’s average content word joy potential, the most positive of the five basic emotions included in the full model and a feature that has been linked to elementary affective decisions and liking before (Dodds et al., 2011; Jacobs et al., 2016). Joy potential, sonority, and AAP also showed significant positive relationships with the valence ratings on their own in our additional linear regression analyses. All of these findings fit well with the concept of beauty appreciation, positive valence, and an aesthetic processing trajectory being closely linked with one another. The number of higher frequency neighbors (NHFN; Grainger and Jacobs, 1996) and TTR are both surface features, and their importance in predicting the valence ratings reinforces the idea that comprehensibility in poetry processing heavily depends on surface features (Xue et al., 2020, 2023). Given the independent linear alignment of both valence and comprehensibility ratings with liking ratings, it seems plausible that low-level surface understanding and the appreciation of positive valence both contribute to the likeability of a poetic text. Indeed, NHFN also showed an accordingly negative significant relationship with the valence ratings on its own in our additional linear regression analysis.
The other two predictors in the reduced model for valence ratings are the arousal shift, i.e., the difference between consecutive pages in terms of the average content word arousal, and the number of probable narrative event boundaries. Given their direct links to emotional arcs (cf. Reagan et al., 2016) and narrative structure (Radvansky and Zacks, 2017; Geerligs et al., 2021), respectively, we would view both of them as related mainly to the immersive processing trajectory. The positive significant relationship of the number of probable event boundaries on its own with the valence ratings that we found in our additional linear regression analysis implies that more narrative density leads to higher valence ratings. This also fits well with the dysfluency prediction for the aesthetic trajectory of the NCPM: More narrative density leads to more information to process and therefore less reading fluency, which co-occurs with higher valence ratings.
The reader’s emotional evaluation of narrative poetry in terms of valence thus seems to depend on both the textual esthetics of the poetry, and the immersive content of the narrative. Of course, it could also just be that the immersive trajectory is stronger than the aesthetic trajectory in narrative poetry reading, thereby also “leaking” into this rating dimension. To further investigate this, we are currently conducting research with methodology that allows us to explore the neurocognitive underpinnings of these processing trajectories.
Features predicting arousal ratings
Two aspects are remarkable about the reduced prediction model for arousal ratings (Figure 7B). First, GLEAN’s arousal potential is the most important predictor, neatly validating this specific text analysis tool. The arousal potential also showed the only significant relationship with the valence ratings on its own in our additional linear regression analysis – not so surprisingly, a positive one. The model also stands out because five of the remaining six features can be directly associated with the narrative; the only exception being semantic neighborhood density (SND). This fits perfectly with the notion that arousal ratings serve as a window to the immersive processing trajectory, through the detour of suspense correlations (cf. Wallentin et al., 2011; Kaakinen and Simola, 2020; Jacobs et al., 2016; Jacobs and Lüdtke, 2017). To be clear, as argued above, specifically the features page number and semantic similarity to the previous page probably do not only relate to the narrative, but also to certain aspects of comprehensibility. However, these are aspects of comprehensibility that do not relate to the text surface, but to higher-level understanding. Given its decidedly semantic aspects, this is also true for SND. Arousal ratings are thus clearly dependent on higher-level processing and likely to be related to the immersive processing trajectory.
Features predicting liking ratings
In the prediction of liking ratings, the seven predictors of the reduced model were somewhat close to each other in their average importance, creating a rather flat comparison of total effects, and generally rather low main effects (Figure 8B). This suggest that many aspects of text processing have to be accounted for when predicting liking ratings (cf. Jacobs et al., 2016; Jacobs, 2023), including individual differences (Jacobs, 2023; Lei et al., 2023). Using the same logic, we applied in the interpretation of the other reduced models, we can observe four features that clearly relate to the immersive processing trajectory, namely arousal shift, the number of probable narrative event boundaries, and the respective contributions to two different semantic topics. Interestingly, one of these two topics also showed the only significant relationship with the liking ratings on its own in our additional linear regression analyses – and it was a negative one: Our subject seemed to have particularly liked this on specific topic. Next to that, we find two features that have proven to be rather hard to classify between comprehensibility and immersion so far, namely page number in the excerpt, and semantic similarity to the previous page. It can thus be stated without a doubt that text features related to higher-level narrative understanding played an important role for our liking ratings.
According to previous research AAP should also be an important predictor for narrative liking (Jacobs and Kinder, 2019). However, this was not the case in our dataset. At the same time, the features we did observe to play a role for the liking ratings were still related to the narrative. The liking ratings were also linearly aligned with the comprehensibility ratings, and both dimensions share four out of seven predictors in their reduced models. Then again, the liking ratings were also linearly aligned with the valence ratings, but not with the arousal ratings. With all of that in mind, the complexity of liking ratings for narrative poetry cannot be overstated. If there are separate processing trajectories for immersive and aesthetic aspects of literary reception, they at the very least interact in the emergence of liking, if nowhere else.
Just like for arousal ratings, surface readability also seems to play less of a role for the liking ratings than higher-level comprehension does. In that regard, liking ratings for narrative poetry mostly seem to depend on higher-level processing. That being said, the sonority score feature can be linked to both pronounceability (and therefore readability) and aesthetic processing. Higher-level comprehension, the immersive processing trajectory, and sonority – possibly related to the poetic aspects of the text – are thus the central elements that predict liking ratings for narrative poetry, but not the basic emotions and stylistics that we expected to be important based on previous studies (Simonton, 1990; Delmonte, 2016; Jacobs et al., 2016). Again, liking ratings seem rather complex in their nature, which makes sense, given the complexity of our stimulus material, and the fact that we ignored reader-specific data in this study.
Outlook
The supra-page smoothing did clearly improve the predictive accuracy for all four reduced models in comparison to the page-level analyses, even when compared to the already performance-improved bagged predictions (cf. Reagan et al., 2016; Jacobs and Kinder, 2019). The permutation analyses showed that these are not effects of the smoothing itself, but effects directly related to this specific order of pages that constitute the narrative structure. Interestingly, this implies that there are effects of the larger narrative on the rating behavior that we could model with our page-based text analyses. That being said, some of our page-based features related to the larger narrative to begin with. Thus, our findings also serve as a reminder to consider various levels of text features when predicting reader responses. This of course is especially relevant for future research on other methodological levels: For instance, eye-tracking has a higher temporal resolution for reading research and provides reader response data even on the sub-lexical level. However, researchers limiting analyses of eye-tracking studies on literary reading to such a detailed level might for example miss possible effects of higher-level comprehensibility on reading speed.
Of course, there were also limitations to our approach. This study’s stimulus material is all taken from the same literary work. This was prudent in the sense that it allowed us to do comparisons between pages from one coherent text body, which freed us from having to account for idiosyncratic differences between different texts. However, it also limits the generalizability of our results to other the processing of other literary texts now. In the same vein, using stimulus material of the genre of narrative poetry allowed us to compare processing trajectories that would not interact so clearly in other text types, but it again limits the generalizability of our results for other genres. After all, different text types are also received differently by the reader (Quezada Gaponov et al., 2024).
We would like to claim that we investigated the processing of a naturalistic text in an experimental setting, but in reality, we had to walk a fine line in deciding where to compromise on which of those two aspects. Subjects were paid to read a certain text at a certain time, and did so in an experimental setting, constantly interrupting their reading flow to rate each page. All of this could be argued to be detrimental to a true naturalistic reading experience, but was necessary for experimental control. The subjects also read on a digital screen rather than on paper, although this hardly seems to be a limitation in terms of the reading experience (Sorrentino, 2021).
Finally, it is worth mentioning that for this study, we focused on the average reader response to literary texts rather than individual differences between readers in their responses. In our view, this was a necessary first step in this specific niche of empirical literary studies for now. We considered it worthwhile to first identify common denominators in the cognitive processes underlying literary reading before looking into individual nuances influencing these processes. The focus on the average reader allowed us to use sophisticated multivariate analyses rather than simpler mixed models, thereby creating a stronger theoretical base for future research, hopefully including many individual differences studies.
Now that we have a better insight into which text features drive which aspects of direct-subjective reader responses, it is easier to conduct more much-needed research on the indirect-objective level to find out more about the interaction between different processing trajectories, specifically with other behavioral and neuroimaging methods. Especially neuroimaging has proven to function as a detailed window into naturalistic narrative reception processes (Altmann et al., 2012, 2014; Hsu et al., 2015; Lehne et al., 2015; Geerligs et al., 2021; Le et al., 2022; Oetringer et al., 2024).
Conclusion
To summarize, it seems clear that the average subjective reader response, as sampled in the rating dimensions comprehensibility, valence, arousal, and liking, can successfully be predicted by quantitative text features, even if the text is a multi-layered and complex one like a work of narrative poetry. The multivariate ANN models all achieved high prediction accuracies, revealing the crucial aspects of reader’s text processing that influence their experiential judgments. Valence ratings were affected by features associated with both esthetics and narrative immersion, arousal ratings mostly by those associated with narrative immersion, and comprehensibility ratings by both surface features and high-level narrative aspects of the texts.
The possibly hardest to interpret results were obtained for liking ratings, which were predicted by a broad variety of as features, including those associated with narrative immersion and high-level comprehensibility. The more complex a text is, the more complex it seems to be to accurately predict how much subjects will like it (cf. Jacobs, 2015, 2023; Jacobs and Willems, 2018; Willems and Jacobs, 2016). That notwithstanding, we conclude that liking ratings for narrative poetry are mostly a result of the reader’s evaluation of a text’s comprehensibility, structural cohesion, and narrative content. In comparison, the role of aesthetic and word-level emotion potentials in this text evaluation is less clear. At present, we interpret this as readers treating narrative poetry more like a story than like a poem – although both aspects are present in the reader response.
In any case, it is clear that a sort of higher-level processing affected all rating dimensions through narrative structures, likely similar to the immersive processing trajectory of the NCPM. The NCPM’s aesthetic processing trajectory was less dominant in our subjective reader responses to narrative poetry, but it seemed to have played a role, given the observed importance of the AAP feature for predicting valence ratings. Comprehensibility does not seem to be strictly linked to the immersive trajectory, at least for the average reader. Rather, there seems to be a distinction between higher- and surface-level processing as well, possibly also related to both trajectories of the NCPM. If so, liking with its high linear correlation with both comprehensibility and emotional text aspects seems to be a potential window into the interaction between both trajectories that in most realistic reading scenarios do not exclude each other (Jacobs, 2021, 2023). The Liking ratings thereby also have proven to be heavily dependent on comprehensibility, but not exclusively, which fits nicely with the theoretical implications of the pleasure-interaction model of aesthetic liking (Graf and Landwehr, 2015). Future research using neuroimaging will help to account better for these nuances.
Data availability statement
The raw data supporting the conclusion of this article are available from the corresponding author on request.
Ethics statement
The studies involving humans were approved by Department of Education and Psychology, Freie Universität Berlin, reference nr. 006/2021. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
MT: Conceptualization, Data curation, Methodology, Visualization, Writing – original draft, Formal analysis, Investigation, Validation, Writing – review & editing. JL: Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Supervision, Writing – review & editing, Validation. AJ: Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. MT was funded by a grant within the ‘Empirical Study of Literature training network’ and European Union Horizon 2020 program (Marie Skłodowska-Curie, grant no. 860516).
Acknowledgments
In addition, we are grateful to Linda Geerligs from the Donders Center for Cognition in Nijmegen, The Netherlands, for her comments on an earlier version of the manuscript.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2024.1431764/full#supplementary-material
Footnotes
References
Ali, A., Ahmad, N., De Groot, E., Johannes Van Gerven, M. A., and Kietzmann, T. C. (2022). Predictive coding is a consequence of energy efficiency in recurrent neural networks. Patterns 3:100639. doi: 10.1016/j.patter.2022.100639
Altmann, U., Bohrn, I. C., Lubrich, O., Menninghaus, W., and Jacobs, A. M. (2012). The power of emotional valence—from cognitive to affective processes in reading. Front. Hum. Neurosci. 6. doi: 10.3389/fnhum.2012.00192
Altmann, U., Bohrn, I. C., Lubrich, O., Menninghaus, W., and Jacobs, A. M. (2014). Fact vs fiction-how paratextual information shapes our reading processes. Soc. Cogn. Affect. Neurosci. 9, 22–29. doi: 10.1093/scan/nss098
Berlyne, D. E. (1970). Novelty, complexity, and hedonic value. Percept. Psychophys. 8, 279–286. doi: 10.3758/BF03212593
Bestgen, Y. (1994). Can emotional valence in stories be determined from words? Cognit. Emot. 8, 21–36. doi: 10.1080/02699939408408926
Borsboom, D., Deserno, M. K., Rhemtulla, M., Epskamp, S., Fried, E. I., McNally, R. J., et al. (2021). Network analysis of multivariate data in psychological science. Nat. Rev. Methods Prim. 1:58. doi: 10.1038/s43586-021-00055-w
Bradley, M. M., and Lang, P. J. (1994). Measuring emotion: the self-assessment manikin and the semantic differential. J. Behav. Ther. Exp. Psychiatry 25, 49–59. doi: 10.1016/0005-7916(94)90063-9
Crevecoeur, F., Bollens, B., Detrembleur, C., and Lejeune, T. M. (2010). Towards a “gold-standard” approach to address the presence of long-range auto-correlation in physiological time series. J. Neurosci. Methods 192, 163–172. doi: 10.1016/j.jneumeth.2010.07.017
Crossley, S. A., Allen, D. B., and McNamara, D. S. (2011). Text readability and intuitive simplification: a comparison of readability formulas. Read. Foreign Lang 23, 84–101.
Crossley, S. A., Kyle, K., and McNamara, D. S. (2017). Sentiment analysis and social cognition engine (SEANCE): an automatic tool for sentiment, social cognition, and social-order analysis. Behav. Res. Methods 49, 803–821. doi: 10.3758/s13428-016-0743-z
Delmonte, R. (2016). Exploring Shakespeare’s sonnets with SPARSAR. Linguist. Lit. Stud. 4, 61–95. doi: 10.13189/lls.2016.040110
Dodds, P. S., Harris, K. D., Kloumann, I. M., Bliss, C. A., and Danforth, C. M. (2011). Temporal patterns of happiness and information in a global social network: Hedonometrics and twitter. PLoS One 6:e26752. doi: 10.1371/journal.pone.0026752
Eckstein, M. K., Summerfield, C., Daw, N. D., and Miller, K. J. (2023). Predictive and interpretable: combining artificial neural networks and classic cognitive models to understand human learning and decision making. bioRxiv 2023:541226. doi: 10.1101/2023.05.17.541226
Eekhof, L. S., Kuijpers, M. M., Faber, M., Gao, X., Mak, M., van den Hoven, E., et al. (2021). Lost in a story, detached from the words. Discourse Process. 58, 595–616. doi: 10.1080/0163853X.2020.1857619
Elkins, K., and Chun, J. (2019). Can sentiment analysis reveal structure in a plotless novel? arXiv 2019:1441. doi: 10.48550/arXiv.1910.01441
Errington, P. J., Thye, M., and Mirman, D. (2022). Difficulty and pleasure in the comprehension of verb-based metaphor sentences: a behavioral study. PLoS One 17:e0263781. doi: 10.1371/journal.pone.0263781
Fieder, N., Wartenburger, I., and Abdel Rahman, R. (2019). A close call: interference from semantic neighbourhood density and similarity in language production. Mem. Cogn. 47, 145–168. doi: 10.3758/s13421-018-0856-y
Geerligs, L., van Gerven, M., and Güçlü, U. (2021). Detecting neural state transitions underlying event segmentation. NeuroImage 236:118085. doi: 10.1016/j.neuroimage.2021.118085
Graesser, A. C., McNamara, D. S., Louwerse, M. M., and Cai, Z. (2004). Coh-Metrix: analysis of text on cohesion and language. Behav. Res. Methods Instrum. Comput. 36, 193–202. doi: 10.3758/BF03195564
Graf, L. K. M., and Landwehr, J. R. (2015). A dual-process perspective on fluency-based aesthetics: the pleasure-interest model of aesthetic liking. Personal. Soc. Psychol. Rev. 19, 395–410. doi: 10.1177/1088868315574978
Grolig, L., Tiffin-Richards, S. P., and Schroeder, S. (2020). Print exposure across the reading life span. Reading and Writing, 33, 1423–1441. doi: 10.1007/s11145-019-10014-3
Grainger, J., and Jacobs, A. M. (1996). Orthographic processing in visual word recognition: a multiple read-out model. Psychol. Rev. 103, 518–565. doi: 10.1037//0033-295x.103.3
Güçlütürk, Y., Jacobs, R. H. A. H., and Lier, R. V. (2016). Liking versus complexity: decomposing the inverted U-curve. Front. Hum. Neurosci. 10:112. doi: 10.3389/fnhum.2016.00112
Hameau, S., Nickels, L., and Biedermann, B. (2019). Effects of semantic neighbourhood density on spoken word production. Q. J. Exp. Psychol. 72, 2752–2775. doi: 10.1177/1747021819859850
Harash, A. (2022). The model of failed foregrounding. Psychol. Aesthet. Creat. Arts 16, 594–609. doi: 10.1037/aca0000419
Hartung, F., Wang, Y., Mak, M., Willems, R., and Chatterjee, A. (2021). Aesthetic appraisals of literary style and emotional intensity in narrative engagement are neurally dissociable. Commun. Biol. 4:1401. doi: 10.1038/s42003-021-02926-0
Hastie, T., Tibshirani, R., and Friedman, J. (2009). The elements of statistical learning. New York: Springer.
Hsu, C.-T., Conrad, M., and Jacobs, A. M. (2014). Fiction feelings in Harry potter: haemodynamic response in the mid-cingulate cortex correlates with immersive reading experience. Neuroreport 25, 1356–1361. doi: 10.1097/WNR.0000000000000272
Hsu, C.-T., Jacobs, A. M., Altmann, U., and Conrad, M. (2015). The magical activation of left amygdala when Reading Harry potter: an fMRI study on how descriptions of supra-natural events entertain and enchant. PLoS One 10:e0118179. doi: 10.1371/journal.pone.0118179
Hugentobler, K. G., and Lüdtke, J. (2021). Micropoetry meets neurocognitive poetics: influence of associations on the reception of poetry. Front. Psychol. 12:737756. doi: 10.3389/fpsyg.2021.737756
Jacobs, A. M. (2015). Neurocognitive poetics: methods and models for investigating the neuronal and cognitive-affective bases of literature reception. Front. Hum. Neurosci. 9:186. doi: 10.3389/fnhum.2015.00186
Jacobs, A. M. (2017). Quantifying the beauty of words: a neurocognitive poetics perspective. Front. Hum. Neurosci. 11:622. doi: 10.3389/fnhum.2017.00622
Jacobs, A. M. (2021). “The neurocognitive poetics model of literary Reading 10 years after” in Neuroaesthetics in focus. eds. A. Chatterjee and E. Cardillo (Oxford: Oxford University Press).
Jacobs, A. M. (2023). Neurocomputational poetics: How the brain processes verbal art. London: Anthem Press.
Jacobs, A. M., Herrmann, B., Lauer, G., Lüdtke, J., and Schroeder, S. (2020). Sentiment analysis of children and youth literature: is there a Pollyanna effect? Front. Psychol. 11:574746. doi: 10.3389/fpsyg.2020.574746
Jacobs, A., Hofmann, M. J., and Kinder, A. (2016). On elementary affective decisions: to like or not to like, that is the question. Front. Psychol. 7:1836. doi: 10.3389/fpsyg.2016.01836
Jacobs, A. M., and Kinder, A. (2019). Computing the affective-aesthetic potential of literary texts. AI 1, 11–27. doi: 10.3390/ai1010002
Jacobs, A. M., and Kinder, A. (2020). Quasi error-free text classification and authorship recognition in a large corpus of English literature based on a novel feature set. arXiv 2020:10801. doi: 10.48550/arXiv.2010.10801
Jacobs, A. M., and Kinder, A. (2021). Electoral programs of German parties 2021: a computational analysis of their comprehensibility and likeability based on Senti art 2021:12500. doi: 10.48550/ARXIV.2109.12500
Jacobs, A. M., and Lüdtke, J. (2017). “Immersion into narrative and poetic worlds” in Narrative absorption. eds. F. Hakemulder, M. M. Kuijpers, E. S. H. Tan, K. Bálint, and M. M. Doicaru (Amsterdam, Netherlands: Benjamins), 69–96.
Jacobs, A. M., Võ, M. L.-H., Briesemeister, B. B., Conrad, M., Hofmann, M. J., Kuchinke, L., et al. (2015). 10 years of BAWLing into affective and aesthetic processes in reading: what are the echoes? Front. Psychol. 6:714. doi: 10.3389/fpsyg.2015.00714
Jacobs, A. M., and Willems, R. M. (2018). The fictive brain: neurocognitive correlates of engagement in literature. Rev. Gen. Psychol. 22, 147–160. doi: 10.1037/gpr0000106
Kaakinen, J. K., and Simola, J. (2020). Fluctuation in pupil size and spontaneous blinks reflect story transportation. J. Eye Mov. Res. 13:13. doi: 10.16910/jemr.13.3.6
Kaakinen, J. K., Werlen, E., Kammerer, Y., Acartürk, C., Aparicio, X., Baccino, T., et al. (2022). IDEST: international database of emotional short texts. PLoS One 17:e0274480. doi: 10.1371/journal.pone.0274480
Kettunen, K. (2014). Can type-token ratio be used to show morphological complexity of languages? J. Quant. Linguist. 21, 223–245. doi: 10.1080/09296174.2014.911506
Kidd, E., Donnelly, S., and Christiansen, M. H. (2018). Individual differences in language acquisition and processing. Trends Cogn. Sci. 22, 154–169. doi: 10.1016/j.tics.2017.11.006
Kim, E., and Klinger, R. (2019). A survey on sentiment and emotion analysis for computational literary studies. Z. Digit. Geisteswissensch. 2019:8. doi: 10.17175/2019_008
Kraxenberger, M., and Menninghaus, W. (2017). Affinity for poetry and aesthetic appreciation of joyful and sad poems. Front. Psychol. 7:2051. doi: 10.3389/fpsyg.2016.02051
Kron, A., Pilkiw, M., Banaei, J., Goldstein, A., and Anderson, A. K. (2015). Are valence and arousal separable in emotional experience? Emotion 15, 35–44. doi: 10.1037/a0038474
Lang, P. J. (1994). “The motivational organization of emotion: affect-reflex connections” in Emotions: Essays on emotion theory. eds. S. H. M. van Goozen, N. E. Van de Poll, and J. A. Sergeant (New Jersey: Lawrence Erlbaum Associates, Inc.), 61–93.
Le, L., Ambrogioni, L., Seeliger, K., Güçlütürk, Y., Van Gerven, M., and Güçlü, U. (2022). Brain2Pix: fully convolutional naturalistic video frame reconstruction from brain activity. Front. Neurosci. 16:940972. doi: 10.3389/fnins.2022.940972
Lehne, M., Engel, P., Rohrmeier, M., Menninghaus, W., Jacobs, A. M., and Koelsch, S. (2015). Reading a suspenseful literary text activates brain areas related to social cognition and predictive inference. PLoS One 10:e0124550. doi: 10.1371/journal.pone.0124550
Lei, A., Willems, R. M., and Eekhof, L. S. (2023). Emotions, fast and slow: processing of emotion words is affected by individual differences in need for affect and narrative absorption. Cognit. Emot. 37, 997–1005. doi: 10.1080/02699931.2023.2216445
Lüdtke, J., and Hugentobler, K. G. (2022). “Using emotional word ratings to extrapolated norms for valence, arousal, imagebility and concreteness: the German list of extrapolated affective norms (GLEAN)” in Proceedings of KogWis2022, the 5th biannual conference of the German Society for Cognitive Science. eds. E. Ferstl, L. Konieczny, R. von Stülpnagel, J. Beck, and L. Zacharski (Freiburg, Germany: Albert-Ludwigs-Universität Freiburg).
Lüdtke, J., Meyer-Sickendieck, B., and Jacobs, A. M. (2014). Immersing in the stillness of an early morning: testing the mood empathy hypothesis of poetry reception. Psychol. Aesthet. Creat. Arts 8, 363–377. doi: 10.1037/a0036826
Mak, M., Faber, M., and Willems, R. M. (2022). Different routes to liking: how readers arrive at narrative evaluations. Cogn. Res. 7:72. doi: 10.1186/s41235-022-00419-0
Mak, M., Faber, M., and Willems, R. M. (2023). Different kinds of simulation during literary reading: insights from a combined fMRI and eye-tracking study. Cortex 162, 115–135. doi: 10.1016/j.cortex.2023.01.014
Mar, R. A., Oatley, K., Djikic, M., and Mullin, J. (2011). Emotion and narrative fiction: interactive influences before, during, and after reading. Cognit. Emot. 25, 818–833. doi: 10.1080/02699931.2010.515151
Mendhakar, A., and Tilmatine, M. (2023). Automatic authorship classification for German lyrics using Naïve bayes. J. Lang. Technol. Comput. Linguist. 36, 171–182. doi: 10.21248/jlcl.36.2023.242
Miall, D. S., and Kuiken, D. (1994a). Beyond text theory: understanding literary response. Discourse Process. 17, 337–352. doi: 10.1080/01638539409544873
Miall, D. S., and Kuiken, D. (1994b). Foregrounding, defamiliarization, and affect: response to literary stories. Poetics 22, 389–407. doi: 10.1016/0304-422X(94)00011-5
Moore, M., and Gordon, P. C. (2015). Reading ability and print exposure: Item response theory analysis of the author recognition test. Behavior Research Methods, 47, 1095–1109. doi: 10.3758/s13428-014-0534-3
Musso, M. F., Hernández, C. F. R., and Cascallar, E. C. (2020). Predicting key educational outcomes in academic trajectories: a machine-learning approach. High. Educ. 80, 875–894. doi: 10.1007/s10734-020-00520-7
Oetringer, D., Gözükara, D., Güçlü, U., and Geerligs, L. (2024). The neural basis of event segmentation: stable features in the environment are reflected by neural states. bioRxiv 1:577369. doi: 10.1101/2024.01.26.577369
Quezada Gaponov, C., Aravena, S., Maldonado, M., and Coloma, C. J. (2024). Desarrollo de las habilidades lingüísticas y lectoras y su relación con la comprensión de textos narrativos y expositivos en estudiantes de segundo y tercer grado. Investig. Sobre Lect. 18, 115–144. doi: 10.24310/isl.2.18.2023.16511
Radford, A., Jozefowicz, R., and Sutskever, I. (2017). Learning to generate reviews and discovering sentiment 2017:1444. doi: 10.48550/ARXIV.1704.01444
Radvansky, G. A., and Zacks, J. M. (2017). Event boundaries in memory and cognition. Curr. Opin. Behav. Sci. 17, 133–140. doi: 10.1016/j.cobeha.2017.08.006
Reagan, A. J., Mitchell, L., Kiley, D., Danforth, C. M., and Dodds, P. S. (2016). The emotional arcs of stories are dominated by six basic shapes. EPJ Data Sci. 5:31. doi: 10.1140/epjds/s13688-016-0093-1
Rubin, D. C., and Talarico, J. M. (2009). A comparison of dimensional models of emotion: evidence from emotions, prototypical events, autobiographical memories, and words. Memory 17, 802–808. doi: 10.1080/09658210903130764
Russell, J. A., and Barrett, L. F. (1999). Core affect, prototypical emotional episodes, and other things called emotion: dissecting the elephant. J. Pers. Soc. Psychol. 76, 805–819. doi: 10.1037/0022-3514.76.5.805
Sadoski, M., Goetz, E. T., and Rodriguez, M. (2000). Engaging texts: effects of concreteness on comprehensibility, interest, and recall in four text types. J. Educ. Psychol. 92, 85–95. doi: 10.1037/0022-0663.92.1.85
Scapin, G., Loi, C., Hakemulder, F., Bálint, K., and Konijn, E. (2023). The role of processing foregrounding in empathic reactions in literary reading. Discourse Process. 1, 1–21. doi: 10.1080/0163853X.2023.2198813
Schrott, R., and Jacobs, A. M. (2011). Gehirn und Gedicht: Wie wir unsere Wirklichkeiten konstruieren. Munich, Germany: Carl Hanser Verlag.
Schwarz-Friesel, M. (2015). “Language and emotion: the cognitive linguistic perspective” in Consciousness & Emotion Book Series. ed. U. M. Lüdtke , vol. 10 (Amsterdam, Netherlands: John Benjamins Publishing Company), 157–174.
Simonton, D. K. (1990). Lexical choices and aesthetic success: a computer content analysis of 154 Shakespeare sonnets. Comput. Hum. 24, 251–264. doi: 10.1007/BF00123412
Smith, N. J., and Levy, R. (2013). The effect of word predictability on reading time is logarithmic. Cognition 128, 302–319. doi: 10.1016/j.cognition.2013.02.013
Sorrentino, P. (2021). Does the medium matter? Digital vs. paper reading for leisure and foreign language learning doctoral dissertation, Dissertation. Göttingen, Georg-august Universität.
Thissen, B. A. K., Menninghaus, W., and Schlotz, W. (2018). Measuring optimal Reading experiences: the Reading flow short scale. Front. Psychol. 9:2542. doi: 10.3389/fpsyg.2018.02542
Tilmatine, M., Hubers, F., and Hintz, F. (2021). Exploring individual differences in recognizing idiomatic expressions in context. J. Cogn. 4:37. doi: 10.5334/joc.183
Ullrich, S., Aryani, A., Kraxenberger, M., Jacobs, A. M., and Conrad, M. (2017). On the relation between the general affective meaning and the basic sublexical, lexical, and inter-lexical features of poetic texts–a case study using 57 poems of H. M. Enzensberger. Front. Psychol. 7:2073. doi: 10.3389/fpsyg.2016.02073
Van Peer, W., Hakemulder, J., and Zyngier, S. (2007). Lines on feeling: foregrounding, aesthetics and meaning. Lang. Lit. 16, 197–213. doi: 10.1177/0963947007075985
Van Peer, W., Sopčák, P., Castiglione, D., Fialho, O., Jacobs, A. M., and Hakemulder, F. (2021). “Foregrounding” in Handbook of empirical literary studies. eds. D. Kuiken and A. M. Jacobs (Berlin, Germany: De Gruyter), 145–176.
Vennemann, T. (1987). Preference laws for syllable structure: And the explanation of sound change with special reference to German, Germanic, Italian, and Latin. de Gruyter.
Wallentin, M., Nielsen, A. H., Vuust, P., Dohn, A., Roepstorff, A., and Lund, T. E. (2011). Amygdala and heart rate variability responses from listening to emotionally intense parts of a story. NeuroImage 58, 963–973. doi: 10.1016/j.neuroimage.2011.06.077
Wassiliwizky, E., Koelsch, S., Wagner, V., Jacobsen, T., and Menninghaus, W. (2017). The emotional power of poetry: neural circuitry, psychophysiology and compositional principles. Soc. Cogn. Affect. Neurosci. 12, 1229–1240. doi: 10.1093/scan/nsx069
Willems, R. M., and Jacobs, A. M. (2016). Caring about Dostoyevsky: the untapped potential of studying literature. Trends Cogn. Sci. 20, 243–245. doi: 10.1016/j.tics.2015.12.009
Winko, S. (2023). “68 literature and emotion” in Handbücher zur Sprach- und Kommunikationswissenschaft/handbooks of linguistics and communication science [HSK] 46/3. eds. G. L. Schiewer, J. Altarriba, and B. C. Ng (Berlin, Germany: De Gruyter), 1417–1436.
Xue, S., Jacobs, A. M., and Lüdtke, J. (2020). What is the difference? Rereading Shakespeare’s sonnets–an eye tracking study. Front. Psychol. 11:421. doi: 10.3389/fpsyg.2020.00421
Xue, S., Lüdtke, J., and Jacobs, A. M. (2023). Once known, twice hedonic: enjoying Shakespeare’s sonnets through rereading–a deep learning perspective. Psychol. Aesthet. Creat. Arts 1:621. doi: 10.1037/aca0000621
Zhang, L., Wang, S., and Liu, B. (2018). Deep learning for sentiment analysis: a survey. WIREs. Data Min. Knowl. Disc. 8:1253. doi: 10.1002/widm.1253
Keywords: reading, narrative cognition, empirical study of literature, aesthetic perception, rating prediction, quantitative text analysis, naturalistic stimuli
Citation: Tilmatine M, Lüdtke J and Jacobs AM (2024) Predicting subjective ratings of affect and comprehensibility with text features: a reader response study of narrative poetry. Front. Psychol. 15:1431764. doi: 10.3389/fpsyg.2024.1431764
Edited by:
Elena Jiménez-Pérez, University of Malaga, SpainReviewed by:
Rocío Díaz Zavala, National University of Saint Augustine, PeruPilar Ferré Romeu, University of Rovira i Virgili, Spain
Pedro García Guirao, University of Malaga, Spain
Copyright © 2024 Tilmatine, Lüdtke and Jacobs. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Mesian Tilmatine, m.tilmatine@fu-berlin.de