- Tübingen Center for Mental Health, University Hospital and Faculty of Medicine, University of Tübingen, Tübingen, Germany
Introduction: This study explores the stability of scores on psychometrically validated trait questionnaires over time. We illustrate potential pitfalls through a larger study that used the Ruminative Response Scale (RRS) to categorize participants prior to study inclusion into two groups based on their habitual tendency to ruminate. Surprisingly, when we re-administered the RRS at the start of an experimental session, significant score changes occurred, resulting in participants shifting between the two groups.
Methods: To address this, we modified our recruitment process, aiming to reduce careless responses, including an online RRS assessment a week before the lab appointment. We analyzed the different samples prior to and after changing the recruitment procedure, as well as the total sample regarding the psychometric properties of the RRS. We also explored various indices to identify and predict score changes due to careless responding; however, only a subgroup of participants was successfully identified.
Results: Our findings suggest that Mahalanobis distances are effective for identifying substantial score changes, with baseline state rumination emerging as a marginally significant predictor.
Discussion: We discuss the importance of conducting manipulation checks and offer practical implications for research involving psychometrically validated trait questionnaires.
Introduction
Whenever questionnaire data are assessed, there will be a certain percentage of invalid data. Johnson (2005) summarizes three main classes of invalid data: linguistic incompetence/misunderstanding, misrepresentation, and careless response. Linguistic incompetence or misunderstanding refers to invalid data due to insufficient basic verbal comprehension. Misrepresentation, on the other hand, refers to presenting oneself in a way that is misleading or inaccurate (e.g., “faking good” and “faking bad”). Finally, careless responding is defined as follows: “Careless responding occurs when respondents fail to read or give sufficient attention to item content, resulting in data that may not accurately reflect respondents’ true levels of the constructs being measured (Meade and Craig 2012; Ward and Meade 2018)” (Ward and Meade, 2023, p. 578). Because it is difficult to find a clear definition and unique characteristics of careless responding (e.g., invariability of responses, fast responses, inconsistency), estimates of prevalence also vary substantially across studies. A recent investigation by Jones et al. (2022), where 48 crowdsourced alcohol-related studies were analyzed regarding the occurrence of careless responses, estimated the pooled prevalence rate at ∼11.7%, which is assumed to be generalizable to other fields of research (Ward and Meade, 2023). Despite the increasing knowledge and awareness of the topic, careless responding is so far rarely investigated in psychological studies, whose scientific aim is different from the explicit investigation of careless responses. That means psychological research is not routinely screened for invalid data in general. This may be due to the lack of clear guidelines concerning effective detection and elimination (Ward and Meade, 2023).
In psychology, a trait is defined as “an enduring personality characteristic that describes or determines an individual’s behavior across a range of situations” (American Psychological Association, n.d.). Usually, traits are assessed using self-report questionnaires, where different statements are rated in terms of how participants typically think or behave. An implicit assumption of these questionnaires is that they are answered in more or less the same way regardless of the situation in which they are administered, which manifests the difference between state measures (Geiser et al., 2017).
Currently, several theories exist regarding ruminative thinking (Nolen-Hoeksema et al., 2008), each with distinct emphases and implications for temporal stability (Smith and Alloy, 2009). One of the earliest theories, known as the Response Style Theory proposed by Nolen-Hoeksema (1991), primarily focuses on depressive rumination (Papageorgiou and Wells, 2004). This theory posits that rumination is a trait-like behavior, defined as “thoughts and behaviors that center on one’s depressive symptoms and their implications” (Nolen-Hoeksema and Morrow, 1991, p. 569). Nolen-Hoeksema and Morrow (1991) also introduced the Ruminative Response Scale (RRS), which remains widely used for assessing trait rumination to this day. Rumination is nowadays regarded as a transdiagnostic process apparent in many psychopathologies, which manifests the importance and relevance of investigation in the context of improving treatment options for mental disorders.
Using the data from a recent study by our group investigating the neural correlates of ruminative thinking in response to a social stress induction using the Trier Social Stress Test (TSST) (Kirschbaum et al., 1993) and the impact of Theta-Burst Stimulation (TBS), we wanted to investigate careless responding using a real dataset. In the aforementioned study, participants were screened using the RRS (Treynor and Gonzalez, 2003) in order to assess participants’ habitual ruminative tendencies and assign them to two stratified groups (low vs. high trait ruminators).
The aims of the current investigation were to evaluate the psychometric properties of the Ruminative Response Scale, to explore the predictive value of different indicators of careless responses, and finally to evaluate the efficacy of the steps we took in changing the recruitment procedure to minimize careless responding in the first place.
Methods
Participants
The sample analyzed here was originally recruited within a larger project investigating the effects of Theta-Burst Stimulation on the stress response in low and high ruminators. For this, a total of 120 right-handed healthy volunteers were recruited via posts spread across the university hospital and social media platforms. Potentially interested participants knew there would be a stress induction and neurostimulation at each of the two appointments at the laboratory, scheduled approximately 5 weeks apart. Inclusion criteria encompassed individuals aged 18–50 with normal or corrected vision, right-handedness, absence of metal in the skull/brain, and proficiency in the German language. Exclusion criteria involved any medical conditions, including diabetes mellitus, renal insufficiency, uncontrolled hypertension, history of traumatic brain injury, cardiac arrhythmias, acute substance abuse, adrenal insufficiency, any acute psychiatric or neurological disorder, and pregnancy in women (for a list of inclusion and exclusion criteria, see Supplementary material S1). All procedures were approved by the ethics committee at the University Hospital and University of Tübingen and were in line with the Declaration of Helsinki in its latest version. Initially, an eligibility screening using an online questionnaire ( ) using SoSci Survey was completed, where demographic and clinical variables as well as the revised Ruminative Response Scale (RRS) (Treynor and Gonzalez, 2003) were assessed. According to an a priori power analysis for the main research question of the larger study aiming to investigate the effect of Theta-Burst Stimulation on stress-reactive rumination, we aimed to recruit a stratified sample of 44 low- and 44 high trait ruminators [low trait ruminators: mean RRS ≤ 1.82 (percentile rank 27); high trait ruminators: mean RRS ≥ 2.36 (percentile rank 65)]. For a sensitivity analysis of the reported analyses, we refer to Supplementary material S2. These corresponding cutoffs are based on the combined data of 983 participants from prior studies of our group (reference deleted for blind peer review) (Rosenbaum et al., 2018a,b, 2021). All eligible volunteers received an invitation to participate in the study. At the beginning of the first of two experimental sessions, participants completed the RRS again in paper–pencil format ( ). After completion of both appointments at the laboratory, participants were compensated with 100 € or course credit. We observed substantial changes in RRS scores from the online screening ( ) to the assessment of the RRS at the laboratory ( ) after having recruited and partly assessed a total of n = 52 participants. As a consequence, we changed the recruitment procedure as follows: To minimize the likelihood of the questionnaire assessing momentary states, we introduced an additional instruction before the initial online Ruminative Response Scale (RRS) assessment ( ). This instruction explicitly asked participants to consider how they typically handle negative emotions, not just how they dealt with them in the past week. We decided to do so as per this instruction because we explicitly wanted the participants to fill out the following questionnaire, especially conscientiously, to prevent misunderstandings of the questionnaire instruction as well as careless responses in general. We retained the standard RRS completion instructions both before and after adding this new instruction (see Supplementary material S3). Furthermore, we conducted telephone interviews with all participants before their inclusion in the study. During these interviews, we asked participants about their subjective opinions regarding whether they considered themselves to be low- or high trait ruminators and how they personally defined rumination. This was done to assess the alignment between their self-perceived traits and their RRS scores obtained during the initial online screening ( ). Finally, approximately 1 week before their scheduled laboratory appointments ( ), we administered the RRS as an online questionnaire once again ( ). This was done to check for significant changes between the second assessment ( ) and the initial online screening ( ). If participants were now categorized into the opposite trait group (e.g., low trait ruminators becoming high trait ruminators) or if their second RRS assessment placed them closer to the opposite group’s score range (e.g., low ruminators no longer scoring below the low RRS cutoff and approaching the high RRS cutoff), they were excluded from the study before their first laboratory appointment. These same criteria were applied when assessing the RRS at the laboratory ( ). Applying these rules to the already assessed participants, a total of seven participants substantially changed RRS scores between and . We wrote an email to all of the aforementioned participants in order to get feedback and assess the RRS again (post-hoc RRS) to check for the “real” group. Four out of the seven participants reported having answered the RRS at the lab regarding the current state. According to this post-hoc RRS, these participants were categorized as high trait ruminators, which was in line with their first assessment at . Two of the participants being asked about the category change, however, subsequently withdrew from the study. One participant was further excluded prior to the first appointment at the lab as she already filled in and changed from high to medium but closer to low. Additionally, seven participants declined to participate prior to their first appointment due to loss of interest, and one declined to participate after the first lab appointment as participation was too stressful, which resulted in 11 dropouts and n = 41 completers for sample S1 (see Figure 1).
Figure 1. CONSORT diagram of the recruitment procedure. S1 = sample recruited prior to changing the recruitment procedure. S2 = sample recruited after changing the recruitment procedure.
After changing the recruitment procedure, another 68 participants were recruited (sample S2). The aforementioned rules resulted in five participants being excluded after , and two participants being excluded after . Out of the five, one participant was accidentally excluded by the study team after T2 due to a false calculation of the sum score, whereas there was no substantial change in RRS scores. Furthermore, a total of nine participants declined to participate prior to their first appointment at the lab due to a loss of interest in participation; four declined to participate due to the experimental session (stress induction or TBS); and two were excluded due to circulatory problems during the session. This resulted in a total of 22 dropouts and 46 completers. Considering only participants never changing categories, the final total sample (sample S3) resulted in n = 82 participants. For an overview of the recruitment procedure, different samples, and exclusion of participants, please find the CONSORT diagram (Moher et al., 2009) in Figure 1. For a more detailed description and illustration of changes in RRS scores, see the Results section, “Number of participants changing categories”.
Ruminative Response Scale
In order to assess inter-individual levels of trait rumination, the self-report Ruminative Response Scale (RRS) (Nolen-Hoeksema and Morrow, 1991), a subscale of the Response Style Questionnaire (RSQ) (Nolen-Hoeksema and Morrow, 1991), was used. The RRS consists of a total of 22 items, which are rated on 4-point Likert scales ranging from 1 = “almost never” to 4 = “almost always” and resulting in a total score ranging between 22 and 88 and consequently a mean ranging between 1 and 4. A high internal consistency has been observed in several studies and samples (Cronbach’s α > 0.88) (Nolen-Hoeksema and Morrow, 1991; Just and Alloy, 1997; Kasch et al., 2001; Moberly and Watkins, 2008), including studies using the German version of the RRS (Cronbach’s α = 0.89–0.92) (Wahl et al., 2011). Test–retest reliability, however, has been proven to fluctuate across different time spans as well as clinical and non-clinical samples: in case of non-clinical samples, test–retest reliability typically ranges between r = 0.80 over 6 months (Nolen-Hoeksema et al., 1994) and r = 0.67 over 1 year (Nolen-Hoeksema et al., 1999). In clinical samples, test–retest scores ranged between r = 0.36 over 6 months (Kasch et al., 2001) and r = 0.47 over 1 year (Just and Alloy, 1997). The original version of the Ruminative Response Scale (RRS) by Nolen-Hoeksema and Morrow (1991) was initially developed to assess depressive rumination. Treynor and Gonzalez (2003) revised the original scale due to high confounding with depression symptoms. Therefore, we used the revised scale. The psychometric properties were found to be comparable to the original RRS (Schoofs et al., 2010; Hasegawa, 2013).
State rumination questionnaire
We assessed stress-reactive rumination using adapted items from the RRS (Nolen-Hoeksema and Morrow, 1991), the Amsterdam Resting-State Questionnaire (Diaz et al., 2013), and the Perseverative Thinking Questionnaire (Ehring et al., 2011), as well as a questionnaire by de Jong-Meyer et al. (2009). The 18 items were answered using a 5-point Likert scale ranging from 1 (“not at all”) to 5 (“very often”), totaling a score between 18 and 90. Subjects were instructed to rate the items regarding the last 10 min. This questionnaire was used in our group in several studies (Rosenbaum et al., 2021; Laicher et al., 2023), and internal consistency has proven to be high (Cronbach’s α > 0.94) (for a full list of items, see Supplementary material S4).
Data analysis
In order to assess the effectiveness of our changes in the recruitment procedure, the following analyses are based on three samples. Sample S1: This sample comprises all participants who were recruited prior to the changes, excluding dropouts (n = 41), as well as participants who were excluded from the total sample later on due to changing RRS categories. Sample S2: Analogously, this sample comprised all participants who were recruited after changing the recruitment procedure without participants declining to participate (n = 53). We included all data entries of all participants until they were excluded, as in the case of the seven participants excluded by the study team. Sample S3: This sample comprised all completers that were not excluded as they never changed RRS categories.
We first report demographic variables as well as descriptive information on the mean number of days between RRS assessments. Next, we report their psychometric properties for each sample. To assess reliability, we report Cronbach’s ɑ as an index of internal consistency and the correlation of scale scores and intraclass correlations (ICC) as an index of test–retest reliability. We interpret the ICC concerning potential measurement bias (Liljequist et al., 2019) using the cutoffs proposed by Koo and Li (2016).
Finally, we fitted logistic regression models (RRS category change yes vs. no) in order to predict changes in RRS scores dependent on the defined cutoffs as described earlier. For this, we used the data of all participants assessed in the study with (n = 12) and without category changes (n = 82), while we abstained from the analysis of each sample as changing categories was a rather uncommon phenomenon and would result in fairly unequal samples. All analyses were also repeated using the R package JTRCI (Kruijt, 2023) in order to calculate Jacobson–Truax and reliable change indices and category changers identified using reliable change in contrast to the aforementioned cutoffs. However, this analysis did not yield systematically different results, which is why we included it in Supplementary material S5.
For the logistic regression models, we investigated different parameters. The report on the results is structured accordingly. First, we investigate the effect of time by fitting a model using the number of days between and as a predictor. Secondly, we investigated the predictive value of an index implemented in SPSS, the Anomaly Case Index List, which reflects the unusualness of a record with respect to the group deviation it belongs to, which is determined using cluster analysis. In a second step, we again descriptively analyzed which participants changing categories were detected according to the use of the cutoff score that was suggested by the authors of this algorithm (remove cases with an index >2). Thirdly, we investigated the predictive value of indexes of careless responses described in Curran (2016), namely the longest string of identical consecutive responses, the Intra-individual Response Variability (the standard deviation of responses across a set of consecutive item responses for an individual, i.e., between the items of the RRS), as well as the Mahalanobis distance of each participant’s RRS ratings at the respective assessment ( vs. vs. ). All of the aforementioned indexes (Anomaly Case Index, longest string, Intra-individual Response Variability, and Mahalanobis distances) are calculated by entering raw data of all RRS items per participant and, as a consequence, identifying anomalies on item level rather than total score level. Again, as a second step, we descriptively analyzed which category changers were detected as multivariate outliers by comparing the Mahalanobis distance to the critical quantile of the corresponding distribution. Due to the absence of meaningful cutoff scores, we abstained from such analyses in the case of the longest string and Intra-individual Response Variability. Fourthly, we analyzed response time indexes that are automatically implemented in the online tool (SoSci Survey) we were using for the assessment of and . This was, on the one hand, an index penalizing extremely fast completion (DEG_TIME), which is normed such that values of more than 100 points indicate low-quality data. On the other hand, the relative speed index, as described in Leiner (2019), is computed. After fitting the logistic regression model, we descriptively investigated the identification of category changers dependent on the cutoff scores proposed by the authors (time index ≥50 and relative speed index >2). The two time indexes are calculated using the time it took the subjects to complete the respective online questionnaire.
Fifthly, we investigated whether participants changed categories because they answered the RRS with respect to their momentary rumination, which is why we used baseline state rumination scores at and the interaction with RRS means at vs. vs. as a predictor in our logistic regression model.
Finally, in order to evaluate the effectiveness of the changes we made in the recruitment procedure, we used test of homogeneity (RRS category change between and yes vs. no prior to vs. after changing the recruitment procedure).
Data analysis was done using SPSS and R (R Core Team, 2023) and RStudio (RStudio Team, 2023) using the packages psych (Revelle, 2023a) and psychTools (Revelle, 2023b). We further used the packages ggplot2 (Wickham, 2016), ggThemes (Arnold, 2021), ggExtra (Attali and Baker, 2022), and networkD3 (Allaire et al., 2017) for plotting.
Results
Demographic data and descriptive information
The sample prior to changing the recruitment procedure (S1, n = 41) comprised 75.61% females; the sample after changing the recruitment procedure (S2, n = 53) comprised 84.91% females; and the final total sample (S3, n = 82) 78.05%. The average age of the samples was M = 24.93 years (SD = 5.38) for S1, M = 23.60 years (SD = 4.31) for S2, and M = 24.40 years (SD = 4.94) for S3. Overall, approximately 8.5 weeks (M = 59.70 days, SD = 31.48 days) after the online screening ( ) was completed, participants had their appointment at the lab ( ). About 74.62 days (SD = 27.54) passed between the first ( ) and the second online assessment of the RRS ( ) and on average 5.68 days (SD = 3.62) between the second online assessment of the RRS ( ) and the appointment at the lab ( ). Samples prior to and after changes in the recruitment were comparable concerning the sex distribution, (1) = 0.348, p = 0.555, as well as their age, F(1, 86) = 1.660, p = 0.201, = 0.019. However, they differed concerning the average number of days between and , F(1, 86) = 69.850, p < 0.001, = 0.448, which was reflected by an average 37.27 days (SD = 15.43) between and in the case of the sample prior to changing the recruitment procedure (S1) compared to M = 79.28 days (SD = 28.76) between and in case of the sample recruited after the changes (S2). Concerning RRS scores, both samples were comparable concerning the distribution of mean RRS scores at , F(1, 86) = 1.298, p = 0.258, = 0.015, however, not concerning RRS scores at , F(1, 86) = 6.481, p < 0.05, = 0.070. This was reflected by overall higher RRS scores in the case of the sample after changing the recruitment procedure (M = 2.14, SD = 0.70) compared to prior (M = 1.81, SD = 0.52) (for an illustration of mean RRS scores dependent on group (low vs. high RRS) and sample (S1, S2, and S3), please see Figure 2). Further detailed information on demographic data is to be found in Table 1.
Figure 2. Line, boxplots, and marginal distributions of mean RRS total scores dependent on sample (A) Sample S1, prior to changing the recruitment procedure, (B) Sample S2, after changing the recruitment procedure, (C) Sample S3, total sample of completers never changing categories). Transparent dots indicate the raw data. Bold dots indicate the estimated marginal mean of the fitted model (time (T1 vs. T2 vs. Tlab) × group (low RRS vs. high RRS). The lower and upper hinges of the boxplot correspond to the first and third quartiles, respectively. Whiskers extend from the hinge to the largest value, no further than 1.5 × interquartile range.
Table 1. Means and standard deviations of demographic variables and a summary of reliability measures dependent on each sample.
Number of participants changing categories
In the case of sample S1 (n = 41), 23 participants (56.1% of the sample) were categorized as low trait ruminator at and remained low at while eight high trait ruminators (19.5% of the sample) remained high at . Four participants (9.8%) changed from high at to low at , another four participants (9.8%) changed from high at to medium but remained closer to high at while one participant (2.4%) changed from high to medium but closer to low at . Finally, one participant (2.4%) originally categorized as a low trait ruminator changed to medium but remained closer to low. Accordingly, following the rules of recruitment applied in the case of sample S2, five (12.2%) participants changed categories from to .
In the case of sample S2 (n = 53), 21 participants (39.6%) who were high trait ruminators at remained in this group also at and . Analogously, this was the case for 15 (28.3%) low trait ruminators at which remained in this group also at and . Five (9.4%) participants that were formerly grouped as low trait ruminators at fell between the cutoffs but closer to their original group at . Three out of those five changed to the low group again at (5.7%), two remained medium but closer to low (3.8%). Two high trait ruminators (3.8%) remained high at but eventually fell between the cutoffs but closer to high at . Three participants (5.7%) who were formerly grouped as high trait ruminators at fell between the cutoffs but closer to their original group at and remained medium but closer to high. Two (3.8%) further high trait ruminators at fell between the cutoffs but closer to their original group at but were excluded at after falling between the cutoffs but closer to low. One participant (1.9%) was excluded at after changing to the low trait ruminator group albeit being recruited as and remaining a high trait ruminator at . Four (7.5%) participants were excluded after : one participant after changing from low to high, one participant after changing from low to medium but closer to high, one participant after changing from high to medium but closer to low, and one participant after changing from high to low. One high trait ruminator was excluded after due to a false calculation of the sum score whereas there was no substantial change in RRS scores. For an illustration of the changes of RRS means and corresponding groups, see Figure 3.
Figure 3. Sankey diagrams illustrating the changes of mean RRS scores and corresponding groups dependent on sample (A) Sample S1. (B) Sample S2. (C) Sample S3.
In the case of sample S3 (n = 82), where all participants changing categories were excluded, a total of 38 participants (46.3%) remained low trait ruminators throughout all RRS assessments (23 assessed only two times at and , 15 three times at , and ), while this was true for 29 high trait ruminators (35.4%) (eight assessed two times at and , 21 three times at , and ). A total of six participants (7.3%) being categorized as low trait ruminators at changed to medium trait ruminators (but closer to the low compared to the high cutoff) at the next RRS assessment (one assessed two times at and , five assessed three times at , and ). Out of the five being assessed three times, three changed again to the low group, while two remained medium ruminators at . Analogously, seven high trait ruminators at (8.5%) changed from high to medium at the next RRS assessment (four assessed two times at and , three assessed three times at , , and ). All of the three participants being assessed three times remained medium trait ruminators at . Finally, two participants (2.5%) remained high trait ruminators at and but fell in between the cutoffs at .
Predicting change in RRS categories
Time
Fitting our first model using only the number of days between and as a predictor, we found that time between measurements did not predict changing RRS categories ( = −0.004, z = −0.289, p = 0.772).
SPSS Anomaly Index
Next, we fitted three models, each including the Anomaly Case Index generated by SPSS for the respective RRS assessment ( vs. vs. ). While the Anomaly Index of ( = −0.136, z = −0.103, p = 0.918) and the Anomaly Index of did not yield significant predictors ( = 1.883, z = 1.293, p = 0.196), the Anomaly Index of ( = 2.422, z = 1.953, p = 0.051) yielded marginal significance. That means, descriptively, for a one-unit increase in the Anomaly Index of , the odds of changing RRS categories (vs. not) increase by a factor of 11.26. According to the authors of the algorithm, observations with Anomaly Case Indexes >2 should be excluded. Descriptively, this was only the case for the RRS score at of one participant (out of a total of eight participants changing categories between and ) who was excluded at the lab after changing categories between (high), (medium but closer to high) and (medium but closer to low). For crosstables of the absolute and relative frequency participants changing and not changing categories were flagged, see Table 2.
Careless package
Next, we used indexes of the careless package implemented in R. Entering the corresponding indices of the respective RRS assessments ( vs. vs. ) separately and as an interaction with each other did not yield any significant predictor. When fitting logistic regression models using the Mahalanobis distances of the respective RRS assessment ( vs. vs. ), we did not observe any significant predictors (all p’s > 0.326). Next, we investigated the Mahalanobis distances descriptively: Comparing RRS ratings at with the critical quantile for ɑ = 0.05, 14 out of 95 participants were identified as multivariate outliers. Out of these 14, two participants out of a total of 12 category changers were identified [one changing between and (out of a total of four changing; 25% identified); one changing between and (out of a total of eight participants changing; 12.5% identified)]. Concerning, RRS ratings at 4 out of 54 participants were identified as multivariate outliers. Out of these four, two participants (out of a total of four participants changing categories between and ) were identified as multivariate outliers (p < 0.05). One of the two participants has already been identified as a multivariate outlier according to the Mahalanobis distance of RRS ratings at . The other one was also identified by the Anomaly Index of SPSS at . For RRS ratings at , 12 out of 95 participants were identified as multivariate outliers (p < 0.05). Out of those 12, two participants changing categories between and were identified whereas one of those was the participant already identified as a multivariate outlier concerning his RRS ratings at and (see Table 3).
SoSci survey response time indexes
Finally, we fitted logistic regression models using the time indexes of the SoSci Survey (consequently, they were only available for the online assessments at and ). Again, we entered each predictor on its own and as an interaction of both predictors; however, we observed no significant effects (all p’s > 0.140). We first investigated the time index and the proposed cutoff of time index ≥50. This resulted in ratings of 15 participants out of 95 in total being flagged. Out of these 15, three participants changed categories: Two changed categories between and while in one of them ratings were not available as this participant was assessed prior to the changes in recruitment procedure. Those two participants were also identified by the Mahalanobis distances and one of them also using the SPSS Anomaly Index. In the case of ratings, 4 out of 54 participants were flagged, however, none of them changed categories at any time. When using proposed cutoff score of the relative speed index (relative speed index >2), one participant’s ratings and one participant’s ratings were flagged, however, none of them were category changers (see Tables 4, 5).
Baseline state rumination
In order to investigate a potential bias in RRS ratings due to a confound with current state rumination, we finally fitted logistic regression models using the interaction of RRS scores at vs. vs. with baseline state rumination ratings at . As a result, we observed marginally significant effects in the case of the main effect of RRS score at ( = 8.282, z = 1.920, p = 0.055) and the interaction effect of RRS score at and state rumination ( = −3.935, z = −1.653, p = 0.098). For every one-unit change in RRS scores at , the log odds of changing categories versus not increases by 8.28, and descriptively, the effect of RRS scores on the odds of changing categories versus not decreases in case of higher state rumination levels at the lab (and the other way around: the effect of state rumination at the lab decreases in case of higher RRS scores at ).
Effectiveness of the changes made in the recruitment procedure
Finally, we investigated whether the distribution of category changers was different prior to vs. after changing the recruitment procedure, which turned out to not be the case, (1) = 1.254, p = 0.263.
Discussion
This investigation aimed to evaluate the predictive value of different indexes for careless responses. Out of all indexes analyzed, Mahalanobis distances seem to be an easy-to-use tool with an acceptable trade-off of sensitivity and specificity that is applicable in most cases.
Scientific progress inevitably requires that findings build on each other. In particular, research in applied subjects often relies on fundamentals that have not yet been fully explored or are currently being investigated. The data presented in this article may serve as a reminder of this fact and bring more awareness to the corresponding consequences. In a recent study of our group, we aimed to investigate the neural correlates of rumination in response to social stress and the impact of Theta-Burst Stimulation over the left dorsolateral prefrontal cortex. For this, participants were screened using a trait questionnaire of ruminative thinking [Ruminative Response Scale (RRS); Treynor and Gonzalez, 2003] in order to assign them to two stratified groups (low- vs. high trait ruminators). These groups were assumed to differ regarding their habitual tendency to ruminate; however, administering the same questionnaire again at the beginning of the experimental session yielded significant changes. As trait measures are generally assumed to have a negligible amount of situation-specific variance, we expected more or less concordant RRS scores. It is further assumed that participants are comparable according to the traits indicated by the questionnaire scores. In the absence of any systematic differential item functioning, this implicates that participants having the same total scores in the RRS are comparable according to their tendency to ruminate (DeVellis, 2006). Whereas the philosophical debate concerning a person’s ability to introspectively assess their own habitual tendencies regarding their whole lifetime and the comparability of this introspection across everyone else is out of the scope of this article, these axioms should also be kept in mind as they underlie the aforementioned stratification and recruitment.
Being confronted with this non-negligible amount of variance, we aimed to investigate potential predictors of participants substantially changing in trait measures and report practical implications for other researchers. As we became aware of this problem throughout the data collection of the aforementioned study, we changed the recruitment procedure henceforth as follows: We introduced more “manipulation checks” for consistent answering in order to eliminate potential misunderstandings. For instance, by adding an extra disclaimer to the online screening stating that participants should think about how they typically handle negative emotions and not just regarding, for instance, the last 2 weeks, there is a chance of participants being more aware of this questionnaire aiming to assess a habitual tendency, a trait, rather than a momentary state. We further checked for misunderstandings of the term “rumination” using a telephone interview and introduced another RRS online assessment 1 week prior to the lab appointment. Unfortunately, we changed several aspects of the recruitment and experimental procedures at once, which is why we are unable to differentiate the efficacy and impact of each change regarding the reduction of careless responses. We did this because we attempted to induce a strong reduction in careless responses as much as possible. In the following, we report our results dependent on samples prior to and after changing the recruitment procedure, as well as the total sample.
We first analyzed the reliability of the different RRS assessments in order to investigate potential conspicuities compared to the psychometric properties reported in the literature. While Cronbach’s ɑ was nearly the same across the RRS assessments and samples, test–retest reliability seemed to substantially differ. While equal ICC values dependent on ICC type indicate the absence of systematic error, which was the case when we analyzed the final total sample where participants changing categories were excluded, there were different ICC values indicating systematic error, especially in the case of the sample assessed prior to changing the recruitment procedure.
As a next step, we investigated predictors of participants whose RRS scores substantially changed within the weeks between the different assessments by fitting logistic regression models (category change: yes vs. no). The number of days, i.e., the sheer amount of time between assessments, had no significant effect on whether participants switched categories, which is consistent with the assumption that the RRS claims to measure a trait that should remain relatively stable over time. Considering the frequency and number of iterations used in our case, this also implicates a negligible effect of memorizing previous answers.
Next, we investigated different indexes of careless responses provided by the software used to collect the data as well as implemented in the software used to analyze the data. This comprised, for instance, an index of the unusualness of cases according to a cluster analysis, namely the Anomaly Case Index generated by SPSS. Using this index and the proposed cutoff scores, we identified the RRS ratings at the second online assessment of one participant who changed RRS categories. While the sensitivity of this index was really low as a result of the number of participants and not changing RRS categories being flagged, the specificity was rather high. Two other indexes used response times to identify careless responses. Again, both did not yield significant predictors in our fitted logistic regression models; however, using the proposed cutoff scores, ratings of 15 participants out of 95 (~16%) in total were flagged. A third of those indeed changed categories later in the study.
As a last step, we used several indices proposed by Curran (2016) and implemented in the statistical software R (R Core Team, 2023), namely the longest string of identical consecutive responses, the Intra-individual Response Variability, and the Mahalanobis distance of each participant’s RRS ratings at the respective assessments. While none of these indexes yielded significant predictors of category change, Mahalanobis distances were able to identify several participants changing categories. Albeit rather many participants were flagged in general (~14% of all participants’ RRS ratings at , ~7% of all participants RRS ratings at and ~ 12% of all participants RRS ratings at ), up to 50% of all participants changing categories after the respective RRS assessment were identified. Especially in the light of the aforementioned results of the other indexes proposed in the literature, Mahalanobis distances seem to be an easy-to-use tool with an acceptable trade-off of sensitivity and specificity that is applicable in the case of paper–pencil as well as online questionnaire data and further does not require the use of a commercial software.
As several participants changing categories qualitatively reported to have answered the RRS as a state measure, we further conducted an analysis investigating the predictive value of baseline state rumination, which was assessed shortly after participants completed the RRS at the laboratory session. We observed a marginally significant main effect of RRS score at and interaction of RRS score at and state rumination, which suggests that there is most probably not only one underlying cause of the observed changes in RRS scores. Presumably, some participants reliably answer the RRS as a state measure, while others show careless responses to some degree, and others might show distortions due to other reasons. As we observed a significant impact of state rumination, future studies should further assess state rumination concurrently with trait rumination in order to disentangle these interrelationships. The precise reasons for these substantially differing RRS scores remain unclear. And while we investigated some potential predictors in this analysis, our sample size is too small to give an answer to the question of the source of the variability. What these analyses, however, are able to show is the need for predetermined rules, a “manipulation check,” in order to assess variation in the data regardless of the cause. Like this, corresponding participants can be excluded prior to their lab appointment, and results are more trustworthy compared to data where underlying assumptions have never been tested. Ultimately, we cannot assume participants not changing categories to be “better” or “more reliable” only because there is less intra-person variability. Most probably, additional qualitative data could be a valuable source of information concerning the motivation of participants (e.g., asking about thoughts during resting/waiting phases). While some participants might report thinking about the aim and method of the study, there might also be first hints to participants reporting not caring and “doing it all solely for the monetary compensation.” We asked participants about these changes in their RRS scores, and while some reported having answered the RRS as a state and their ratings being tinted by their current bad mood, others have reported that they have not “learnt the questionnaire by heart.” Specially to prevent participants from answering the RRS as a state by accident, we introduced an instruction to conscientiously read the following instructions and complete the questionnaire. In addition to potentially increased data quality, this however, might also have unintended consequences, like, for instance, participants taking longer to complete the questionnaire, which in turn has an impact on the analyzed time indexes generated by the SoSci Survey. Furthermore, participants thinking about their answers more thoroughly might make their answers incomparable to those of participants who also completed the questionnaire truthfully and in accordance with the instructions, but rather intuitively. It would be an interesting endeavor to estimate the credibility of the data assessed post-hoc using, for instance, the Randomized Response Technique (Lensvelt-Mulders et al., 2005); however, this is beyond the scope of this article.
One major limitation of the current investigation is the sample size. Although comparable to other studies in psychological research, the rather seldom occurrence of category-changing participants (i.e., little variance to be explained) and consequently different sample sizes (category change vs. no category change) is most probably one reason for problems in fitting our logistic regression models. The distinction of “category change” further followed a rather arbitrary cutoff and the reliable change index, while numeric changes of questionnaire scores might also be of particular interest. As this investigation, however, resulted from a stratified sampling, we found this approach the most appropriate. The assessment of two extreme groups (low and high ruminators) rather than a larger sample also including medium ruminators, however, prevents conclusions that can be drawn for the moderate group into which most of the normal population falls. Another limitation to keep in mind is that simultaneous changes were made to recruitment procedures, instructions, and other aspects, making it challenging to pinpoint the exact cause of the variations in result stability. When evaluating the efficacy of the changes made, we found no significant differences in the distribution of category changers prior to and after changing the recruitment procedure. Future research should systematically alter individual aspects to isolate their effects, explore alternative and potentially more efficient adjustments, and consider potential differential impacts resulting from the interplay of various modifications. A last point to consider is further the variance induced via the form of assessment: face-to-face and online assessments of questionnaires each come with their own set of advantages and disadvantages. Especially important in the context of the current investigation is that the former allows for clarifications and a controlled environment, while online assessments, for instance, offer metadata on completion time (Evans and Mathur, 2018). For a comprehensive discussion of the advantages and disadvantages of administration format with a focus on personality questionnaires, see Hertel et al. (2002). This study, however, was not primarily designed to systematically investigate inattentive responses but to report the usability of several indexes of careless responses using a typical clinical dataset in psychological science and give practical implications for other researchers. Further studies concerning the identification of careless responses are desperately needed. Moreover, scientists and readers should be aware of the underlying assumptions that are made using, for instance, a single “trait” questionnaire in order to stratify groups that are believed to differ concerning their responses. Recent investigations aimed at evaluating the extent of trait and state aspects in well-known questionnaires such as Spielberger’s state and trait anxiety and anger scales (Lance et al., 2021) found state and trait assessments were both dominated by stable trait-like variance. This gives rise to reopening the discussion of shared and unique variance of state and trait measures in general, as well as the need for corresponding scrutinizing of well-established measures. New tools, such as ecological momentary assessments of a variable of interest, might be used to screen participants prior to study inclusion for a more reliable differentiation of traits and states. One promising approach might also be latent state–trait theory (Steyer et al., 1992; Geiser et al., 2017), which has been proven to be superior to classical test theory in individual change detection (Jabrayilov et al., 2016).
Summing it up, the authors would like to raise awareness about the topic of careless responses, which many researchers are not thoroughly aware of when it comes to their own data. We would like to emphasize the importance of manipulation checks using, for instance, multiple assessments of questionnaires and different data sources (e.g., qualitative data, interviews prior to study inclusion). While it may not be feasible for all studies to implement manipulation checks in the first place, post-hoc investigation of potential careless responding is always applicable (DeSimone and Harms, 2018). Following our results, we recommend conducting Mahalanobis distance analyses to identify multivariate outliers, allowing researchers to assess the credibility of their results as the comparability of studies with different and undetected relative frequencies of reliable data is most probably limited and might also be one cause of failed replications (Maniaci and Rogge, 2014).
Data availability statement
The raw data supporting the conclusions of this article will be made available upon request by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Ethics Committee at the University Hospital and University of Tübingen. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
II-V: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing. A-CE: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing. AF: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing. DR: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Acknowledgments
We acknowledge support from the Open Access Publication Fund of the University of Tübingen.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2024.1368390/full#supplementary-material
References
Allaire, J. J., Gandrud, C., Russell, K., and Yetman, C. J. (2017). networkD3: D3 JavaScript network graphs from R. Available at: https://CRAN.R-project.org/package=networkD3
American Psychological Association (n.d.). Trait. APA dictionary of psychology. Available at: https://dictionary.apa.org/trait (Accessed January 25, 2023).
Arnold, J. B. (2021). ggthemes: Extra Themes, Scales and Geoms for “ggplot2.” Available at: https://CRAN.R-project.org/package=ggthemes
Attali, D., and Baker, C. (2022). ggExtra: add marginal histograms to “ggplot2”, and more “ggplot2” enhancements. Available at: https://CRAN.R-project.org/package=ggExtra
Curran, P. G. (2016). Methods for the detection of carelessly invalid responses in survey data. J. Exp. Soc. Psychol. 66, 4–19. doi: 10.1016/j.jesp.2015.07.006
de Jong-Meyer, R., and Parthe, T.Projektgruppe (2009). Einfluss von Achtsamkeitsübung und Dezentrierung auf Rumination und Spezifität autobiographischer Erinnerungen. Z. Klin. Psychol. Psychother. 38, 240–249. doi: 10.1026/1616-3443.38.4.240
DeSimone, J. A., and Harms, P. (2018). Dirty data: the effects of screening respondents who provide low-quality data in survey research. J. Bus. Psychol. 33, 559–577. doi: 10.1007/s10869-017-9514-9
DeVellis, R. F. (2006). Classical test theory. Med. Care 44, S50–S59. doi: 10.1097/01.mlr.0000245426.10853.30
Diaz, B. A., Van Der Sluis, S., Moens, S., Benjamins, J. S., Migliorati, F., Stoffers, D., et al. (2013). The Amsterdam resting-state questionnaire reveals multiple phenotypes of resting-state cognition. Front. Hum. Neurosci. 7:446. doi: 10.3389/fnhum.2013.00446
Ehring, T., Zetsche, U., Weidacker, K., Wahl, K., Schönfeld, S., and Ehlers, A. (2011). The perseverative thinking questionnaire (PTQ): validation of a content-independent measure of repetitive negative thinking. J. Behav. Ther. Exp. Psychiatry 42, 225–232. doi: 10.1016/j.jbtep.2010.12.003
Evans, J. R., and Mathur, A. (2018). The value of online surveys: a look back and a look ahead. Internet Res. 28, 854–887. doi: 10.1108/IntR-03-2018-0089
Geiser, C., Götz, T., Preckel, F., and Freund, P. A. (2017). States and traits: theories, models, and assessment. Eur. J. Psychol. Assess. 33, 219–223. doi: 10.1027/1015-5759/a000413
Hasegawa, A. (2013). Translation and initial validation of the Japanese version of the ruminative responses scale. Psychol. Rep. 112, 716–726. doi: 10.2466/02.08.PR0.112.3.716-726
Hertel, G., Naumann, S., Konradt, U., and Batinic, B. (2002). “Personality assessment via Internet: Comparing online and paper-and-pencil questionnaires,” in Online social sciences, 115–133.
Jabrayilov, R., Emons, W. H. M., and Sijtsma, K. (2016). Comparison of classical test theory and item response theory in individual change assessment. Appl. Psychol. Meas. 40, 559–572. doi: 10.1177/0146621616664046
Johnson, J. A. (2005). Ascertaining the validity of individual protocols from web-based personality inventories. J. Res. Pers. 39, 103–129. doi: 10.1016/j.jrp.2004.09.009
Jones, A., Earnest, J., Adam, M., Clarke, R., Yates, J., and Pennington, C. R. (2022). Careless responding in crowdsourced alcohol research: a systematic review and meta-analysis of practices and prevalence. Exp. Clin. Psychopharmacol. 30, 381–399. doi: 10.1037/pha0000546
Just, N., and Alloy, L. B. (1997). The response styles theory of depression: tests and an extension of the theory. J. Abnorm. Psychol. 106, 221–229. doi: 10.1037/0021-843X.106.2.221
Kasch, K. L., Klein, D. N., and Lara, M. E. (2001). A construct validation study of the response styles questionnaire rumination scale in participants with a recent-onset major depressive episode. Psychol. Assess. 13, 375–383. doi: 10.1037/1040-3590.13.3.375
Kirschbaum, C., Pirke, K. M., and Hellhammer, D. (1993). The ‘Trier social stress test’ – a tool for investigating psychobiological stress responses in a laboratory setting. Neuropsychobiology 28, 76–81. doi: 10.1159/000119004
Koo, T. K., and Li, M. Y. (2016). A guideline of selecting and reporting Intraclass correlation coefficients for reliability research. J. Chiropr. Med. 15, 155–163. doi: 10.1016/j.jcm.2016.02.012
Kruijt, A.-W. (2023). JTRCI: obtain and plot Jacobson-Truax and reliable change indices. Available at: http://awkruijt/JTRCI/
Laicher, H., Int-Veen, I., Woloszyn, L., Wiegand, A., Kroczek, A., Sippel, D., et al. (2023). In situ fNIRS measurements during cognitive behavioral emotion regulation training in rumination-focused therapy: a randomized-controlled trial. NeuroImage Clin. 40:103525. doi: 10.1016/j.nicl.2023.103525
Lance, C. E., Christie, J., and Williamson, G. M. (2021). Do state and trait measures measure states and traits? The case of community-dwelling caregivers of older adults. Assessment 28, 829–844. doi: 10.1177/1073191119888582
Leiner, D. J. (2019). Too fast, too straight, too weird: non-reactive indicators for meaningless data in internet surveys. Survey Res. Methods, 229–248. doi: 10.18148/SRM/2019.V13I3.7403
Lensvelt-Mulders, G. J. L. M., Hox, J. J., Van Der Heijden, P. G. M., and Maas, C. J. M. (2005). Meta-analysis of randomized response research: thirty-five years of validation. Sociol. Methods Res. 33, 319–348. doi: 10.1177/0049124104268664
Liljequist, D., Elfving, B., and Skavberg Roaldsen, K. (2019). Intraclass correlation – a discussion and demonstration of basic features. PLoS One 14:e0219854. doi: 10.1371/journal.pone.0219854
Maniaci, M. R., and Rogge, R. D. (2014). Caring about carelessness: participant inattention and its effects on research. J. Res. Pers. 48, 61–83. doi: 10.1016/j.jrp.2013.09.008
Meade, A., and Craig, S. (2012). Identifying careless responses in survey data. Psychol. Methods 17, 437–455.
Moberly, N. J., and Watkins, E. R. (2008). Ruminative self-focus, negative life events, and negative affect. Behav. Res. Ther. 46, 1034–1039. doi: 10.1016/j.brat.2008.06.004
Moher, D., Liberati, A., Tetzlaff, J., and Altman, D. G.The PRISMA Group (2009). Preferred reporting items for systematic reviews and Meta-analyses: the PRISMA statement. PLoS Med. 6:e1000097. doi: 10.1371/journal.pmed.1000097
Nolen-Hoeksema, S. (1991). Responses to depression and their effects on the duration of depressive episodes. J. Abnorm. Psychol. 100, 569–582. doi: 10.1037/0021-843X.100.4.569
Nolen-Hoeksema, S., Larson, J., and Grayson, C. (1999). Explaining the gender difference in depressive symptoms. J. Pers. Soc. Psychol. 77, 1061–1072. doi: 10.1037//0022-3514.77.5.1061
Nolen-Hoeksema, S., and Morrow, J. (1991). A prospective study of depression and posttraumatic stress symptoms after a natural disaster: the 1989 Loma Prieta earthquake. J. Pers. Soc. Psychol. 61, 115–121. doi: 10.1037/0022-3514.61.1.115
Nolen-Hoeksema, S., Parker, L. E., and Larson, J. (1994). Ruminative coping with depressed mood following loss. J. Pers. Soc. Psychol. 67, 92–104. doi: 10.1037//0022-3514.67.1.92
Nolen-Hoeksema, S., Wisco, B. E., and Lyubomirsky, S. (2008). Rethinking rumination. Perspect. Psychol. Sci. 3, 400–424. doi: 10.1111/j.1745-6924.2008.00088.x
Papageorgiou, C., and Wells, A. (2004). Depressive rumination: nature, theory and treatment. Available at: http://public.eblib.com/choice/publicfullrecord.aspx?p=738756 (Accessed January 27, 2021).
R Core Team (2023). R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing Available at: https://www.R-project.org/.
Revelle, W. (2023a). Psych: Procedures for psychological, psychometric, and personality research. Evanston, Illinois: Northwestern University Available at: https://CRAN.R-project.org/package=psych.
Revelle, W. (2023b). psychTools: Tools to accompany the “psych” package for psychological research. Evanston, Illinois: Northwestern University Available at: https://CRAN.R-project.org/package=psychTools.
Rosenbaum, D., Hilsendegen, P., Thomas, M., Haeussinger, F. B., Metzger, F. G., Nuerk, H.-C., et al. (2018a). Cortical hemodynamic changes during the Trier Social Stress Test: An fNIRS study. NeuroImage 171, 107–115. doi: 10.1016/j.neuroimage.2017.12.061
Rosenbaum, D., Hilsendegen, P., Thomas, M., Haeussinger, F. B., Nuerk, H.-C., Fallgatter, A. J., et al. (2018b). Disrupted prefrontal functional connectivity during post-stress adaption in high ruminators. Sci. Rep. 8:15588. doi: 10.1038/s41598-018-33777-1
Rosenbaum, D., Int-Veen, I., Laicher, H., Torka, F., Kroczek, A., Rubel, J., et al. (2021). Insights from a laboratory and naturalistic investigation on stress, rumination and frontal brain functioning in MDD: an fNIRS study. Neurobiology Stress 15:100344. doi: 10.1016/j.ynstr.2021.100344
RStudio Team (2023). RStudio: Integrated development environment for R (version 2023.03.1). Boston, MA: RStudio, PBC.
Schoofs, H., Hermans, D., and Raes, F. (2010). Brooding and reflection as subtypes of rumination: evidence from confirmatory factor analysis in nonclinical samples using the Dutch ruminative response scale. J. Psychopathol. Behav. Assess. 32, 609–617. doi: 10.1007/s10862-010-9182-9
Smith, J. M., and Alloy, L. B. (2009). A roadmap to rumination: a review of the definition, assessment, and conceptualization of this multifaceted construct. Clin. Psychol. Rev. 29, 116–128. doi: 10.1016/j.cpr.2008.10.003
Steyer, R., Ferring, D., and Schmitt, M. J. (1992). States and traits in psychological assessment. Eur. J. Psychol. Assess. 8, 79–98,
Treynor, W., and Gonzalez, R. (2003). Rumination reconsidered: a psychometric analysis. Cogn. Ther. Res. 27, 247–259. doi: 10.1023/A:1023910315561
Wahl, K., Ertle, A., Bohne, A., Zurowski, B., and Kordon, A. (2011). Relations between a ruminative thinking style and obsessive-compulsive symptoms in non-clinical samples. Anxiety Stress Coping 24, 217–225. doi: 10.1080/10615806.2010.482985
Ward, M., and Meade, A. (2018). Applying social psychology to prevent careless responding during online surveys. Appl. Psychol. 67, 231–263.
Ward, M. K., and Meade, A. W. (2023). Dealing with careless responding in survey data: prevention, identification, and recommended best practices. Annu. Rev. Psychol. 74, 577–596. doi: 10.1146/annurev-psych-040422-045007
Wickham, H. (2016). ggplot2: Elegant graphics for data analysis. New York: Springer-Verlag Available at: https://ggplot2.tidyverse.org.
Keywords: rumination, careless responding, psychometric analysis, ruminative response scale, invalid data
Citation: Int-Veen I, Ehlis A-C, Fallgatter AJ and Rosenbaum D (2024) On assessing trait rumination using the Ruminative Response Scale. Front. Psychol. 15:1368390. doi: 10.3389/fpsyg.2024.1368390
Edited by:
Juan Francisco Rodríguez-Testal, Sevilla University, SpainReviewed by:
Carlos Fernandez-Rubio, Provincial Drug Addiction Center, SpainXiang Wang, McGill University, Canada
Copyright © 2024 Int-Veen, Ehlis, Fallgatter and Rosenbaum. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Isabell Int-Veen, aXNhYmVsbC5pbnQtdmVlbkBtZWQudW5pLXR1ZWJpbmdlbi5kZQ==