![Man ultramarathon runner in the mountains he trains at sunset](https://d2csxpduxe849s.cloudfront.net/media/E32629C6-9347-4F84-81FEAEF7BFA342B3/0B4B1380-42EB-4FD5-9D7E2DBC603E79F8/webimage-C4875379-1478-416F-B03DF68FE3D8DBB5.png)
94% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Psychiatry , 10 February 2025
Sec. Psychopathology
Volume 15 - 2024 | https://doi.org/10.3389/fpsyt.2024.1463116
This article is part of the Research Topic The Role of Self-Perception in Mental Health: Current Insights on Self-Esteem and Self-Schemas View all 4 articles
Introduction: The self-referential encoding task (SRET) has a number of implicit measures which are associated with various facets of depression, including depressive symptoms. While some measures have proven robust in predicting depressive symptoms, their effectiveness can vary depending on the methodology used. Hence, understanding the relative contributions of population differences, word lists and calculation methods to these associations with depression, is crucial for translating the SRET into a clinical screening tool.
Methods: This study systematically investigated the predictive accuracy of various SRET measures across different samples, including one clinical population matched with healthy controls and two university student populations, exposed to differing word lists. Participants completed the standard SRET and its variations, including Likert scales and matrix formats. Both standard and novel SRET measures were calculated and compared for their relative and incremental contribution to their associations with depression, with mean squared error (MSE) used as the primary metric for measuring predictive accuracy.
Results: Results showed that most SRET measures significantly predicted depressive symptoms in clinical populations but not in healthy populations. Notably, models with task modifications, such as Matrix Endorsement Bias and Likert Endorsement Sum Bias, achieved the lowest mean squared error (MSE), indicating better predictive accuracy compared to standard Endorsement Bias measures.
Discussion: These findings imply that task modifications such as utilising Likert-response options and the use of longer word lists may enhance the effectiveness of screening methods in both clinical and research settings, potentially improving early detection and intervention for depression.
Screening and monitoring of Depression have traditionally relied on self-reported measures of depressive symptoms, capturing the individual’s current mood but potentially overlooking underlying vulnerability or future risk (1). Self-referential processing (SRP), the processing of information related to one’s self, provides a measure of the individual’s perception of themselves, which is a core feature of depression (2–5). The self-referential encoding task (SRET) has been widely used to measure individuals’ self-schemas. In the SRET, an adjective is presented and participants are required to make binary decisions about whether the word describes themselves. Various measures including number of word endorsements, reaction time to decide, and number of words recalled are then collected to derive information regarding the participant’s emotional self-biases (6, 7).
Numerous studies have explored the associations between SRET measures and depression, including symptoms, longitudinal course, treatment response, relapse and remission (8–11). For instance, Dainer-Best et al. (12) found that depressed individuals tended to endorse more negative words and nondepressed individuals endorsed more positive words. Negative biases in SRP have also been linked to increased risk of recurrent depressive episodes (10), while deficits in SRP, such as slower drift rates in rejecting negative stimuli, have been shown to persist into remission among individuals with depression (8). Clustering approaches based on the SRET have also shown to be useful in subgrouping patients with depressive and anxiety symptoms, providing a novel transdiagnostic framework (13).
Nonetheless, several SRET variables have been inconsistent in their associations with depressive symptoms across different studies. A meta-analysis conducted on endorsement bias variables showed a mixed pattern of results (14): Some studies found that depressed individuals were more likely to endorse negative words (12), while others observed that depressed individuals had greater endorsement of positive words (9). On the other hand, Kiang et al. (15) found no significant differences in the endorsement of positive and negative words among depressed individuals. Nevertheless, those endorsing severe levels of depressive symptoms tended to endorse fewer positive words, suggesting a potential link with symptom severity.
This ambiguity in results extends to other SRET measures. Reaction time (RT), while initially shown to have no significant difference in clinically healthy and clinically depressed populations (16), was found to be slower in depressed individuals when processing self-referential adjectives (14, 17). In addition, several studies enhanced the interpretability of the RT data by incorporating drift rates in their analysis, revealing a significant correlation between mean drift rates for positive and negative words and baseline depression levels. Similarly, recall bias, initially shown to be significant in individuals with recurrent depressive episodes (10), showed no significant association with depressive symptom severity in a study by Dainer-Best et al. (12). These findings complicate the use of endorsement bias, RT and recall bias as reliable markers for depressive symptoms and underscores the complexity of assessing cognitive processes in MDD.
The inconsistencies between studies may be accounted for by population differences, procedural differences such as variations in the word lists used in the different studies and the method of calculating each variable across studies. For example, in Joorman et al.’s study (18) paper, endorsement was calculated by taking the number of words endorsed in a valence category divided by the total number of words endorsed while in LeMoult et al.’s study (10), endorsement was operationalised by taking the number of endorsed words in each valence category, thus introducing methodological inconsistencies.
There is a need to better understand the relative contributions of population differences, word lists and calculation methods to these associations with depression if the SRET is to be clinically translated. In this study, we aim to systematically investigate the predictive validity of the various SRET measures across different samples exposed to differing word lists. The primary aim is to evaluate the relative strength of association with depressive symptoms between SRET-based measures. To this end, we draw together data collected from clinical, community and healthy populations where participants completed a common self-report measure for depressive symptoms and a self-referential encoding task encompassing a range of measures. The self-referential encoding task includes a wide range of words encompassing common word lists from other studies such as LeMoult et al. (10), and Frewen and Lundberg (19). Each measure is then compared for its relative and incremental contribution to their associations with depression.
In this study, three primary datasets were utilised for the comparative analysis:
The first dataset (Dataset A) comprised a Self-Referential Encoding Task (SRET) consisting of 60 words, incorporating endorsement data, latencies, recall data, recognition data as well as Matrix and Likert data for a separate set of 200 words. This task was administered to a sample of 188 participants, including 144 patients from the Institute of Mental Health (IMH) and 44 healthy controls. The clinical participants were recruited from IMH outpatient clinics if they had a history of depressive symptoms or were currently experiencing depressive symptoms, while the control group was recruited from the community. Details on the patients’ diagnostic status and other sample characteristics are provided in Supplementary Table 1.
Participants were drawn from three studies conducted at IMH, including 84 IMH patients from the “Understanding the person, exploring change across psychotherapies” (XChange) study, which also included 52 participants from the “Understanding the Person, Improving Psychotherapy: Preventing Relapse by targeting Emotional bias Modulation in PsychoTherapy” (PRE-EMPT) and 34 patients and 18 healthy controls from “The role of cholinergic dysfunction in the progression of depression” (CholDep) study. In the Choldep study, healthy controls were also recruited by word of mouth.
The second dataset (Dataset B) comprised another iteration of the SRET that consisted of 185 words, and has endorsement data, latencies and also components of an Other-Referential Processing (ORP) task. It was administered to 61 participants, who were recruited from the National University of Singapore (NUS) as part of an undergraduate thesis project.
The third dataset (Dataset C) also employed a SRET that consisted of 179 words and was administered to a separate sample of 97 participants, recruited at NUS for another undergraduate thesis project. This dataset included endorsement data, latencies and recall data.
The SRET is a task designed to access an individual’s self-relevant schemas (7) that typically comprises three sequential segments: endorsement, distractor task, and incidental recall.
In the endorsement task, participants judged whether an adjective described them. Participants responded by pressing keys representing “not me” or “me” on a computer keyboard. Reaction times, measured in milliseconds were also recorded.
Following the endorsement task, participants engaged in a five-minute distractor task to minimise interference and memory consolidation of endorsed words. Subsequently, only participants in sample A and C completed an incidental recall task lasting one minute, during which they were prompted to recall the words that have been displayed to them in the endorsement task. Afterward, participants in sample A were also presented with a list of 120 words and asked to indicate whether each word had been presented during the endorsement task. The list comprised 60 words from the endorsement task and 60 distractor words. Finally, sample A participants were presented with a matrix containing 200 words and asked to select which words described them. For each selected word, participants from Dataset A rated how accurately the word described them on a scale of 1-4 (1 = Not at all, 4 = Completely accurate).
The following measures derived from the SRET were calculated to evaluate the observed responses.
Multiple endorsement variables were derived using varying calculation methods found in existing literature.
The Proportion of Negative Words Endorsed and Proportion of Positive Words Endorsed were computed as the number of positive/negative words endorsed divided by the number of positive/negative words presented, respectively.
Endorsement bias was operationalised as the number of positive/negative words endorsed divided by the total number of words endorsed. Additionally, a variable representing the difference between negative and positive biases was calculated by subtracting the positive endorsement bias from the negative endorsement bias.
Two reaction time (RT) variables were analysed. Positive RT bias and Negative RT bias were calculated to assess the differences in reaction times between endorsing and rejecting negative or positive words.
Positive RT bias was calculated by using the formula: Positive RT Bias = (Mean RT of Endorsement of Positive Words - Mean RT of Non-Endorsement of Positive Words)/Average RT Across all Trial Types. Similarly, Negative RT bias was calculated with the formula: Negative RT Bias = (Mean RT of Endorsement of Negative Words - Mean RT of Non-Endorsement of Negative Words)/Average RT Across all Trial Types.
This method of determining RT bias corresponds with the approach employed in previous SRET studies (20).
Only words that were correctly recalled were considered. Negative Recall Bias was calculated by dividing the number of negative recalled and endorsed words by the total recalled and endorsed words. Additionally, the Proportion of Negative Endorsed and Recalled Words to Total Endorsed Words and the Proportion of Positive Endorsed and Recalled Words to Total Endorsed Words variables were also calculated.
Negative Recognition Bias was operationalised as the number of correctly recognised and endorsed negative words divided by total number of words recognised correctly while Positive Recognition Bias was defined as the number of correctly recognised and endorsed positive words divided by total number of words recognised correctly.
Signal Detection Theory (SDT) examines how individuals distinguish meaningful information from “noise” (21). The response bias metric, c, measures an individual’s tendency to respond affirmatively or negatively. In this study, c reflects the tendency of a particiapnt to respond whether or not they have seen the word in the word list of the preceding endorsement task. A value of c < 0 represents a strong bias towards “Yes”, while c > 0 represents a strong bias towards “No” (22). As such, we calculated the measures c+, c- and c+ minus c- using the following formula:
Reference formula (23):
Drift rates (v) for sample A were estimated using the drift diffusion model and it represents the speed and direction of information accumulation, with positive values suggesting a preference for the upper threshold (“Yes”). In the context of SRET, drift rate reflects the efficiency of processing whether negative or positive self-descriptive adjectives describe oneself.
Endorsement data and RT data underwent analysis using FAST-dm software, with parameters determined by the trial size of 60 words. Given the small sample size and expected presence of contaminated data, the Kolmogorov-Smirnov estimation test was chosen to run the program. Previous studies have also explored the association between SRET and depressive symptoms, incorporating drift rates as a measure (12, 24). Two drift rate measures were derived: 1) Drift rates towards negative words (v-), indicating the speed and direction of information accumulation when processing negative stimuli, and 2) Drift rates towards positive words (v+), indicating the speed and direction of information accumulation when processing positive stimuli.
Similar to the endorsement measures, the Proportion of Matrix Negative Words Endorsed was calculated by taking the number of negative words endorsed in the matrix divided by the number of negative words presented in the matrix, and the same calculation was done for Proportion of Matrix Positive Words Endorsed. Another variable, Negative Matrix Endorsement Bias, was calculated by taking the number of negative words endorsed in the matrix divided by total number of words endorsed in the matrix.
For the Likert measures, the Proportion of Likert Sum Negative Words and Proportion of Likert Sum Positive Words were calculated by dividing the sum of the ratings of the negative or positive words endorsed by the total number of negative or positive words presented. The Negative Likert Endorsement Sum Bias was computed by dividing the sum of the ratings of negative words endorsed by the total number of non-zero responses, and the same method was applied to calculate the Positive Likert Endorsement Sum Bias.
Next, the Likert data underwent recoding: responses of 1 were adjusted to 0, while responses of 2 to 4 were recoded as 1. This was to avoid the assumption that there are equal intervals between scale points and to facilitate more intuitive interpretations of the results. Following recoding, the sum of negative and positive words was calculated. The same calculation was done for the Proportion of Likert Negative words and Proportion of Likert Positive words. Negative Likert Endorsement bias was computed by dividing the sum of negative words by the total number of non-zero responses, while Positive Likert Endorsement bias was calculated in a similar manner.
The mean rating of negative words endorsed on the Likert scale was calculated as the Mean Negative Likert Rating Bias. This was calculated by averaging the Likert ratings for all negative words endorsed and dividing by the total number of negative words presented.
As each of the three datasets utilised a unique set of word lists, with overlaps in words between them, we conducted regressions using only the overlapping words to investigate the influence of participant characteristics versus word lists on predictive differences. Across all three datasets, we identified a total of 23 overlapping words.
The Quick Inventory of Depressive Symptomatology (QIDS-16-SR) is a 16-item self-report measure that assesses the 9 criterion domains used to diagnose a major depressive disorder (25) and was a shortened version derived from the Inventory of Depressive Symptomatology (IDS). Participants were asked to rate the severity of each of the 30 symptoms in the preceding seven days on a scale of 0–3, with higher scores indicating greater symptom severity. The total score is calculated by summing 9 of the 30 items. The total score ranges from 0 to 27. Participants were categorised into depressive symptom severity levels based on their total QIDS-16-SR scores: 0–5 (no depression), 6–10 (mild depression), 11–15 (moderate depression), 16–20 (severe depression), and 21–27 (very severe depression).
QIDS-16-SR was shown to have satisfactory psychometric properties: Cronbach’s alpha was.875 for Dataset A,.774 for Dataset B and.837 for Dataset C, displaying good reliability in the current study. QIDS-SR-16 total scores were also highly correlated with IDS-30-SR (r = .96) and HAM-D (r = .86) total scores. Overall, the QIDS-16-SR exhibited excellent psychometric properties, suggesting its utility as a brief assessment tool for depressive symptom severity in both clinical and research settings (25).
We conducted multiple linear regression analyses to examine the associations between the different SRET variables such as negative/positive endorsement bias, negative/positive latency bias with depressive symptoms, while controlling for demographic variables. Every SRET variable was entered as independent variables separately in regression models. Demographic variables, including age, gender, and education level, were included as covariates in the models to account for potential confounding effects.
Regression analyses were performed separately for each dataset (Dataset A, Dataset B, and Dataset C) to explore potential variations in the associations across different samples. R2 values of the regression models will be utilised as a measure to evaluate the relative strength of associations between the SRET variables and depressive symptoms.
Mean square error (MSE) is measured as the average of the squared differences between each of the actual values and the predicted values by a model. The formula for calculating MSE is as follows:
, where yi is the ith instance of the actual value and pi is the predicted value, and n is the total number of values (26).
For regression tasks, MSE evaluates how well the actual data is represented by a model’s predictions, with lower MSE values indicating better model performance. Hence, MSE values were used to select the models as they measure the prediction error of SRET measures (27). In this context, the model with a lower MSE value would indicate a better fit in determining the severity of depressive symptoms.
R2 quantifies the proportion of variance in depressive symptoms that is explained by the different SRET measures. Higher R2 values indicate that a larger proportion of the variance in depressive symptoms can be accounted for by the SRET measures, suggesting stronger associations and provision of better fit, hence allowing for a comparison of the relative predictive power of the SRET variables in explaining variations in depressive symptoms (28). However, as Ford (29) highlighted, R2 can be arbitrarily low even with a well-fitted model and arbitrarily close to 1 with a flawed model, underscoring the decision to rely on additional metrics like MSE to examine predictive ability.
All analyses were conducted using R (30) on RStudio version 12.0.353.
In this study, we compared the SRET variables between three datasets: a clinical population with matched healthy controls (Dataset A) and two university populations (Datasets B and C) using R2 values of endorsement, RT and recall. QIDS-16-SR Symptom Severity was calculated for the participants of each dataset. In Dataset A, the breakdown of depression severity was analysed separately for the patient (N = 144) and healthy control (N = 44) samples. In the patient sample, 6.94% presented with no depression, 24.31% with mild depression, 36.81% with moderate depression, 20.83% with severe depression and 7.64% with very severe depression. In contrast, the healthy control sample had 54.55% of the participants presenting with no depression, 29.55% with mild depression, 4.55% with moderate depression, 2.27% with severe depression and no healthy controls were categorised as having very severe depression. There were 9 participants in Dataset A with missing data. In Dataset B (N = 61), 32.79% reported no depression, 49.18% reported mild depression, 9.84% reported moderate depression and 8.20% reported severe depression. In Dataset C (N = 97), there were 3 participants with missing QIDS-16-SR data. 49.48% of the participants reported no depression, 25.77% reported mild depression, 19.59% reported moderate depression, 1.03% reported severe depression and 1.03% reported very severe depression. The demographic characteristics of participants of each dataset are presented in Supplementary Table 1.
Regression analyses showed that endorsement bias variables and RT bias variables were significantly associated with depressive symptoms only for Dataset A, while no significant associations were found in datasets B and C, representing the non-clinical population. Please refer to Supplementary Tables 2, 3 for the regression models of the effect of endorsement bias and RT bias for full word list on depressive symptoms, across all three datasets and Supplementary Table 4 for the comparison of effects of recall bias variables for Datasets A and C.
This also held true when only calculating measures from the 23 overlapping words between the datasets. Regression analysis conducted on the 23 overlapping words between datasets revealed that the regression models were statistically non-significant when Dataset B and C were combined. Similarly, the models for all seven predictors of Endorsement and RT bias were statistically non-significant for Dataset B and Dataset C. Only the regression models in Dataset A were statistically significant, both when the data were analysed separately for the Patient and Healthy Control subgroups, and when the total sample was analysed. Please refer to Supplementary Tables 5, 6 for the regression models of the effect of endorsement measures and RT measures on depressive symptoms for the 23 overlapping words, across all populations.
In Dataset A alone, regression analysis on the 23 overlapping words showed that Proportion of Negative Words Endorsed had an MSE value of 22.01 (F(8, 166) = 16.73, p <.001, R2 = 0.414) and Negative Endorsement Bias with an MSE value of 22.73 (F(8, 166) = 15.44, p <.001, R2 = 0.394). These two models had the lowest and comparable MSE values. This was followed by the Difference in Endorsement Bias variable with an MSE value of 25.46 (F(8, 159) = 9.36, p <.001, R2 = 0.292), Positive Endorsement Bias with an MSE value of 25.62 (F(8, 159) = 9.16, p <.001, R2 = 0.287) and lastly, Proportion of Positive Words Endorsed with an MSE value of 27.10 (F(8, 166) = 9.13, p <.001, R2 = 0.278). Comparatively, the RT bias variables performed worse than endorsement bias variables, with both variables yielding non-significant p-values. These findings suggest that endorsement bias variables displayed stronger predictive power compared to RT bias variables in Dataset A. Please refer to Supplementary Table 7 for the full regression models.
To investigate the consistent lack of significance observed in Datasets B and C, we conducted an F-test comparing the variance in depressive symptoms between Dataset A and the combined Datasets B and C level of significance and there appeared to be a significant trend towards greater variance in depressive symptoms in Dataset A (F(183, 157) =1.75, p <.001). Subsequent analyses were conducted on the full word list in Dataset A alone as it demonstrated more robust associations with depressive symptoms.
The model with the Negative Endorsement Bias as a predictor had a MSE value of 20.76 and R2 value of 0.447 (F(8, 166) = 19.17, p <.001) in Dataset A. Overall, Negative Endorsement Bias ranked 7th in its MSE values among the other predictors. On the other hand, the model with Positive Endorsement Bias as a predictor had a MSE value of 33.60 and R2 value of 0.105 (F(8, 166) = 2.78, p <.001) and the model with the Difference in Endorsement Bias as a predictor had a MSE value of 29.65 and R2 value of 0.210 (F(8, 166) = 6.30, p <.001). These measures ranked 28th and 24th respectively. The Proportion of Negative Words Endorsed and Proportion of Positive Words Endorsed had a MSE of 22.06 (F(8, 166) = 16.63, p <.001, R2 = 0.412) and 24.92 (F(8, 166) = 12.00, p <.001, R2 = 0.336) respectively, ranking 10th and 15th among the SRET predictors for Dataset A.
The model with Negative RT Bias had a MSE value of 29.41 and a significant R2 of 0.216 (F(8, 166) = 6.55, p <.001) in Dataset A. Positive RT bias had a slightly higher MSE value of 29.44 and a significant R2 of 0.216 (F(8, 166) = 6.52, p <.001) in Dataset A. However, the MSE values for RT biases are outperformed by other predictors such as Negative Endorsement Bias and Matrix and Likert endorsement bias. This suggests that the endorsement bias predictors provide a better fit in predicting depressive symptoms. Overall, Positive and Negative RT Bias ranked 23rd and 22nd respectively in terms of MSE values among the other SRET predictors.
Conversely, the Proportion of Negative Endorsed and Recalled Words to Total Endorsed Words model had a higher MSE value than RT Bias variables of 31.77 and a significant R2 of 0.154 (F(8, 166) = 4.30, p <.001). Similarly for the model with Proportion of Positive Endorsed and Recalled Words to Total Endorsed Words as a predictor, it had a higher MSE value of 33.83 but a lower R2 of 0.099 (F(8, 166) = 2.59, p <.001). Comparatively, Negative Recall Bias had a lower MSE value of 28.39 and a higher R2 value of 0.241 (F(8, 148) = 6.72, p <.001). Similarly to RT Bias, the MSE values of Recall Bias are outperformed by the Endorsement Bias predictors. Overall, Proportion of Negative Endorsed and Recalled Words to Total Endorsed Words model ranked 26th, Proportion of Positive Endorsed and Recalled Words to Total Endorsed Words ranked 29th and Negative Recall Bias ranked 20th. The Proportion of Positive Endorsed and Recalled Words to Total Endorsed Words ranked the lowest out of all SRET predictors.
The model with Negative Recognition Bias had a MSE value of 22.50 (F (8, 165) = 15.84, p <.001, R2 = 0.402), while the model with Positive Recognition Bias yielded an MSE of 26.48 (F (8, 165) = 9.91, p <.001, R2 = 0.296). The models rank 12th and 18th respectively.
Response biases for recognition of endorsed negative words (c-) and recognition of endorsed positive words (c+) yielded a MSE value of 29.94 (F (8, 162) = 5.81, p <.001, R2 = 0.202) and 32.10 (F (8, 161) = 3.87, p <.001, R2 = 0.144) respectively. Notably, the Difference between the two response Biases (c+ minus c-) model yielded the lowest MSE value of 26.13 (F (8, 161) = 10.01, p <.001, R2 = 0.303) of the SDT measures and ranked 17th followed by Response Bias for Recognition of Endorsed Negative Words (c-) at 25th and Response Bias for Recognition of Endorsed Positive Words (c+) that ranked 27th among the other SRET predictors.
Furthermore, drift rates towards negative words and positive words were analysed in Dataset A to understand their relationship with depressive symptoms. Drift rate towards Negative words (v-) yielded a MSE value of 28.72 (F (8, 162) = 6.90, p <.001, R2 = 0.230). Drift Rate towards positive words (v+) had a lower MSE value of 27.32 (F (8, 162) = 8.44, p <.001, R2 = 0.267). The MSE value suggests that Drift Rate towards positive words (v+) provides a slightly better fit for predicting depressive symptoms, compared to Drift rate towards negative words (v-). However, overall, the model with Drift Rate towards positive words (v+) ranked 19th among the other SRET predictors.
Negative Matrix Endorsement Bias in Dataset A reported a MSE value of 17.70 and a R2 value of 0.528 (F(8, 166) = 26.58, p <.001). The Negative Matrix Endorsement Bias model has the second lowest MSE value among all the models tested on Dataset A. The Proportion of Matrix Negative Words Endorsed reported a higher MSE value of 21.96 and a R2 of 0.415 (F(8, 166) = 16.82, p <.001) and the Proportion of Matrix Positive Words Endorsed reported an even higher MSE value of 24.74 and R2 of 0.341 (F(8, 166) = 12.26, p <.001). Overall, the Proportion of Matrix words endorsed variables ranked 9th and 14th respectively. This suggests that the Negative Matrix Endorsement Bias model is a good fit and is one of the best models for predicting depressive symptoms.
Difference in Likert Endorsement Sum Bias ranked first among the SRET predictors with the lowest MSE of 17.51 and the highest R2 value of 0.533 (F (8, 166) = 27.11, p <.001), suggesting that it is the best fit model for predicting depressive symptoms in Dataset A. Comparatively, Difference in Likert Endorsement Bias ranked 5th among the SRET predictors with a MSE of 18.40 (F (8, 166) = 24.66, p <.001, R2 = 0.510). Taken together, these results suggest that Difference in Likert Endorsement Sum Bias, obtained using the sum of Likert responses, is a better predictor than the Difference in Likert Endorsement Bias, where the Likert responses were recoded into a binary variable.
The Negative Likert Endorsement Sum Bias and Positive Likert Endorsement Sum Bias yielded MSE values of 17.92 (F (8, 166) = 25.94, p <.001, R2 = 0.522) and 19.51 (F (8, 166) = 21.91, p <.001, R2 = 0.480) respectively, ranking 4th and 6th among the predictors. In comparison, the Negative Likert Endorsement Bias and Positive Likert Endorsement Bias reported MSE values of 21.93 (F (8, 166) = 16.88, p <.001, R2 = 0.416) and 24.26 (F (8, 166) = 12.98, p <.001, R2 = 0.354) respectively, ranking 8th and 13th among all predictors. When comparing the valenced measures, Negative measures yielded lower MSE values compared to Positive measures. These results also follow the same pattern as the Difference scores, where Likert Endorsement Sum Bias reported lower MSEs compared to Likert Endorsement Bias, with both Likert Scale measures yielding lower MSEs compared to regular and Matrix format. Figure 1 shows the comparison of MSE values between Likert-response variables and binary-coded or binary-response variables.
Figure 1. Scatter plot comparing MSE values of binary and Likert measures for predicting depressive symptoms.
Proportion of Likert Negative Words Endorsed ranked 3rd among the SRET predictors with a MSE of 17.78 and R2 value of 0.526 (F (8, 166) = 26.35, p <.001), suggesting that the model is a good fit for predicting depressive symptoms in Dataset A. The Mean Negative Likert Bias reported a MSE of 16.50 and R2 of 0.410 (F (8, 166) = 16.50, p <.001), ranking 11th overall. The Proportion of Likert Sum Negative Words reported a MSE value of 22.13 (F (8, 166) = 16.50, p <.001, R2 = 0.410) and the Proportion of Likert Sum Positive Words reported a MSE value of 25.19 (F (8, 166) = 11.62, p <.001, R2 = 0.329), ranking 11th and 16th among the SRET variables respectively.
Overall, the MSE values reported from all the models tested in Dataset A are consistent with the R2 values, indicating which models provide the best fit for predicting depressive symptoms. The model with the lowest MSE value was the Difference in Likert Endorsement Sum Bias, followed by Negative Matrix Endorsement Bias and the Proportion of Likert Negative Words Endorsed. Negative Likert Endorsement Sum Bias ranked fourth and the Difference in Likert Endorsement Bias ranked fifth in terms of lowest MSE values.
Conversely, the models with the highest MSE values included Proportion of Positive Endorsed and Recalled Words to Total Endorsed Words, Positive Endorsement Bias and Response Bias for Recognition of Endorsed Positive Words (c+). These models indicated a poorer fit for predicting depressive symptoms compared to those with lower MSE values.
All MSE rankings of each model are noted down in Table 1.
Table 1. Regression analysis of endorsement bias for full word list with depressive symptoms for Dataset A.
The full correlation matrix between Age, QIDS-16-SR scores, and all SRET measures in Dataset A is provided in Supplementary Table 8. For clarity, only the correlations for the best-performing variables within each group are highlighted here. Negative Endorsement Bias showed a strong correlation with Negative Matrix Endorsement Bias (r = 0.88, p <.001), and Negative Matrix Endorsement Bias was most strongly correlated with the Difference in Likert Endorsement Sum (r = 0.98, p <.001). Additionally, Negative Recall Bias was also highly correlated with Difference in Likert Endorsement Sum (r = 0.66, p <.001). Negative RT Bias was highly correlated with Negative Likert Endorsement Bias (r = 0.58, p <.001) while drift rates to positive words (v+) was the most highly correlated with Positive RT Bias (r = 0.66, p <.001).
Given that the endorsement variables are among the most commonly used and strongest predictors of depressive symptoms in the SRET literature, the count of negative and positive words endorsed were entered as predictors in the base regression model. This base regression model yielded a MSE value of 21.92 (F(2, 149) = 52.28, p <.001) and a R2 value of 0.412. Subsequently, each of the remaining variables were individually added to the base regression model, and the change in the MSE values was examined to assess whether additional variables offer incremental predictive validity beyond the count of endorsed words.
The results show that the Negative Matrix Endorsement Bias (Δ MSE = -3.28, Δ R2 = 0.088) and Likert Endorsement Bias variables (Proportion of Likert Negative Words Endorsed (Δ MSE = -3.11, Δ R2 = 0.083), Difference in Likert Endorsement Sum Bias (Δ MSE = -3.01, Δ R2 = 0.081), Negative Likert Endorsement Sum Bias (Δ MSE = 2.70, Δ R2 = 0.072), Difference in Likert Endorsement Bias (Δ MSE = -2.28, Δ R2 = 0.061), Negative Likert Endorsement Bias (Δ MSE =-1.76, Δ R2 = 0.047)) showed the most significant changes in MSE values from the base regression model. A greater decrease in MSE values upon adding each measure to the baseline model suggests that these measures provide substantial improvements in predicting depressive symptoms. Specifically, Negative Matrix Endorsement Bias improved the model’s performance, resulting in a reduction of the MSE value by 3.28 compared to the baseline model. In contrast, the addition of Negative Recall Bias and Positive Recognition Bias to the baseline model did not yield improvements in predictive ability of the model, but instead reported an opposite trend where MSE increased by 0.07 and 0.01 respectively. The corresponding change in R2 was minimal at 0.002, 0.001 and 0.001 respectively. Positive RT Bias, Positive Endorsement Bias, Response Bias for Recognition of Endorsed Positive Words (c+) as well as the simple proportions of negative and positive words endorsed did not yield significant improvements in predictive ability, reporting zero change in MSE and R2, suggesting that none of these predictors significantly enhanced the model beyond the total count of endorsed words.
Please refer to Table 2 for the MSE values as predictors are iteratively added individually to the base model, and the corresponding change in MSE is observed.
Considering the wide range of predictors in our dataset, it was valuable to identify a subset of variables that best explain the variability in depressive symptoms. To achieve this, we employed both forward step regression and backward step regression techniques. These methods iteratively add or remove predictors based on their contribution to the model’s predictive power.
Four models were generated to identify the optimal subset of variables using different criterions: (1) Forward Stepwise Regression with AIC as criterion, (2) Backward Stepwise Regression with AIC as criterion, (3) Forward Stepwise Regression with BIC as criterion and (4) Backward Stepwise Regression with BIC as criterion.
The selected models of each method are presented in Table 3.
The best model is (2) Backward Stepwise Regression with AIC as the criterion as the model has the lowest MSE value of 16.56 and highest R2 of 0.56 (AIC = 880.06, BIC = 913.33). Although both the AIC and BIC value for this model is not the smallest, it is comparable to the BIC values of the other models. The selected predictors for (2) were Negative Matrix Endorsement Bias and Negative Likert Endorsement Bias. Notably, Negative Matrix Endorsement Bias was also included in the (1) Forward Stepwise Regression with AIC as criterion final model and (3) Forward Stepwise Regression with BIC as criterion final model. Taken together, these findings suggest that having a larger number of words may serve as a more robust predictor for depressive symptoms. The implications of these results will be further discussed in the Discussion section.
Regression analysis was conducted to assess the predictive validity of endorsement bias across four different word lists: the 23 overlapping words list, 40 words list from LeMoult’s study, the 60 words list used in Dataset A and the 200 words list used in the matrix for Dataset A. Table 4 presents the effect of different word lists on endorsement bias measures in Dataset A.
The results showed that the MSE value for the 200 words list was the lowest among all the word lists analysed across various measures of Endorsement Bias. Specifically, the 200 words list yielded an MSE value of 17.70 (F (8, 166) = 26.58, p <.001, R2 = 0.528) for Negative Endorsement Bias while the MSE values for the 23 words, 40 words and 60 words were 23.31 (F (5, 169) = 15.78, p <.001, R2 = 0.379), 21.85 (F (5, 169) = 30.32, p <.001, R2 = 0.418) and 19.17 (F (8, 166) = 20.76, p <.001, R2 = 0.447), respectively. For Positive Endorsement Bias and Difference in Endorsement Bias measures, a similar trend was observed as the MSE value of the 200 words list was the lowest compared to the other word lists. For the Positive Endorsement Bias model, the MSE value was 23.44 (F (5,169) = 25.41, p <.001, R2 = 0.376) in the 23-word list, 21.85 (F (5,169) = 30.32, p <.001, R2 = 0.418) in the 40-word list, 33.60 (F (8, 166) = 2.78, p <.001, R2 = 0.105) in the 60-word list and 17.70 (F (8, 166) = 26.58, p <.001, R2 = 0.528) in the 200-word list. Similarly, for the Difference in Endorsement Bias model, the rank order of MSE values mirrored that of the Positive Endorsement Bias model, with MSE values of 23.37 (F (5,169) = 25.78, p <.001, R2 = 0.377), 21.85 (F (5,169) = 30.32, p <.001, R2 = 0.418), 29.65 (F (8, 166) = 6.30, p <.001, R2 = 0.210) and 17.70 (F (8, 166) = 26.58, p <.001, R2 = 0.528) for the 23-word, 40-word, 60-word and 200-word list respectively.
The SRET offers multiple measures for assessing SRP and its associations with depressive symptoms. In this study, we investigated several key measures derived from the SRET and its task variations. These measures included endorsement bias measures, RT measures, drift rates, recall measures, recognition measures, as well as matrix measures and Likert measures. The aim was to determine which of these measures had the highest predictive accuracy in predicting depressive symptoms across different samples.
The comparison of SRET measures across the three distinct datasets revealed a trend: while SRET measures demonstrated significant predictive utility within the clinical and healthy controls population (Dataset A), this association was not observed in the university populations (Datasets B and C). An F-test was then conducted to compare the severity of depressive symptoms between the clinical population and the university population. The results revealed a significant difference in depressive symptom severity, indicating that the clinical population exhibited greater depressive symptom severity compared to the university population. This finding suggests that the SRET may hold greater predictive value within populations already experiencing more severe depressive symptoms, rather than in non-clinical samples. This is corroborated by previous research (14, 31) which found that individuals with major depressive disorder (MDD) exhibit greater negative SRP bias, possessing more negative self-schemas compared to non-depressed individuals. Consequently, subsequent analyses were conducted on Dataset A alone as it demonstrated more robust associations with depressive symptoms. Future studies could explore the efficacy of the SRET in non-clinical samples with larger sample sizes, to further validate the use of the SRET beyond clinical populations.
Within Dataset A, the comparative analysis revealed that endorsement bias emerged as one of the strongest SRET measures in predicting depressive symptoms, ranking seventh overall. Notably, it performed better than the proportion of negative and positive words endorsed. This is corroborated by Dainer-Best et al. (12), whose research indicated that endorsement measures were among the strongest predictors of depressive symptoms. Given the multitude of calculation methods for endorsement data, this finding suggests that utilising endorsement bias may enhance the predictive power of endorsement measures in assessing depressive symptoms.
Although conventional RT bias measures significantly predicted depressive symptoms, our findings suggest that drift rates may be better predictors of depressive symptoms. Previous research indicates that depressed individuals often show a heightened attentional bias towards negative stimuli, which can be hypothesised to manifest as quicker responses to negative words due to increased salience for depressed individuals (32). This is supported by Fritzsche et al. (17) and Collins and Winer (14), who revealed that depressed individuals consistently exhibit slower reaction times than their non-depressed counterparts, indicating a prolonged decision-making process when processing self-referential adjectives. These findings complicate the use of RT as a reliable marker for depressive symptoms and underscores the complexity of assessing cognitive processes in MDD. The drift-diffusion model emerged as a better predictor than conventional SRET measures, such as RT bias, ranking at nineteenth place overall. Furthermore, the drift rates also offered reductions in MSE when incrementally added to the base regression model beyond endorsement measures. Consistent with findings from previous research (12), our study did not find robust associations between recall measures and depressive symptoms. This suggests that recall performance may not be a reliable predictor of depressive symptoms across different populations and methodologies. Further investigations are warranted to better understand the role of recall measures in assessing depressive symptomatology.
In our current study, we expanded the conventional SRET task by incorporating additional components such as a recognition task, presenting the SRET in a matrix format and including a Likert scale for a more nuanced analysis of SRP.
For the recognition task, the use of SDT to analyse participants’ recognition of self-descriptive adjectives unexpectedly did not yield better results as compared to calculating the proportion of correctly recognised and endorsed negative and positive words. Negative Recognition Bias has the lowest MSE value out of the recognition bias measures and was ranked 12th overall. It was calculated as the proportion of negative words that were both correctly recognised and endorsed, divided by the total number of negative words correctly recognised as being presented in the previous list. This metric takes into account both recognition accuracy and endorsement tendencies, which have shown to be key for understanding cognitive biases related to depressive symptom severity.
On the other hand, response bias (c+ or c-) quantifies the extent to which a “yes, the word was presented in the previous list” versus “no, the word was not presented” is more probable for positive or negative words endorsed independent of sensitivity. Hence, response bias measures provide insight into participant’s general decision-making tendencies but does not directly account for accuracy or interaction between recognition and endorsement. Therefore, Negative Recognition Bias, which integrate accuracy and endorsement measures more directly might have performed better at capturing the maladaptive negative self-schemas.
The present findings suggest that the Negative Matrix Endorsement Bias and the Difference in Likert Endorsement Sum bias variables are the most robust SRET measures in terms of their predictive accuracy for depressive symptoms as they have the lowest MSE values and highest R2 values, ranking the top among other SRET measures. Furthermore, when these measures were incrementally added to the base regression model, the MSE consistently decreased, and the R2 value increased, indicating a significant improvement in the model’s explanatory power to depressive symptoms.
The study employed forward and backward stepwise regression models to determine the best subset of measures. Interestingly, Negative Matrix Endorsement Bias consistently appeared in three of the four models, regardless of whether AIC or BIC was used as a stopping criterion, suggesting its robust predictive power across different model selection techniques. Therefore, Negative Matrix Endorsement Bias, which is indicative of a depressotypic processing style (33), might be a reliable marker for screening for depressive symptoms. Consistent with our findings, the Endorsement Bias measures derived from the 200 words list demonstrated superior predictive validity, as evidenced by having the lowest MSE value compared to the other words lists.
These results have several implications to the SRET. First, expanding the word list in the SRET appears to be more beneficial for predicting depressive symptoms. Existing literature has well-documented that depressogenic patterns of self-schemas are associated with concurrent and prospective depressive symptoms in clinical samples of depressed adults (16, 17), adolescents (34), as well as in non-clinical samples of youth (20, 35). By expanding the word list, our study may have captured a broader spectrum of depressogenic self-referential representations, allowing for a richer representation of individuals’ self-schemas, potentially enhancing the sensitivity of the SRET for screening for depressive symptoms (7). On average, each of the task components lasted less than 5 minutes regardless of the length of the word list, suggesting that the 200-word list did substantially increase respondent burden. Furthermore, the expanded word list could help in identifying subtle nuances in self-referential processing that might be missed with a shorter list. Future research should explore the optimal length and content of word lists in SRET to maximise its predictive power. Clinically, using a more comprehensive word list in SRET could improve the accuracy of depressive symptom screening and better inform treatment strategies. Next, the presentation of the SRET stimuli in a matrix format might have facilitated the integration of self-referential information as the adjectives were shown simultaneously. A study by Bharti et al. (36) found that simultaneous presentation affects the organisation of stimuli in the visual working memory differently than sequential presentation and these cognitive processes of memory are closely related to self-referential processing (37, 38). Hence, the matrix format may have allowed for more effective encoding and retrieval of self-related information. Although there are existing studies that have investigated the neural activity associated with SRET (39), future research could explore the neural underpinnings associated with sequential versus simultaneous presentation of the self–descriptive words to better understand these cognitive processes and their impact on self-referential processing.
Utilising a Likert-response format for the SRET demonstrated better predictive validity compared to the traditional binary-response format of the SRET. The Difference in Likert Endorsement Sum Bias variable performed the best among the various SRET measures. While previous research suggests that a binary response format often imposes the least response burden on participants (40), in our study, we found that the Likert-response options allows for more nuanced self-referential judgements. A study conducted with older adults endorsing depressive symptoms tested both Likert response options and binary response options and the authors concluded that binary response options may have increased response burden and cognitive load on their participants as the individual must decide if their experience was sufficient enough to warrant a complete endorsement of the symptom (41). Similarly, in the context of the SRET, binary responses may fail to fully capture the subtleties of the participants’ self-schemas. In our study, administering Likert response options after participants selected words in the matrix may have provided the participants the opportunity to clarify the degree of endorsement, capturing the complexity of the participants’ self-schemas more effectively than binary options. This is supported by the current results where using the full range of Likert scores, rather than collapsing the responses into binary categories produced the best predictive measure, with the Difference in Likert Endorsement Sum Bias emerging as the best model for predicting depressive symptom categories.
Furthermore, as seen in Figure 1, we hypothesise that the differences arising between Likert response categories and binary-coded response categories may be associated with depression. Notably, these differences do not appear to have been adequately captured by novel measures such as SDT or drift rates derived from reaction time (RT) data. The findings suggest that a Likert response format may offer superior sensitivity compared to a binary response format, not only in this study but potentially in other similar tasks. Future research could further explore the utility of Likert-response formats in the SRET and examine whether they confer additional benefits as compared to the traditional binary-response options in capturing SRP.
In summary, our study provides a head-to-head comparison of the predictive accuracy and incremental predictive value between SRET measures in the prediction of depressive symptoms. We find that an expanded word list and a Likert response format rather than binary response format improved the sensitivity of the SRET for depressive symptoms. Self-descriptive judgements without recall or memory measures may be sufficient for prediction of depressive symptoms. The implications of this study are significant for both research and the development of translational applications of the SRET. By refining the SRET with innovative measures and formats, we can improve its efficiency as a potential assessment in depression. Future research should continue to explore the neural and cognitive mechanisms underlying different presentation formats and further optimise the SRET by refining word selection to maximise its predictive accuracy. In conclusion, our study contributes to a deeper understanding of cognitive biases in depression and demonstrates the value of our adaptations to the SRET.
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Both IMH studies, Study Reference Number: 2015/00545, 2018/01184 and 2021/00005, were approved by the IMH Institutional Research Review Committee (IRRC) and NHG Domain Specific Review Board (DSRB). The NUS sample was collected with approval from the NUS Institutional Review Board and Psychology Department Ethics Review Committee (DERC) under Psych-DERC Reference Code: 2018-August-86. The studies were conducted in accordance with the local legislation and institutional requirements. Participants provided written informed consent for participation in this study.
ET: Supervision, Writing – original draft, Writing – review & editing. HT: Methodology, Supervision, Writing – original draft, Writing – review & editing. KF: Writing – original draft. ST: Project administration, Writing – original draft, Writing – review & editing. NR: Project administration, Writing – original draft, Writing – review & editing. CH: Writing – original draft, Writing – review & editing. ZT: Writing – original draft, Writing – review & editing. RO: Writing – original draft, Writing – review & editing. CT: Writing – original draft, Writing – review & editing. JY: Writing – original draft, Writing – review & editing. ML: Writing – original draft, Writing – review & editing. AT: Writing – original draft, Writing – review & editing. SO: Project administration, Writing – original draft. XYL: Project administration, Writing – original draft, Writing – review & editing. JLK: Project administration, Writing – original draft, Writing – review & editing. JK: Supervision, Writing – original draft, Writing – review & editing. GT: Conceptualization, Formal Analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Writing – original draft, Writing – review & editing.
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research is supported by the National Research Foundation, Singapore, and the Agency for Science Technology and Research (A*STAR), Singapore, under its Prenatal/ Early Childhood Grant (Grant No. H22P0M0005). This research is also supported by A*STAR Brain-Body Initiative (BBI) (#21718), IMH Research Seed Funding (642-2018), Ministry of Health (MOH) Clinician Scientist Residency Seed Funding (April 2016 call), LKCMed-NUSMed-NHG Collaborative Mental Health Research Pilot Grant Call 2020 (MHRPG/2003).
We thank the study participants of the Xchange, PRE-EMPT and CholDep studies. This manuscript has been submitted as a preprint (https://doi.org/10.31234/osf.io/6fsu3). The authors thank Poorva Harshang Pandya, Hui Wen Ong, Rui Yi Tang, and Kar Yuan Lee for their contributions to reviewing and editing this article.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2024.1463116/full#supplementary-material
1. El-Den S, Chen TF, Gan Y-L, Wong E, O’Reilly CL. The psychometric properties of depression screening tools in primary healthcare settings: A systematic review. J Affect Disord. (2018) 225:503–22. doi: 10.1016/j.jad.2017.08.060
2. Bentley SV, Greenaway KH, Haslam SA. An online paradigm for exploring the self-reference effect. PloS One. (2017) 12:e0176611. doi: 10.1371/journal.pone.0176611
3. Northoff G, Heinzel A, de Greck M, Bermpohl F, Dobrowolny H, Panksepp J. Self-referential processing in our brain—A meta-analysis of imaging studies on the self. NeuroImage. (2006) 31:440–57. doi: 10.1016/j.neuroimage.2005.12.002
4. Rogers TB, Kuiper NA, Kirker WS. Self-reference and the encoding of personal information. J Pers Soc Psychol. (1977) 35:677–88. doi: 10.1037/0022-3514.35.9.677
5. Symons CS, Johnson BT. The self-reference effect in memory: A meta-analysis. psychol Bull. (1997) 121:371–94. doi: 10.1037/0033-2909.121.3.371
6. Bradley B, Mathews A. Negative self-schemata in clinical depression. Br J Clin Psychol. (1983) 22:173–81. doi: 10.1111/j.2044-8260.1983.tb00598.x
7. Derry PA, Kuiper NA. Schematic processing and self-reference in clinical depression. J Abnormal Psychol. (1981) 90:286–97. doi: 10.1037/0021-843X.90.4.286
8. Allison GO, Kamath RA, Carrillo V, Alqueza KL, Pagliaccio D, Slavich GM, et al. Self-referential processing in remitted depression: an event-related potential study. Biol Psychiatry Global Open Sci. (2021) 3:119–29. doi: 10.1016/j.bpsgos.2021.12.005
9. Disner SG, Shumake JD, Beevers CG. Self-referential schemas and attentional bias predict severity and naturalistic course of depression symptoms. Cogn Emotion. (2017) 31:632–44. doi: 10.1080/02699931.2016.1146123
10. LeMoult J, Kircanski K, Prasad G, Gotlib IH. Negative self-referential processing predicts the recurrence of major depressive episodes. Clin psychol Sci: A J Assoc psychol Sci. (2017) 5:174–81. doi: 10.1177/2167702616654898
11. van Kleef RS, Kaushik P, Besten M, Marsman J-BC, Bockting CLH, van Vugt M, et al. Understanding and predicting future relapse in depression from resting state functional connectivity and self-referential processing. J Psychiatr Res. (2023) 165:305–14. doi: 10.1016/j.jpsychires.2023.07.034
12. Dainer-Best J, Lee HY, Shumake JD, Yeager DS, Beevers CG. Determining optimal parameters of the self-referent encoding task: A large-scale examination of self-referent cognition and depression. psychol Assess. (2018) 30:1527–40. doi: 10.1037/pas0000602
13. Tan GC, Wang Z, Tan ESE, Ong RJM, Ooi PE, Lee D, et al. Transdiagnostic clustering of self-schema from self-referential judgements identifies subtypes of healthy personality and depression. Front Neuroinform. (2024) 17:1244347. doi: 10.3389/fninf.2023.1244347
14. Collins AC, Winer ES. Self-referential processing and depression: A systematic review and meta-analysis. Clin Psychol Sci. (2023) 12(4):721–750. doi: 10.1177/21677026231190390
15. Kiang M, Farzan F, Blumberger DM, Kutas M, McKinnon MC, Kansal V, et al. Abnormal self-schema in semantic memory in major depressive disorder: Evidence from event-related brain potentials. Biol Psychol. (2017) 126:41–7. doi: 10.1016/j.biopsycho.2017.04.003
16. Gotlib IH, Kasch KL, Traill S, Joormann J, Arnow BA, Johnson SL. Coherence and specificity of information-processing biases in depression and social phobia. J Abnormal Psychol. (2004) 113:386–98. doi: 10.1037/0021-843X.113.3.386
17. Fritzsche A, Watz H, Magnussen H, Tuinmann G, Löwe B, von Leupoldt A. Cognitive biases in patients with chronic obstructive pulmonary disease and depression – a pilot study. Br J Health Psychol. (2013) 18:827–43. doi: 10.1111/bjhp.12025
18. Joormann J, Dkane M, Gotlib IH. Adaptive and maladaptive components of rumination? Diagnostic specificity and relation to depressive biases. Behav Ther. (2006) 37:269–80. doi: 10.1016/j.beth.2006.01.002
19. Frewen P, Lundberg E. Visual–Verbal Self/Other-Referential Processing Task: Direct vs. Indirect assessment, valence, and experiential correlates. Pers Individ Dif. (2012) 52:509–14. doi: 10.1016/j.paid.2011.11.021
20. Connolly SL, Abramson LY, Alloy LB. Information processing biases concurrently and prospectively predict depressive symptoms in adolescents: Evidence from a self-referent encoding task. Cogn Emotion. (2016) 30:550–60. doi: 10.1080/02699931.2015.1010488
21. Lynn SK, Barrett LF. Utilizing” Signal detection theory. psychol Sci. (2014) 25:1663–73. doi: 10.1177/0956797614541991
22. Stanislaw H, Todorov N. Calculation of signal detection theory measures. Behav Res Methods Instruments Comput. (1999) 31:137–49. doi: 10.3758/BF03207704
23. Macmillan NA. Signal detection theory as data analysis method and psychological decision model. In: A handbook for data analysis in the behavioral sciences: Methodological issues. Hillsdale, NJ: Lawrence Erlbaum Associates, Inc (1993). p. 21–57.
24. Castagna PJ, Waters AC, Edgar EV, Budagzad-Jacobson R, Crowley MJ. Catch the drift: Depressive symptoms track neural response during more efficient decision-making for negative self-referents. J Affect Disord Rep. (2023) 13:100593. doi: 10.1016/j.jadr.2023.100593
25. Rush AJ, Trivedi MH, Ibrahim HM, Carmody TJ, Arnow B, Klein DN, et al. The 16-Item Quick Inventory of Depressive Symptomatology (QIDS), clinician rating (QIDS-C), and self-report (QIDS-SR): A psychometric evaluation in patients with chronic major depression. Biol Psychiatry. (2003) 54:573–83. https://www.sciencedirect.com/science/article/pii/S0006322302018668 (Accessed July 11, 2024).
26. Tyagi K, Rane C, Harshvardhan, Manry M. Chapter 4—Regression analysis. In: Pandey R, Khatri SK, kumar Singh N, Verma P, editors. Artificial Intelligence and Machine Learning for EDGE Computing. Hoboken, NJ: Academic Press (2022). p. 53–63. doi: 10.1016/B978-0-12-824054-0.00007-1
27. Li J. Assessing the accuracy of predictive models for numerical data: Not r nor r2, why not? Then what? PloS One. (2017) 12:e0183250. doi: 10.1371/journal.pone.0183250
28. Montgomery DC, Peck EA, Vining GG. Introduction to Linear Regression Analysis. John Wiley & Sons (2013).
29. Ford C. Understanding Robust Standard Errors. UVA Library StatLab (2015). Available at: https://library.virginia.edu/data/articles/understanding-robust-standard-errors (Accessed November 21, 2024).
30. R Core Team. R: A Language and Environment for Statistical Computing [Computer software]. R Foundation for Statistical Computing (2022). Available at: https://www.R-project.org/ (Accessed November 21, 2024).
31. Segal ZV, Hood JE, Shaw BF, Higgins ET. A structural analysis of the self-schema construct in major depression. Cogn Ther Res. (1988) 12:471–85. doi: 10.1007/BF01173414
32. Donaldson C, Lam D, Mathews A. Rumination and attention in major depression. Behav Res Ther. (2007) 45:2664–78. doi: 10.1016/j.brat.2007.07.002
33. Duyser FA, Van Eijndhoven PFP, Bergman MA, Collard RM, Schene AH, Tendolkar I, et al. Negative memory bias as a transdiagnostic cognitive marker for depression symptom severity. J Affect Disord. (2020) 274:1165–72. doi: 10.1016/j.jad.2020.05.156
34. Auerbach RP, Stanton CH, Proudfit GH, Pizzagalli DA. Self-referential processing in depressed adolescents: A high-density event-related potential study. J Abnormal Psychol. (2015) 124:233. Available at: https://psycnet.apa.org/record/2015-04301-001 (Accessed July 11, 2024).
35. Liu P, Vandemeer MRJ, Joanisse MF, Barch DM, Dozois DJA, Hayden EP. Depressogenic self-schemas are associated with smaller regional grey matter volume in never-depressed preadolescents. NeuroImage: Clin. (2020) 28:102422. doi: 10.1016/j.nicl.2020.102422
36. Bharti AK, Yadav SK, Jaswal S. Feature Binding of Sequentially Presented Stimuli in Visual Working Memory. Frontiers in Psychology (2020) 11. doi: 10.3389/fpsyg.2020.00033
37. Sajonz B, Kahnt T, Margulies DS, Park SQ, Wittmann A, Stoy M, et al. Delineating self-referential processing from episodic memory retrieval: Common and dissociable networks. NeuroImage. (2010) 50(4):1606–17. doi: 10.1016/j.neuroimage.2010.01.087
38. Sui J, Humphreys GW. The Integrative Self: How Self-Reference Integrates Perception and Memory. Trends in Cognitive Sciences. (2015) 19(12):719–28. doi: 10.1016/j.tics.2015.08.015
39. Auerbach RP, Bondy E, Stanton CH, Webb CA, Shankman SA, Pizzagalli DA. Self-referential processing in adolescents: Stability of behavioral and ERP markers. Psychophysiology. (2016) 53:1398–406. doi: 10.1111/psyp.12686
40. McLauchlan J, Browne M, Russell AMT, Rockloff M. Evaluating the reliability and validity of the short gambling harm screen: are binary scales worse than likert scales at capturing gambling harm? J Gambling Issues. (2020) 44:103–20. doi: 10.4309/jgi.2020.44.6
Keywords: self-schema, self-concept, self-referential processing, personality, depression
Citation: Tan ESE, Tan HM, Fong KV, Tey SYX, Rane N, Ho CW, Tan ZY, Ong RJM, Teo C, Yu J, Lee M, Teo AR, Ong SK, Lim XY, Kee JL, Keppo J and Tan GC-Y (2025) Evaluating the relative predictive validity of measures of self-referential processing for depressive symptom severity. Front. Psychiatry 15:1463116. doi: 10.3389/fpsyt.2024.1463116
Received: 11 July 2024; Accepted: 11 December 2024;
Published: 10 February 2025.
Edited by:
Regina Espinosa Lopez, Camilo José Cela University, SpainReviewed by:
Guanmin Liu, Tianjin University, ChinaCopyright © 2025 Tan, Tan, Fong, Tey, Rane, Ho, Tan, Ong, Teo, Yu, Lee, Teo, Ong, Lim, Kee, Keppo and Tan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Geoffrey Chern-Yee Tan, UHN5Y2hvdGhlcmFweV9SZXNlYXJjaEBpbWguY29tLnNn
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.