Skip to main content

BRIEF RESEARCH REPORT article

Front. Pain Res., 05 August 2024
Sec. Pain Research Methods

Assessing how individuals conceptualize numeric pain ratings: validity and reliability of the Pain Schema Inventory (PSI–6) Short Form

  • Department of Physical Therapy and Rehabilitation Science, Carver College of Medicine, University of Iowa, Iowa City, IA, United States

Background: While numeric scales to represent pain intensity have been well validated, individuals use various conceptualizations when assigning a number to pain intensity, referred to as pain rating schema. The 18-item Pain Schema Inventory (PSI-18) quantifies pain rating schema by asking for numeric values for multiple mild, moderate or severe pain conditions. This study aimed to assess the validity and reliability of a shortened form of the PSI, using only 6 items (PSI-6).

Methods: A secondary analysis was performed on two existing datasets. The first (n = 641) involved a community-based population that completed the PSI-18. The second (n = 182) included participants with chronic pain who completed the PSI-6 twice, one week apart. We assessed face validity, convergent validity, offset biases, test-retest reliability, and internal consistency of the PSI-6 compared to the PSI-18.

Results: Both the PSI-18 and PSI-6 demonstrated excellent face validity. The PSI-6 demonstrated excellent convergent validity relative to the PSI-18, with correlations from r = 0.88 to 0.92. Bland-Altman plots revealed offset biases near zero (< 0.22 on 0–10 scale) across all categories of mild, moderate, severe and average pain. Internal consistency was excellent, with Cronbach's Alpha = 0.91 and 0.80, for PSI-18 and PSI-6 respectively. Test-retest reliability of the PSI-6 was high with correlations from r = 0.70–0.76.

Conclusion: The PSI-6 is a valid and reliable tool to assess pain rating schema with reduced subject burden, to better interpret individuals’ pain ratings and adjust for inter-individual variability.

1 Introduction

Pain is the most widely reported symptom in the medical literature, with an estimated 56% of American adults reporting some pain in the previous 3 months, and an estimated 20% reporting chronic pain (1, 2). Although pain can vary widely, the International Association for the Study of Pain encompasses this range by defining it as, “An unpleasant sensory and emotional experience associated with, or resembling that associated with, actual or potential tissue damage” (3). Multiple methods have been used to assess the subjective nature of pain including unidimensional numerical intensity ratings (48) and multidimensional assessments (9, 10). While multidimensional approaches provide unique insight, unidimensional pain ratings are widely used for their ease, compliance, and responsiveness (7, 8).

The psychometric validity of pain rating assessment has been well documented (46, 8, 1113). However, underlying pain rating schemas can vary across individuals, i.e., the conceptual frameworks that individuals utilize when transforming pain to a number (14). While moderate pain is consistently rated numerically higher than mild pain, and severe pain higher than moderate pain, the actual numerical pain ratings for each level (i.e., mild, moderate, or severe pain) can greatly differ between individuals. This may add challenges when interpreting either clinical or research pain ratings, as what is a “3” to one person may be a “4” to another, and so on. Identifying these differences in underlying pain rating schema may improve the interpretation of individuals’ numeric pain ratings.

More broadly, generalized pain schemas have been described as pain-related information processing biases or internal sources of information (15) as well as “general beliefs, appraisal and expectations about pain” (16). Although challenging to assess, generalized pain schemas, i.e., the internal frameworks that help interpret anticipated or actual painful experiences, have been evaluated using paired word choices (e.g., “excruciating”—“relieving”), sentence completion (e.g., Sentence Completion Test), and both explicit means of assessment, as well as implicit assessment using the Implicit Association Test (IAT) (17, 18). These general assessments of pain schema often consider the multidimensional nature of pain. Some experts may even consider assessments of pain catastrophizing or pain-related fear of movement to be forms, or components, of generalized pain schema, i.e., assessing beliefs, appraisals or expectations about pain (19, 20).

One specific component of the generalized pain schema is pain rating schema, which considers how one represents pain intensity numerically (14). The evaluation of pain rating schema is distinct from the explicit and implicit methods used to assess more generalized interpretive biases, explicitly attempting to better understand how individuals conceptualize pain intensity. While numerical ratings of pain as a sensory and emotional experience have become routine both in the clinic and in research settings, relatively few attempts have been made to quantify the schema underlying these numerical assignments (8).

Numerous studies have examined individual differences in pain sensitivity, frequently utilizing numerical pain ratings as the primary independent or dependent variable. That is, they evaluate for differences in pain ratings for a given, standardized noxious input, or conversely identifying variation in the noxious input needed to achieve a particular pain intensity (7, 21, 22). Either way, the underlying pain rating schema is inherently a core component of how individuals conceptualize their pain numerically. However, this schema is rarely considered or adjusted for, which can lead to inaccurate or misleading inferences when comparing pain ratings between individuals. Similarly, clinical pain science often relies on self-reported pain intensity to study relationships with pain biomarkers, including imaging, psychosocial, and “omics” variables (23). Further, minimum pain rating intensities are frequently used as a criterion for inclusion in clinical trials (24). Yet, few clinical studies consider adjusting for individual variation in the use of numerical pain scales, i.e., pain rating schema.

In one cross-sectional study, generalized pain rating schema were identified as average cut-off values for mild, moderate, and severe pain based on corresponding functional interference scores (25). While it was not able to identify individual-specific pain schema, cut-off values were related to pain catastrophizing, with no sex differences (25). In another study, individuals were asked to rate their anticipated pain ratings for mild, moderate, and severe pain for five different common acute pain conditions (headache, toothache, joint pain, delayed onset muscle soreness, and burns) and pain in general (14). While not originally named, we refer to this survey as the Pain Schema Inventory (PSI). In a mixed cohort of community-dwelling adults, with and without pain at the time of completion, three pain schema clusters emerged (14). One cluster (“Low rating subgroup”) predominantly utilized the lower portion of the 0–10 scale (i.e., mean mild to severe pain range: −1.5 to −5.5 out of 10). A second cluster (“High rating subgroup”) predominantly utilized the upper portion of the scale (mild to severe range: −2.5 to −7.5 out of 10). While a third cluster (“Low/high rating subgroup”) utilized the fullest range of the 0–10 scale, (mild to severe range: −1.5 to −7.5). The relative consistency in pain ratings across hypothetical pain conditions supports the presence of pain rating schema. However, a recent study identified that individuals sometimes choose to over- or under- report their numeric pain ratings for specific situations or motivations such as not being taken seriously or to meet others’ expectations (26). Even in these cases, individuals must rely on an underlying schema to conceptualize their actual pain.

More recently, Wang, et al. (2023) applied a shortened version of the PSI survey for use as a covariate to adjust for differences in pain-rating schema when evaluating the relationships between multisensory sensitivity and number of chronic pain conditions (27). The shortened version, which assessed only headache and joint pain, was used to reduce subject burden, from 18 items to 6 items. While the original scale has shown good test-retest repeatability, high face validity and high internal consistency (14), whether the psychometric properties of the shortened form are similar was unreported. Thus, the purpose of this study was to examine the validity and reliability of a short form of the Pain Schema Inventory (PSI-6) using existing datasets.

2 Materials and methods

A retrospective secondary analysis was conducted using two datasets (Wang, et al, 2022; Wang, et al, 2023) to assess multiple psychometric properties of the PSI-6 (27, 28).

2.1 Participants

All participants completed multiple surveys in the original studies using the Research Electronic Data Capture interface (REDCap) (27, 28). Both studies were approved by the University of Iowa Institutional Review Board using a waiver of consent. The first dataset, Cohort 1, included 641 participants from a community-based population with and without pain who completed 18 items of the PSI (28). The second dataset, Cohort 2, included participants with chronic pain (n = 182) who completed the PSI-6 twice, one week apart (27). There was no overlap in participants from each cohort.

2.2 Pain Schema Inventory (PSI-18 and PSI-6)

The original study evaluating numerical pain rating schema asked participants to rate what they would perceive as mild, moderate and severe pain (3 categories) for 6 pain conditions: headache, muscle soreness, toothache, joint injury, burn, and pain “in general” (we refer to these as the PSI-18) (14). Pain threshold and tolerance were also originally explored (12 items, 30 total) but threshold and tolerance may have additional implications beyond numerical pain assignment, they have not typically been included in subsequent assessments. The PSI-6 includes the numeric ratings for mild, moderate and severe pain for only 2 of the conditions (headache and joint injury) (27, 28). The first dataset (Cohort 1) asked participants to complete the original 18 mild, moderate, and severe items, which allowed us to extract both the PSI-18 and PSI-6 for comparison, whereas the second dataset (Cohort 2) only assessed the PSI-6, but at two time points. We also explored using only the 3 PSI pain “in general” items (PSI-3) in Cohort 1. The evaluation of a subset extracted from the full instrument has been used previously, e.g., the SF-12 and SF-36 (29, 30).

In the original study, the PSI was assessed using a paper interface with a 10 cm visual analog scale (VAS) where 0 cm = no pain and 10 cm = worst pain imaginable (14). The current datasets (Cohorts 1 and 2), however, used an electronic survey interface where the PSI items were assessed with a 0–10 numeric rating scale (NRS) with the option to use whole or half numbers (0.5 increments). The PSI-18, PSI-6 and PSI-3 items and scoring are provided in the Appendix.

2.3 Statistical analyses

Summary statistics were computed for all primary measures (mean mild pain, moderate pain, severe pain, and overall average pain rating across all three pain levels) for both the PSI-18 and PSI-6.

2.3.1 Validity

Validity, defined as the extent to which a concept is accurately measured (31), can be assessed in multiple ways. We assessed both face validity, “the extent that an instrument measures the concept intended” (31) and convergent validity, a subset of criterion validity, which is typically assessed as how well one instrument measures similar constructs as another (31, 32). Both forms of validity were first assessed using a two-way repeated measures analysis of variance (ANOVA) for the first dataset (Cohort 1). The repeated factors included: category (mild, moderate, and severe) and form (PSI-18 and PSI-6). Additional covariates included sex (male = 0, female = 1) and pain status (no = 0, yes = 1). The Huynh-Feldt correction for non-sphericity was applied as needed. Effect sizes were computed as partial eta squared (η2) values and interpreted as: negligible (η2 < 0.01), small (0.01 ≤ η2 < 0.06), medium (0.06 ≤ η2 < 0.14), and large (η2 ≥ 0.14) (33).

Post hoc tests were performed as appropriate using paired t-tests with Bonferroni corrections. Simple effect sizes (Cohen's d) were also computed for each pair. One would expect mild pain to be numerically less than moderate or severe pain for the PSI-18 and the PSI-6. That is, intensity ratings should increase from mild to moderate to severe, with severe intensities being the highest as an indicator of face validity.

Convergent validity was assessed first by comparing pain ratings between the two forms, PSI-18 and PSI-6 using the two-way repeated measures ANOVA described above. In addition, Pearson's correlation coefficients for each mild, moderate, and severe pain category, as well as the average pain ratings between the PSI-6 and both the original PSI-18 and the remaining items from the PSI-18 omitted from the PSI-6 (i.e., the 12 items not included in the PSI-6) were performed. These approaches enable the evaluation of the consistency between the PSI-6 and its original version (primary outcome) as well as with the portion of the full form without duplication, thereby minimizing over-estimating convergence due to shared items (secondary outcome). We operationally defined acceptable convergent validity as r ≥ 0.7 (32), anticipating convergence with the original full PSI-18 to be higher than convergence with the items excluded from the PSI-6. We also assessed the correlations between the PSI-3 compared to the PSI-18 to explore if a further shortened version might be comparable, or an alternative, to the PSI-6.

Finally, we evaluated differences between PSI-18 and PSI-6 forms using Bland-Altman analyses, again using the first dataset (34, 35). Bland and Altman proposed that mean pairwise differences should be evaluated relative to mean pairwise values to evaluate for overall bias in the magnitude of the new scale. Accordingly, we plotted the differences (PSI-6—PSI-18) on the y-axes relative to the mean of the PSI-6 and PSI-18 scores on the x-axis for each pain severity category (mild, moderate, and severe) as well as the overall average pain rating across all categories. The mean difference (offset bias) and 95 percent limits of agreement for each Bland-Altman plot were computed and compared to the minimal clinically important difference (MCID) of pain, 2.3 on a 0–10 scale (36), as a reasonable estimate for comparison.

2.3.2 Reliability

Reliability is the consistency of a measure, which too can be assessed in multiple ways (31). Internal consistency is described as the extent to which items within an instrument are measuring the same underlying construct (31). We assessed internal consistency for both the PSI-18 and PSI-6 for Cohort 1 and the PSI-6 for Cohort 2 (first visit) with Cronbach's alpha. Optimal internal consistency has been previously reported with good or acceptable values between 0.70 and 0.95 (32), as overly high internal consistency (r > 0.95) may signify unnecessary redundant items within an instrument.

Test-retest reliability assesses the consistency of a measure when given to the same participants more than once under similar circumstances (31). Pearson correlation coefficients were computed between two assessments of the PSI-6, completed one week apart, using the second dataset (Cohort 2). Consistent with convergent validity, we operationally defined acceptable reliability as r ≥ 0.70 (32).

Significance levels were set at alpha ≤0.05 for all analyses, and statistical analyses were performed using IBM SPSS Statistics for Windows, version 28 (IBM Corp., Armonk, N.Y., USA).

3 Results

Details on Cohorts 1 and 2 including sex, age, race, ethnicity, educational level, and current as well as worst pain over the past 7 days are summarized in Table 1 (27, 28).

Table 1
www.frontiersin.org

Table 1. Participant characteristics reported as mean (SD) or n (%) as appropriate.

3.1 Validity

Both the PSI-18 and PSI-6 demonstrated good face validity (see Figure 1A), as the numerical pain ratings were consistently lowest for mild pain, higher for moderate pain, and highest for the severe pain category (p < 0.0001), with an overall very large effect size (η2 = 0.853). While differences between the original (PSI-18) and short form (PSI-6) were observed (p = 0.048, see Table 2), the overall effect size due to form was negligible (η2 = 0.006), evidence for good construct validity. While several two-way interactions were statistically significant, including pain category (mild, moderate, severe) by form (PSI-18, PSI-6); pain category by sex; and form by sex (see Table 2), the effect sizes were again typically small or negligible (η2 ≤ 0.02) for all but category by form which was moderate (η2 = 0.082). These effects were dwarfed by the differences observed between mild, moderate, and severe pain.

Figure 1
www.frontiersin.org

Figure 1. Mean (SD) pain ratings are shown in panel (A) for each pain category (mild, moderate, and severe pain) and the overall average pain rating across all items with PSI-6 (circles) and the PSI-18 (squares) and males in blue and females in orange. Face validity is demonstrated as mild pain ratings were significantly less than moderate, and moderate significantly less than severe pain. Significant sex differences were also noted, but with very small overall effect sizes (p = 0.006, η2 = 0.012) and inconsistent across pain categories. Significance signified by ***p < 0.001. Panels (B–E) show scatterplots of pain ratings using PSI-6 vs. PSI-18 from cohort 1 to evaluate construct validity, where B) mild pain (green), (C) moderate pain (orange), (D) severe pain (red), and (E) overall average pain (black). In panel (F) the overall test-retest stability of the PSI-6 is shown as a scatterplot of average pain ratings obtained from visit 1 (V1) vs. visit 2 (V2) for cohort 2.

Table 2
www.frontiersin.org

Table 2. Statistical analysis results.

Significant main effects of sex were also observed, but with a small effect size (η2 = 0.012, see also Figure 1A). post hoc paired t-tests revealed significant (p < 0.02) but small and inconsistent effect sizes for sex differences at each pain level (Cohen's d): mild (d = 0.10); moderate (d = 0.05); severe (d = −0.15). There were no significant differences in pain ratings by pain status (yes/no) for main (p = 0.12) or any interaction effects (p > 0.22).

Excellent convergent validity of the PSI-6 was demonstrated by high correlations with the PSI-18 for mild, moderate and severe pain categories (r = 0.88–0.90; p < 0.001), as well as for the overall average pain rating (r = 0.92; p < 0.001) (See Table 2). This means the PSI-6 explained 77%–85% of the variance observed in the full PSI-18 with only one-third the items. Similarly, the PSI-6 also demonstrated good to excellent convergent validity when compared to the 12 PSI items omitted from the PSI-6, r = 0.73–0.90 (p < 0.001) for mild, moderate, severe, and average pain (Table 2). While pain “in general” as a single item showed good correlations with the PSI-18 (p = 0.68–0.75; p < .001), they were consistently lower than the for the PSI-6, explaining 46%–56% of the PSI-18 variance.

The Bland-Altman analyses demonstrated mean offset biases close to zero across all pain levels: mild (0.11), moderate (0.06), and severe (−0.22) as well as the overall average (−0.01) (Figures 2A–D). The 95% limits of agreement were all well less than the MCID cut-off of 2.3: with ± limits of 1.01 (mild), 1.16 (moderate), 1.34 (severe) and 0.84 (average). In each plot, only a few select individuals showed differences that exceeded the 2.3 MCID.

Figure 2
www.frontiersin.org

Figure 2. Bland-Altman plots showing the PSI-6 minus the PSI-18 difference on the y-axes and the mean PSI-6 and PSI-18 on the x-axis. The mean offset bias between the two is shown with a dashed line, and 95% limits of agreement shown with dash-dot lines for (A) mild pain; (B) moderate pain; (C) severe pain; and (D) the average pain across all pain levels. Note that the 95% limits of agreement for all four plots are notably less than the reference MCID (2.3) used for this study, with only a few individuals with differences exceeding the MCID.

3.2 Reliability

Internal consistency of the PSI-18 and PSI-6, using the first dataset (Cohort 1), were good to excellent with Cronbach's Alpha = 0.91 and 0.80, respectively (Table 2). Similarly, in Cohort 2 Cronbach's Alpha was 0.88 for PSI-6. Test-retest reliability of the PSI-6 based on mild, moderate, and severe pain Pearson's correlation coefficients was good to high, with values ranging from 0.70 to 0.76, p < 0.001 (Table 2).

4 Discussion

The assessment of pain rating schema using the PSI-6 showed excellent face and convergent validity, high test-retest reliability and internal consistency. Our results indicated a high degree of consistency and agreement between the PSI-18 and the PSI-6. While some variability was noted using the Bland-Altman analyses, and the very small differences observed were statistically significant with the large sample size, no substantial or meaningful offset was observed. Collectively, these assessments suggest the brief, PSI-6 is a valid and reliable tool to assess pain rating schema with low respondent burden.

Numerous efforts have been made to better understand an individual's pain experience, from simple numeric pain ratings to fuller assessments including pain descriptors such as the McGill Pain Questionnaire (9). Indeed, studies comparing others’ perception of an individual's pain experience demonstrate often poor estimation of pain intensity, that can be influenced by the surrogate raters’ pain or health status (3739). Imaging and other biomarker studies have long been searching for the elusive objective measure of pain (40, 41), yet pain by definition is a sensory and emotional experience (3). Thus, we will always need to rely, at least to some degree, on asking individuals to communicate in some way their perception of their pain. A benefit of utilizing additional pain assessment methods or instruments beyond a NRS or VAS may be a deeper understanding and insight into the individual's pain experience (42).

Characterizing one's pain rating schema may aid in interpreting their ratings of pain intensity. Multiple studies have shown that there is a high degree of variability in how different individuals report their pain (7, 14, 16, 43), including intentional over- or under-reporting of pain for a variety of reasons (26). Patients with pain may have information-processing biases that can be at least partially explained by their pain rating schemas (14). Despite the widespread use of numerical scales to assess pain, there are very limited tools available to assess the construct of pain rating schema (42). The PSI-6 retains the ability of the PSI-18 to assess for an overall tendency to rate pain intensity across all levels (mild through severe) using a single mean score, or evaluate levels separately (e.g., mild vs. moderate vs. severe). Previous work demonstrated individual differences in pain rating schema, suggesting that the 0–10 rating that one individual perceives as mild may be another's moderate pain (14). Thus, differences in pain rating schema between individuals, if unrecognized, may lead to misunderstanding or misinterpretation of the others’ pain experience. For example, clinicians may undervalue patient report of pain intensity, for a variety of reasons, including implicit biases (26, 39, 44). Certainly, concerns of not being believed or having their pain taken seriously by patients compounds these challenges in communicating pain ratings (44). As pain is, by definition, a subjective experience (3), the identification of underlying pain schemas is not meant to reduce subjectivity but rather improve interpretation of the personalized pain experience. While not able to fully correct for the many nuances that may influence pain ratings, this assessment may help increase awareness of looking beyond unidimensional measures of pain as individuals sometimes find assigning numeric values to their pain inherently challenging, as we have observed anecdotally in both research and clinical settings.

Assessing hypothetical pain schema vs. simply asking an individual to also categorize their pain as mild, moderate, or severe provides two different but useful paradigms. An instrument comprised of hypothetical pain conditions provides a means to estimate how any individual uses a numeric pain scale, whether they are currently in pain. In the large cohort in which pain status (presence/absence) was considered, there was no effect on pain schema. This suggests that pain does not inherently bias individuals to change their underlying schema. The use of hypothetical pain items may be particularly helpful for research applications in which assessment across the severity range may serve to reduce pain rating heterogeneity and subsequently increase the precision for identifying underlying neurophysiologic pain mechanisms [e.g., see Wang, et al. (27)]. However, in clinical applications, it may be more feasible and immediately helpful to ask patients to report both numeric pain and whether it is best identified as mild, moderate or severe, as previously recommended (14). This approach has also been used to support cohort-wide pain rating schema determination, i.e., through the identification of mild, moderate, and sever pain cut-points (25), but is unable to generate individual-specific insights on numeric pain scale use.

When the PSI was first introduced, the authors used a 0–10 cm VAS for the pain ratings, finding good to high test-retest reliability and high internal consistency (14). Using a 0–10 NRS with 0.5 increments (21-point scale), we found similarly good to excellent psychometric properties of both the PSI-18 and PSI-6. This suggests that the PSI-18 and the PSI-6 are valid using either the 10 cm VAS or a 0–10 NRS (at least with the higher precision of 0.5 increments).

While longer questionnaires may provide the ability to detect more nuanced evaluation of any construct, the benefit of short forms is reduced subject burden and potentially, higher subject satisfaction and completion rates (45, 46). For example, the Patient-Reported Outcome Measurement Information System (PROMIS) measures typically have multiple versions, including several brief (i.e., 4-, 6- or 8-item) short forms. The PSI-6 was found to be a valid and reliable tool without a substantial loss of information from the PSI-18, making it desirable for both research and clinical applications. Further reducing burden using the PSI-3, relying only on single mild, moderate, and severe pain “in general” ratings, provides a reasonable estimate of the full PSI-18 (explained 46%–56% of the variance in PSI-18), yet was substantially worse than using the PSI-6 (explained 77%–85% of the variance in PSI-18). Thus, while both may be useful, the PSI-6 appears to represent the PSI-18 more consistently.

There are several limitations of the current study. First, the PSI assessment of pain rating schema does not reduce subjectivity as both numeric pain ratings and the categorization of pain as mild, moderate, or severe are subjective. However, their comparison provides at least a contextual framework to interpret numeric pain ratings. Another limitation is that the PSI-6 was extracted from the full PSI-18 and not collected separately. However, this approach has been used previously to assess shortened versions of instruments (29, 30, 45). Additionally, the optimal choice of which 6 items to use for the PSI-6 was not evaluated, but rather an assessment of the 6 items previously used (27, 28). It may be that a different reduced set of items (and/or number of items) could result in even greater convergence with the original PSI-18, however, the use of 2 pain conditions (headache and joint pain) was superior to just the single item of pain “in general.” Another challenge is that cohort diversity was limited. There were a limited number of men in both datasets, despite the minimal sex differences, we cannot rule out that the small number of men in our data is less representative of the larger population. Similarly, the cohorts were predominantly Caucasian and from the United States, thus may not generalize to all races, ethnicities, or other cultures.

5 Conclusion

This secondary analysis revealed that the PSI-6 is a valid and reliable tool to assess pain rating schema. This measure may be used to help adjust for variability in pain ratings among individuals, using the overall mean pain rating or the combination of pain ratings at each pain level (mild, moderate, severe), particularly for investigations in which pain ratings are a primary study outcome. Additional research is recommended to help further our understanding of pain rating schemas, including examining for systematic biases across pain populations, or with time following the development of chronic pain.

Data availability statement

The data analyzed in this study is subject to the following licenses/restrictions. Requests to access these datasets should be directed to laura-freylaw@uiowa.edu.

Ethics statement

The studies involving humans were approved by University of Iowa Biomedical Institutional Review Board. The studies were conducted in accordance with the local legislation and institutional requirements. The Ethics Committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants’ legal guardians/next of kin because this study only involved survey data obtained electronically. No direct contact or access to medical records was involved.

Author contributions

RW: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing. DW: Conceptualization, Data curation, Methodology, Writing – review & editing. LF-L: Conceptualization, Formal Analysis, Investigation, Methodology, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article.

Research reported in this publication was supported by the National Center For Advancing Translational Sciences of the National Institutes of Health under Award Number UM1TR004403 (supporting REDCap). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Acknowledgments

We would like to thank Maria Carrera, Katherine Douglas, Julia Hansen, Nathan Keele, Lexi Shaffer, Conor Swim, and Rachel Tebbe for their contribution to this manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Nahin RL. Estimates of pain prevalence and severity in adults: United States, 2012. J Pain. (2015) 16(8):769–80. doi: 10.1016/j.jpain.2015.05.002

PubMed Abstract | Crossref Full Text | Google Scholar

2. Dahlhamer J, Lucas J, Zelaya C, Nahin R, Mackey S, DeBar L, et al. Prevalence of chronic pain and high-impact chronic pain among adults - United States, 2016. MMWR Morb Mortal Wkly Rep. (2018) 67(36):1001–6. doi: 10.15585/mmwr.mm6736a2

Crossref Full Text | Google Scholar

3. Raja SN, Carr DB, Cohen M, Finnerup NB, Flor H, Gibson S, et al. The revised international association for the study of pain definition of pain: concepts, challenges, and compromises. Pain. (2020) 161(9):1976–82. doi: 10.1097/j.pain.0000000000001939

PubMed Abstract | Crossref Full Text | Google Scholar

4. Price DD, McGrath PA, Rafii A, Buckingham B. The validation of visual analogue scales as ratio scale measures for chronic and experimental pain. Pain. (1983) 17(1):45–56. doi: 10.1016/0304-3959(83)90126-4

PubMed Abstract | Crossref Full Text | Google Scholar

5. Jensen MP, Karoly P, Braver S. The measurement of clinical pain intensity: a comparison of six methods. Pain. (1986) 27(1):117–26. doi: 10.1016/0304-3959(86)90228-9

PubMed Abstract | Crossref Full Text | Google Scholar

6. Ho K, Spence J, Murphy MF. Review of pain-measurement tools. Ann Emerg Med. (1996) 27(4):427–32. doi: 10.1016/S0196-0644(96)70223-8

PubMed Abstract | Crossref Full Text | Google Scholar

7. Nielsen CS, Staud R, Price DD. Individual differences in pain sensitivity: measurement, causation, and consequences. J Pain. (2009) 10(3):231–7. doi: 10.1016/j.jpain.2008.09.010

PubMed Abstract | Crossref Full Text | Google Scholar

8. Hjermstad MJ, Fayers PM, Haugen DF, Caraceni A, Hanks GW, Loge JH, et al. Studies comparing numerical rating scales, verbal rating scales, and visual analogue scales for assessment of pain intensity in adults: a systematic literature review. J Pain Symptom Manage. (2011) 41(6):1073–93. doi: 10.1016/j.jpainsymman.2010.08.016

PubMed Abstract | Crossref Full Text | Google Scholar

9. Melzack R. The Mcgill pain questionnaire: from description to measurement. Anesthesiology. (2005) 103(1):199–202. doi: 10.1097/00000542-200507000-00028

PubMed Abstract | Crossref Full Text | Google Scholar

10. Poquet N, Lin C. The brief pain inventory (bpi). J Physiother. (2016) 62(1):52. doi: 10.1016/j.jphys.2015.07.001

PubMed Abstract | Crossref Full Text | Google Scholar

11. Jensen MP, Karoly P, O'Riordan EF, Bland F Jr, Burns RS. The subjective experience of acute pain. An assessment of the utility of 10 indices. Clin J Pain. (1989) 5(2):153–9. doi: 10.1097/00002508-198906000-00005

Crossref Full Text | Google Scholar

12. Farrar JT, Pritchett YL, Robinson M, Prakash A, Chappell A. The clinical importance of changes in the 0 to 10 numeric rating scale for worst, least, and average pain intensity: analyses of data from clinical trials of duloxetine in pain disorders. J Pain. (2010) 11(2):109–18. doi: 10.1016/j.jpain.2009.06.007

PubMed Abstract | Crossref Full Text | Google Scholar

13. Farrar JT, Young JP Jr., LaMoreaux L, Werth JL, Poole MR. Clinical importance of changes in chronic pain intensity measured on an 11-point numerical pain rating scale. Pain. (2001) 94(2):149–58. doi: 10.1016/s0304-3959(01)00349-9

PubMed Abstract | Crossref Full Text | Google Scholar

14. Frey-Law LA, Lee JE, Wittry AM, Melyon M. Pain rating schema: three distinct subgroups of individuals emerge when rating mild, moderate, and severe pain. J Pain Res. (2013) 7:13–23. doi: 10.2147/JPR.S52556

PubMed Abstract | Crossref Full Text | Google Scholar

15. Pincus T, Morley S. Cognitive-processing bias in chronic pain: a review and integration. Psychol Bull. (2001) 127(5):599. doi: 10.1037/0033-2909.127.5.599

PubMed Abstract | Crossref Full Text | Google Scholar

16. Turk DC, Rudy TE. Cognitive factors and persistent pain: a glimpse into Pandora’s box. Cognit Ther Res. (1992) 16(2):99–122. doi: 10.1007/BF01173484

Crossref Full Text | Google Scholar

17. Rusu AC, Pincus T, Morley S. Depressed pain patients differ from other depressed groups: examination of cognitive content in a sentence completion task. Pain. (2012) 153(9):1898–904. doi: 10.1016/j.pain.2012.05.034

PubMed Abstract | Crossref Full Text | Google Scholar

18. Greenwald AG, McGhee DE, Schwartz JL. Measuring individual differences in implicit cognition: the implicit association test. J Pers Soc Psychol. (1998) 74(6):1464–80. doi: 10.1037//0022-3514.74.6.1464

PubMed Abstract | Crossref Full Text | Google Scholar

19. Sullivan MJL, Bishop SR, Pivik J. The pain catastrophizing scale: development and validation. Psychol Assess. (1995) 7(4):524–32. doi: 10.1037/1040-3590.7.4.524

Crossref Full Text | Google Scholar

20. Hadlandsmyth K, Dailey DL, Rakel BA, Zimmerman MB, Vance CG, Merriwether EN, et al. Somatic symptom presentations in women with fibromyalgia are differentially associated with elevated depression and anxiety. J Health Psychol. (2020) 25(6):819–29. doi: 10.1177/1359105317736577

PubMed Abstract | Crossref Full Text | Google Scholar

21. Lee JE, Watson D, Frey Law LA. Lower-order pain-related constructs are more predictive of cold pressor pain ratings than higher-order personality traits. J Pain. (2010) 11(7):681–91. doi: 10.1016/j.jpain.2009.10.013

PubMed Abstract | Crossref Full Text | Google Scholar

22. Valencia C, Fillingim RB, George SZ. Suprathreshold heat pain response is associated with clinical pain intensity for patients with shoulder pain. J Pain. (2011) 12(1):133–40. doi: 10.1016/j.jpain.2010.06.002

PubMed Abstract | Crossref Full Text | Google Scholar

23. Morris P, Ali K, Merritt M, Pelletier J, Macedo LG. A systematic review of the role of inflammatory biomarkers in acute, subacute and chronic non-specific low back pain. BMC Musculoskelet Disord. (2020) 21(1):142. doi: 10.1186/s12891-020-3154-3

PubMed Abstract | Crossref Full Text | Google Scholar

24. O'Connell NE, Ferraro MC, Gibson W, Rice AS, Vase L, Coyle D, et al. Implanted spinal neuromodulation interventions for chronic pain in adults. Cochrane Database Syst Rev. (2021) 12(12):Cd013756. doi: 10.1002/14651858.CD013756.pub2

PubMed Abstract | Crossref Full Text | Google Scholar

25. Boonstra AM, Stewart RE, Köke AJ, Oosterwijk RF, Swaan JL, Schreurs KM, et al. Cut-off points for mild, moderate, and severe pain on the numeric rating scale for pain in patients with chronic musculoskeletal pain: variability and influence of sex and catastrophizing. Front Psychol. (2016) 7:1466. doi: 10.3389/fpsyg.2016.01466

PubMed Abstract | Crossref Full Text | Google Scholar

26. Boring BL, Walsh KT, Nanavaty N, Ng BW, Mathur VA. How and why patient concerns influence pain reporting: a qualitative analysis of personal accounts and perceptions of others’ use of numerical pain scales. Front Psychol. (2021) 12:663890. doi: 10.3389/fpsyg.2021.663890

PubMed Abstract | Crossref Full Text | Google Scholar

27. Wang D, Frey-Law LA. Multisensory sensitivity differentiates between multiple chronic pain conditions and pain-free individuals. Pain. (2023) 164(2):e91–e102. doi: 10.1097/j.pain.0000000000002696

PubMed Abstract | Crossref Full Text | Google Scholar

28. Wang D, Casares S, Eilers K, Hitchcock S, Iverson R, Lahn E, et al. Assessing multisensory sensitivity across scales: using the resulting core factors to create the multisensory amplification scale. J Pain. (2022) 23(2):276–88. doi: 10.1016/j.jpain.2021.07.013

PubMed Abstract | Crossref Full Text | Google Scholar

29. Lin Y, Yu Y, Zeng J, Zhao X, Wan C. Comparing the reliability and validity of the sf-36 and sf-12 in measuring quality of life among adolescents in China: a large sample cross-sectional study. Health Qual Life Outcomes. (2020) 18(1):360. doi: 10.1186/s12955-020-01605-8

PubMed Abstract | Crossref Full Text | Google Scholar

30. Müller-Nordhorn J, Roll S, Willich SN. Comparison of the short form (sf)-12 health status instrument with the sf-36 in patients with coronary heart disease. Heart. (2004) 90(5):523. doi: 10.1136/hrt.2003.013995

PubMed Abstract | Crossref Full Text | Google Scholar

31. Heale R, Twycross A. Validity and reliability in quantitative studies. Evid Based Nurs. (2015) 18(3):66–7. doi: 10.1136/eb-2015-102129

PubMed Abstract | Crossref Full Text | Google Scholar

32. Terwee CB, Bot SD, de Boer MR, van der Windt DA, Knol DL, Dekker J, et al. Quality criteria were proposed for measurement properties of health Status questionnaires. J Clin Epidemiol. (2007) 60(1):34–42. doi: 10.1016/j.jclinepi.2006.03.012

PubMed Abstract | Crossref Full Text | Google Scholar

33. Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2nd edn New York: Routledge (1988).

Google Scholar

34. Altman DG, Bland JM. Measurement in medicine: the analysis of method comparison studies. J R Stat Soc Ser D Stat. (1983) 32(3):307–17. doi: 10.2307/2987937

Crossref Full Text | Google Scholar

35. Bland J M, Altman D. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet. (1986) 327(8476):307–10. doi: 10.1016/S0140-6736(86)90837-8

PubMed Abstract | Crossref Full Text | Google Scholar

36. Olsen MF, Bjerre E, Hansen MD, Tendal B, Hilden J, Hróbjartsson A. Minimum clinically important differences in chronic pain vary considerably by baseline pain and methodological factors: systematic review of empirical studies. J Clin Epidemiol. (2018) 101:87–106.e2. doi: 10.1016/j.jclinepi.2018.05.007

PubMed Abstract | Crossref Full Text | Google Scholar

37. Desbiens NA, Mueller-Rizner N. How well do surrogates assess the pain of seriously ill patients? Crit Care Med. (2000) 28(5):1347–52. doi: 10.1097/00003246-200005000-00015

PubMed Abstract | Crossref Full Text | Google Scholar

38. Martire LM, Keefe FJ, Schulz R, Ready R, Beach SR, Rudy TE, et al. Older spouses’ perceptions of partners’ chronic arthritis pain: implications for spousal responses, support provision, and caregiving experiences. Psychol Aging. (2006) 21(2):222–30. doi: 10.1037/0882-7974.21.2.222

PubMed Abstract | Crossref Full Text | Google Scholar

39. Guru V, Dubinsky I. The patient vs. caregiver perception of acute pain in the emergency department. J Emerg Med. (2000) 18(1):7–12. doi: 10.1016/s0736-4679(99)00153-5

PubMed Abstract | Crossref Full Text | Google Scholar

40. van der Miesen MM, Lindquist MA, Wager TD. Neuroimaging-based biomarkers for pain: state of the field and current directions. Pain Rep. (2019) 4(4):e751. doi: 10.1097/pr9.0000000000000751

PubMed Abstract | Crossref Full Text | Google Scholar

41. Zhang Z, Gewandter JS, Geha P. Brain imaging biomarkers for chronic pain. Front Neurol. (2021) 12:734821. doi: 10.3389/fneur.2021.734821

PubMed Abstract | Crossref Full Text | Google Scholar

42. Salaffi F, Ciapetti A, Carotti M. Pain assessment strategies in patients with musculoskeletal conditions. Reumatismo. (2012) 64(4):216–29. doi: 10.4081/reumatismo.2012.216

PubMed Abstract | Crossref Full Text | Google Scholar

43. Van Ryckeghem DML, De Houwer J, Van Bockstaele B, Van Damme S, De Schryver M, Crombez G. Implicit associations between pain and self-schema in patients with chronic pain. Pain. (2013) 154(12):2700–6. doi: 10.1016/j.pain.2013.07.055

PubMed Abstract | Crossref Full Text | Google Scholar

44. Buchman DZ, Ho A, Illes J. You present like a drug addict: patient and clinician perspectives on trust and trustworthiness in chronic pain management. Pain Med. (2016) 17(8):1394–406. doi: 10.1093/pm/pnv083

PubMed Abstract | Crossref Full Text | Google Scholar

45. Baron G, Tubach F, Ravaud P, Logeart I, Dougados M. Validation of a short form of the Western Ontario and mcmaster universities osteoarthritis index function subscale in hip and knee osteoarthritis. Arthritis Care Res (Hoboken). (2007) 57(4):633–8. doi: 10.1002/art.22685

PubMed Abstract | Crossref Full Text | Google Scholar

46. Le Carré J, Luthi F, Burrus C, Konzelmann M, Vuistiner P, Léger B, et al. Development and validation of short forms of the pain catastrophizing scale (F-pcs-5) and tampa scale for kinesiophobia (F-tsk-6) in musculoskeletal chronic pain patients. J Pain Res. (2023) 16:153–67. doi: 10.2147/jpr.S379337

PubMed Abstract | Crossref Full Text | Google Scholar

Appendix

Pain Schema Inventory (PSI-18, PSI-6*, PSI-3†).

Please rate the following items on a 0–10 scale, with 0 being no pain and 10 being the worst pain imaginable, using half or whole numbers.

1. What would you rate as a moderate burn (e.g., from a hot stove)?

2. What would you rate as a severe joint injury (e.g., ankle sprain or arthritis)? *

3. In general, what would you rate as a mild pain?

4. What would you rate as a moderate headache? *

5. What would you rate as a severe burn (e.g., from a hot stove)?

6. What would you rate as a mild toothache?

7. What would you rate as severe muscle soreness (1–2 days after activity)?

8. What would you rate as moderate muscle soreness (1–2 days after activity)?

9. What would you rate as mild muscle soreness (1–2 days after activity)?

10. In general, what would you rate as a moderate pain?

11. What would you rate as a severe toothache?

12. What would you rate as a moderate joint injury (e.g., ankle sprain or arthritis)? *

13. What would you rate as a mild headache? *

14. What would you rate as a mild joint injury (e.g., ankle sprain or arthritis)? *

15. In general, what would you rate as a severe pain?

16. What would you rate as a mild burn (e.g., from a hot stove)?

17. What would you rate as a severe headache? *

18. What would you rate as a moderate toothache?


Notes and Scoring Instructions:

• Differences in pain rating schema may be characterized using the overall pain schema value as a covariate to adjust for overall tendency to use higher vs. lower range of the pain scale.

• For PSI-6 or PSI-3, only include the respective 6 (*) or 3 (†) items noted (see Table A1).

• Alternate pain scales have been used, including a 10 cm visual analog scale and the Borg CR10. We recommend allowing for 0.5 increments rather than only integers if using the 0–10 NRS.


Table A1
www.frontiersin.org

Table A1. Scoring of PSI.

For more information on this scale, see: Frey-Law LA, Lee JE, Wittry AM, Melyon M Pain Rating Schema: Three Distinct Subgroups of Individuals Emerge When Rating Mild, Moderate, and Severe Pain. J Pain Res (2013) 7:13-23. Epub 20131223. doi: 10.2147/JPRS52556.

Keywords: pain rating, questionnaire, psychometric property, Bland-Altman, pain schema

Citation: Wiederien RC, Wang D and Frey-Law LA (2024) Assessing how individuals conceptualize numeric pain ratings: validity and reliability of the Pain Schema Inventory (PSI–6) Short Form. Front. Pain Res. 5:1415635. doi: 10.3389/fpain.2024.1415635

Received: 10 April 2024; Accepted: 16 July 2024;
Published: 5 August 2024.

Edited by:

Paul Geha, University of Rochester, United States

Reviewed by:

Brandon Boring, Texas A and M University, United States
Rami Jabakhanji, Northwestern University, United States

© 2024 Wiederien, Wang and Frey-Law. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Laura A. Frey-Law, laura-freylaw@uiowa.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.