Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 22 September 2022
Sec. Personality and Social Psychology

Development of the short Creative Expression Interest Scale based on item response theory

  • 1School of Psychology, Guizhou Normal University, Guiyang, China
  • 2School of Education, Jiangxi Normal University, Nanchang, China
  • 3School of Psychology, Jiangxi Normal University, Nanchang, China

This study develops a short Creative Expression Interest Scale (CEIS) among Chinese freshmen based on the perspective of item response theory (IRT). Nine hundred fifty-nine valid Chinese freshmen participated in the Creative Expression Interest survey. Researchers applied the initial data for unidimensionality, item fit, discrimination parameter, and differential item functioning to obtain a short CEIS. The results show that the Short CEIS meets the psychometric requirements of the IRT. Pearson correlation coefficient of theta between the short and long CEIS is 0.922. The marginal reliability of the short CEIS is 0.799. These indicate that the short CEIS developed in this study among Chinese freshmen, meets the psychometric requirements. Although the Short CEIS can eliminate redundant, uninformative items, save time, and improve the quality of data collection. However, the validity of this short scale needs further validation.

Introduction

Interest is essential in the vocational, organizational, and educational psychology fields. Vocational interest is an individual's preference and good at specific activities, which can stimulate an individual's goal-directed behavior and have greater predictive power for his or her engagement in a particular job (Su et al., 2019a). The current research on vocational interest is mainly based on Holland Vocational Interest Scale to explore the relationship between interest, education, and occupation (Su, 2020). In the occupational field, the type of interest is essential to the characteristics of the unique working environment (Hoff et al., 2020a) and life purpose (Stoll et al., 2021). Interest is also closely related to job performance (Van Iddekinge et al., 2011; Nye et al., 2017) and positively predicts job satisfaction (Hoff et al., 2020b) and career success (James and Su, 2014). In the field of education, interest is an emotional state which is an intrinsic motivation that drives learning and can predict academic achievement (Leung et al., 2014; Nye et al., 2021). Meta-analysis finds that occupational interest is relatively stable (Low et al., 2005; Hoff et al., 2018, 2021; Stoll et al., 2021). Therefore, it is vital for freshmen to objectively and comprehensively evaluate their vocational interests. CEIS is essential for students majoring in art or interested in this field. When a student takes part in the Creative Expression interest test, he or she can make a simple assessment of his/her interest and receive career guidance before applying for a job. At the same time, career interest tests help art students to better understand their majors, improve their professional identity, plan their careers well in advance, and increase employment success.

Although vocational interest has been generally recognized, there is no uniform conclusion on the division of specific dimensions (Gati, 1991). The vocational interest structure mainly focuses on dimensions four, six, seven, and eight. Turnstone developed a four-dimensional vocational interest scale based on a version of the Strong scale (Thurstone, 1931). The four-dimensional vocational interest concludes with Science, Language, People, and Business aspects. The four-dimensional vocational interest concludes with Science, Language, People, and Business aspects. The Language dimension encompasses advertising, art, and news. Guilford used analytic and clustering methods to develop a version with seven dimensions-scientific, aesthetic expression, social welfare, business, clerical, mechanical, and outdoor work. This version is the aesthetic expression that includes music, literature, drama, and artistic performance (Guilford et al., 1954). Jackson (Jackson, 1977) thought vocational interest should have logical, inquiring, expressive communication, and help conventional practical enterprising. Holland (Holland, 1959) proposed a new vocational interest scale which is popular at present. This scale has six dimensions-RIASEC (Realistic, Investigative, Artistic, Social, Enterprising, and Conventional). Based on the previous studies and the 2018 Standard Occupational Classification (SOC) version, Su developed an 8-dimensions model (SETPOINT: Health Science, Creative Expression, Technology, People, Organization, Influence, Nature, and Things). Creative expression has Media, Applied Arts & Design, Music, Visual Arts, Performing Arts, Creative Writing, and Culinary Art (Su et al., 2019b).

With the development of the times, the structure of vocational interest is constantly changing. So far, no vocational interest scale fits all venues (Liu and Rounds, 2003). However, based on existing models, it is easily found that art interest has always existed. Only the way of Expression and content has changed. The search for a dimensional structure of interests began with Thurstone's factor analysis. In this version, the language dimension has advertising, art, law, ministry, and journalism. One might guess that these professions are characterized more or less by an interest in talk (Thurstone, 1931). Later, studies began to use a variety of interest inventories. It is crucial to note Guilford's content-specific essential interest scales. This version's aesthetic Expression includes music, literature, drama, and artistic performance (Guilford et al., 1954). Additionally, in the late 1970s, Rounds et al. used many interest scales and included female and male participants. They labeled a dimension aesthetics such as sculptor, writing a one-act play, music composer, scenario writer, and illustrator (Rounds and Dawis, 1979). Holland defined this aspect as artistic in his six-dimension model (Holland, 1959). In one eight-dimension model, Creative Expression has Media, Applied Arts & Design, Music, Visual Arts, Performing Arts, Creative Writing, and Culinary Art (Su et al., 2019b). These expressions reflect the same underlying interest despite individuals expressing their creativity differently. Therefore, this study uses CEIS, a vocational interest scale developed by (Su et al., 2019b). Creative expression captures a general interest in activities involving expressing imaginative and creative ideas in various forms for art or practical considerations (Su et al., 2019b).

Scholars are always concerned about the quality of questionnaire responses (Kraut et al., 1975; Herzog and Bachman, 1981; Zhao and Kang, 1988; Burisch, 1997; Bowling et al., 2016). Long scales increase the cognitive load of the respondent. Subjects are easily fatigued. When a person is in this state, the quality of his or her answer will decrease (Meade and Craig, 2012; Gibson and Bowling, 2020; Zhong et al., 2021). There are many advantages of short scales, such as saving time, reducing the minimum number of subjects required, and improving the quality of responses to the data (Russell et al., 1989; Robins et al., 2001; Abdel-Khalek, 2006; Postmes et al., 2013).

When conducting studies with large samples, researchers recommend using single-item scales due to time constraints (Zhang and Wei, 2019; Goldammer et al., 2020). Career interest assessment is the basis for career planning. CEIS contains seven dimensions (Su et al., 2019b). If one item is extracted from a dimension, the short version of CEIS maybe contained seven items. Above all, we have a hypothesis: short CEIS has seven items. The seventh dimension is related to food processing and catering. In China, the traditional view is that food processing and catering belong to the service industry. Few people would associate these with art. Thus, there may be six items in the short CEIS.

The concepts and procedures of item response theory (IRT) have much broader applicability for psychological measurement (Steinberg and Thissen, 1996; Ark, 2007). This theory is through the item response curve synthesis of all item analysis data so that we can comprehensively and intuitively see the item difficulty, discrimination, and other characteristics. The most significant advantage of IRT is the invariance of question parameters. That is to say, the estimation of question parameters is independent of the subject group. This advantage helps to revise or develop scale. However, based on IRT, estimating the reliability and validity of the vocational interest scale is very rare (Tay et al., 2009). Therefore, the current research aims to use IRT and develop a short CEIS among Chinese freshmen.

General procedures for scale revision are as follows: first, evaluate assumptions of the IRT Model for unidimensionality, local independence, and monotonicity; second, Fit the IRT model to data, such as examine model fit, evaluate item properties, evaluate scale properties; third, evaluate differential item functioning (DIF) between gender groups (Reeve et al., 2007).

Samples

The researchers used a convenience sampling method to recruit subjects for this study. One thousand and seventy-two Chinese freshmen took part in the questionnaire survey. They came from Henan, Jiangxi, Guizhou, and Guangdong provinces of China. The study involved human participants and was reviewed and approved by the morality and ethics committee of the School of Psychology, Guizhou Normal University. The participants provided oral informed consent to participate in this study. Before conducting the questionnaire, each participant was informed of the individual privacy protection principle, and then they volunteered to participate in the survey. The survey included the basic demographic information (gender), the CEIS, and questions used as the exclusion criteria. In order to obtain accurate and effective response data, we screened the original questionnaires in advance.

The researchers distributed 1,072 questionnaires, and 1,000 completed questionnaires were collected, with a recovery rate of 93.28%. The final valid data for this study was 959, suggesting that the valid rate was 95.9%. If a participant's response at least met one aspect standard in the following, his or her response was regarded as invalid data: The two questions that measure attitudes do not meet the requirements (one was ‘this question is no response, please skip it,’ the other question was ‘Please choose choice A for this question’); there were clear answering rules, not thoughtfully answering; there were more than or equal two missing items. At last, there were 959 valid data. In this study, the researchers adopted a convenient snowball method; simultaneously, participants followed the principle of voluntary. As a result, the ratio of male to female students is unbalanced. Among participants, 282 were male (29.406%), and 677 were female (70.594%). One hundred three majored in engineering, 297 majored in education, 149 majored in science, 125 majored in literature, 139 majored in art, and 139 majored in others.

Measure

This study adopted the CEIS developed by (Su et al., 2019b). This part has 28 items on a five-point Likert scale (1 = Dislike a great deal, 5 = Like a great deal). The original long scale has media, applied arts & design, music, visual arts, performing arts, creative writing, and culinary art 7 dimensions. All the items belong to Creative Expression Interest. That is to say. This scale is a congeneric model (Osburn, 2000). Creative Expression Interest is one dimensional of vocational interest, so the higher the score is, the higher the level of Creative Expression Interest.

Procedure

Before an IRT model can be fit to data, three assumptions must be met. First, Unidimensionality, local independence, and monotonicity are the essential condition for IRT (Drasgow and Parsons, 1983; Embretson and Reise, 2000; Reeve et al., 2007; Acevedo-Mesa et al., 2021). Through the above steps of the IRT analysis, items met all the criteria to assess the person's ability parameter, then evaluate differential item functioning, calculate the correlation of personal parameters (theta) between the original (long) scale and the revised (short) scale.

Unidimensionality

In IRT, unidimensionality implies that one major ability or trait should explain or account for the test performance of examines (Hambleton and Swaminathan, 2014). The criteria for the data might be reasonably well fitted by a unidimensional model through exploratory factor analysis (EFA). The ratio of the first eigenvalue to the second eigenvalue was above 3 (Hattie, 1985). The percentage of variance interpreted by the first factor was more than 20% of the total variance (Reckase, 1979). Only two criteria were satisfied simultaneously; items might fit the unidimensionality model well (Reckase, 1979). The data on Creative Expression Interest might fit the unidimensionality model well. A bifactor model was used for the confirmatory factor analysis (CFA) (Li et al., 2015). Then researchers, via the bifactor function in the psych package of software R4.1.1, conducted a confirmatory factor analysis to test the unidimensionality in the way of high-order bifactor analysis (Gibbons and Hedeker, 1992; Reise et al., 2013; Sunderland et al., 2020). Model fit statistics were used to evaluate model fit and a select optimal number of specific factors and omega.h (ωh) coefficient. The ωh coefficient can be interpreted as the variance in unit-weighted total scores attributable to a single general factor, treating variability in scores due to group factors as measurement error (Sunderland et al., 2020). If ωh > 0.70, we can assume that the overall scores influence a single source primarily and provide support for essential unidimensionality (Sunderland et al., 2020).

Local independence

Local independence assumes that once the dominant factor influencing a person's response to an item is controlled, there should be no significant association among item responses (Ark, 2007). The existence of local independence that influences IRT parameter estimates poses a problem for scale construction. Uncontrolled local independence (LD) among items in an IRT assessment could result in a score different from the measured artistic expression interest construct. High residual correlations (greater than 0.2) will be flagged and considered as possible LD. If two items' LD coefficients were more significant than 0.2, it is better to consider deleting this item (Chen and Thissen, 1997). This function is residuals in the R4.1.1 Lavaan package.

Monotonicity

The assumption of monotonicity means that the probability of endorsing or selecting an item response indicative of a more vital interest status should increase as the underlying level of preference interest. Therefore, monotonicity is an essential requirement for IRT models for items with ordered response categories. In this process, we computed monotonicity via the R4.1.1 Mokken package. In addition, Loevinger was the first to develop an explicit theory of homogeneous tests based on “cumulative” or monotone items (Loevinger, 1948). Homogeneity (H) is one coefficient of monotonicity (Mokken, 1971). In this study, we choose the maximum test score (Hi) to be more than 0.3 (Ark, 2007).

Fit item response theory (IRT) model to data

Estimate IRT model parameters; estimate IRT model relative fitting index; examine model fit; evaluate item properties, category response curves, and item information curves; evaluate scale properties, information function (Reeve et al., 2007).

Estimate IRT model relative fitting index

To assess the degree of fitness for the whole items, the graded response model (GRM) (Samejima, 1969), generalized partial credit model (GPCM) (Muraki, 1992), and Rating Scale Model (RSM) (Andrich, 1978) dealing with polychromous-scored items were used in this study. Given the two test-level model-fit indices: Akaike's information criterion (AIC) (Akaike, 1974), and Bayesian information criterion (BIC) (Schwarz, 1978), the model representing the indices with relatively more minor values were taken as the optimal model for the subsequent analysis. Smaller values of these test-fit indices indicate better model fit (Tan et al., 2018). Therefore, we chose the most miniature model to compare the three model parameters (Reeve et al., 2007). In this process, we use the R4.1.1 Mirt package (Chalmers, 2012).

Item selection

After fitting the selected model, we considered the item content and the resulting psychometric information, such as item response and item information curves. These curves are used to evaluate item quality together with other criteria to identify items for removal from the original version or not.

Item fit

After comparing the relative goodness-of-fit indices of the models, the optimal model was selected, and the overall degree of data fitting was suggested. However, this did not indicate that each item could fit the optimal model well. Accordingly, the degree of item-fit further should be explored. Compare observed and expected response frequencies or examine fit indices, usually using S-χ2 (Reeve, 2003), which quantifies and compares the differences between observed and expected frequencies under the IRT model. The S-χ2 was adopted to show the degree of the item-fit index (Orlando, 2000; Maria and David, 2003). Items with a p-value of S-χ2 <0.001 were considered poor item-fit (Reeve et al., 2007; Flens et al., 2017). It would be considered as a low item-fit one and then removed.

Discrimination

Discrimination represents an item's ability to discriminate between individuals high and low on the latent trait (Li et al., 2015). An item with high discrimination implies that this item is preferable to distinguish whether individuals exhibit signs of preference. Therefore, a high discrimination parameter of one item suggests that this item is of high quality and helps in obtaining a more precise estimation of a latent population trait. We used the pars function by the R stats package to estimate item parameters via the optimal model based on a test-level model-fit check and chose items with discrimination of more than 0.8 (a > 0.8) (Tan et al., 2018).

Evaluate differential item functioning (DIF)

Considering the importance of the equivalence test in practical tests, the authors use DIF analysis to evaluate the systematic error caused by group bias. The DIF analysis can identify systematic errors attributed to different groups. That is, we conducted the DIF analysis to determine whether an individual's response to an item was a function of the underlying latent trait but of gender as well. If an item had DIF, the probability of selecting the same item type might differ for groups at the comparable latent trait level. The current study investigated DIF for gender (male/female). There were lots of methods, such as logistic regression (Crane et al., 2006), Likelihood-ratio (Orlando Edelen et al., 2006), and Wald tests (Millsap and Everson, 1993) to conduct the DIF analysis. Wald test compares separately estimated parameters across the male and female groups to examine whether they are statistically similar. These statistics could be used in tests of polytomous models. These models require more parameters than in the dichotomous case. The p < 0.001 revealed a functional difference in the item, which should be removed (Reeve et al., 2007). In order to reserve more items, we chose the p-value threshold of 0.001. In the process, researchers used the method of Wald to evaluate differential item functioning by the R Mirt package (Chalmers, 2012).

Pearson correlation between short and original CEIS

There are two CEIS. One version is the original long CEIS, and the other is a revised short CEIS. In this section, researchers computed every sample's theta, both long and short versions. First, expect a posteriori (EAP) was used to calculate the sample's theta. Second, researchers computed the theta coefficient of Pearson's correlation between the two versions via the Cor function in the R4.1.1 Psych package. This study then calculated the Pearson correlation between the CTT scores of short CEIS and each item that came from the short version.

Evaluate assumptions of the item response theory (IRT) model

Before applying IRT models, it is essential to evaluate the core assumptions of the model, for example, unidimensionality, monotonicity, and local independence. The ratio of the first and second components was 4.114, exceeding the criteria of 3 (Hattie, 1985). The first factor interpreted about 38.3% of all the variance, which was more than the criteria of 20% (Reckase, 1979). At the same time, researchers used high-order Bifactor analysis and got a coefficient of approximately 0.788, which was more than 0.70 (Sunderland et al., 2020). The above results indicated that the remained items could fit the unidimensional model well. Hi, values of all items were more significant than 0.3 (Ark, 2007), which met the monotonicity condition. The probability of agreeing or selecting a more substantial interest is increased with the level of interest (see Table 1). Pairs of items were reviewed for possible local dependence (LD) when residual correlations were more remarkable than 0.2 (Reeve et al., 2007). From Table 2, according to the Criteria of LD < 0.2, only seven items which are item3, item5, item9, item13, item17, item21, and item25, met the requirements (see Table 1).

TABLE 1
www.frontiersin.org

Table 1. Local independence.

TABLE 2
www.frontiersin.org

Table 2. Estimate item response theory (IRT) model relative fit index.

Test fit and IRT model selection

Test fit statistics of the GRM, the GPCM, and the RSM were documented in Table 2. For the GRM, AIC = 70693.579, and BIC = 71374.804, −2ll = 70413.58 (Log-likelihood = −35206.79). All relative fit indices of the GRM were less than those of the other two IRT models, which suggested that the GRM fitted the data better than the others. Therefore, this study chose the GRM model in the following parts (Reeve et al., 2007).

Item model-fit, discrimination, and different item functioning

Items with p values of P.s-x2 less than 0.001 were poor item-fit and should be eliminated (see Table 3). The P.s-x2 values of all items were more remarkable than 0.001, reaching the index of fitting degree test (Reeve et al., 2007). Therefore, there was no item removed. The discrimination coefficient of all items is more significant than 0.80, whose discrimination values are less than 0.8 and should be excluded from the items (Cho et al., 2015). A high discrimination parameter of one item suggests that this item is of high quality and helps obtain a more precise estimation of a latent population trait. Difficulty coefficient b is another index. From Figure 1, curve 1 decreases monotonically, curves 2, 3, and 4 firstly increase and then decrease, curve 5 monotonically increases, and curve 1 generally has a significant slope. The specific ability of 28 items is better and conforms to the item characteristic curves of the ideal state. The values of the b parameter should be in the range of (– 3, 3) (Steinberg and Thissen, 1996). The first 24 items' b parameter is in this range, while the values of the b parameter of Item25, Item26, Item27, and Item28 are beyond this scope. Samples of the original scale came from the Purdue University and the U.S. workforce (Su et al., 2019b), where a cultural concept is different from China. Few people in China were associated with food processing and catering with the artist. Four item information curves are relatively flat (Figure 2). These are consistent with the mean item information (mean-IIF) in Table 3. Item25, Item26, Item27, and Item28 can't distinguish different subjects' abilities well (Reeve et al., 2007). From this criterion, these four items should be deleted. According to Table 4, it is easily found that if the p value is taken as significance at 0.001 levels, four items should be deleted from the DIF test. They are Item4, Item13, Item14, and Item15. Item13, Item14, and Item15 are three items all belong to visual arts. If we removed all the four items, CEIS would lack visual arts (such as, Sketch a picture). According to the Criteria of LD and other indexes, Item13 fits the requirements. Considering that Item13 meets the criteria in all other criteria and ensures structural integrity; finally, Item13 is still in the expressive art interest scale.

TABLE 3
www.frontiersin.org

Table 3. Graded response model (GRM)'s item parameters.

FIGURE 1
www.frontiersin.org

Figure 1. Item response curves.

FIGURE 2
www.frontiersin.org

Figure 2. Item information curves.

TABLE 4
www.frontiersin.org

Table 4. Gender differential item functioning.

In this study, the marginal reliability of the long CEIS is 0.933, while the short version's marginal reliability was 0.799. Both of them met the requirement of psychometrics. Compared with the original scale, the short version saves more than 75% of the time, effectively reducing samples' cognitive fatigue and improving efficiency. Test information is 22.443 (see Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. Test information and standard errors.

Pearson correlation between short and original CEIS

Expect a posteriori (EAP) was used to calculate the sample's theta. Firstly, we separately estimated the theta of the long and short versions. The former is the original scale which has 28 items (CEIS-28). The latter is the new version with 6 items (CEIS-6). Using IRT, the Pearson correlation of person parameter (theta) between the long (CEIS-28) and short (CEIS-6) version is 0.922. We can find that they have a high correlation. Furthermore, the effect of the short scale was similar to that of the long scale. Therefore, the short scale can be used instead of the long scale in a large sample study, especially when the time is limited. Then, this study calculated the total score of the short CEIS and got a Pearson correlation between CEIS-6 and each item. Items host a radio program, create a piece of artistic and functional furniture, play a musical instrument, sketch a picture, perform on stage for a group of people, and write a novel. Pearson correlation between CEIS-6 total score and its every item is 0.718, 0.656, 0.694, 0.672, 0.703, 0.604 and p < 0.01. These show that this study develops a short CEIS among Chinese freshmen that meet the statistical measurement requirements.

Discussion

In order to reduce the cognitive load of respondents and improve the quality of results, this study develops short CEIS among Chinese freshmen. All the items came from Su's developed vocational interest scale, which came from the existing scale (Su et al., 2019b). Creative Expression Interest consists of seven sub-dimensions. Under the framework of IRT, after the unidimensionality test, monotonicity test, local independence test, item parameter estimation, and gender equivalence test, researchers developed a short CEIS with 6 items. These six items belong to one of the six sub-dimensions, consistent with the Single-item measures hypothesis (Postmes et al., 2013; Zhang and Wei, 2019). The seventh dimension is related to food processing and catering: select ingredients to prepare food, create the recipe for a new dish, create a new cooking technique to enhance flavor, and learn about the required temperature and time for baking pastries. In Chinese culture, the traditional view is that food processing and catering should belong to the service industry. Few people would associate this with art. In fact, with the change in the central contradiction in our society, people are seeking a better quality of life. Citizens may pay more attention to the beauty of appearance. So, the researchers need to pay more attention to this aspect in future studies.

The IRT analyses of unidimensionality, local independence, item fit, discrimination, and DIF were sequentially performed until all remaining items of the expressive art interest scale sufficiently satisfied the above rules (unidimensionality, local independence, good item-fit, high discrimination, and no DIF). Items that satisfied all the following criteria are included in the short CEIS: measuring at least one expressive art interest aspect, satisfying the hypothesis of measuring one primary dimension in IRT, satisfying the hypothesis of local independence in IRT, fitting the IRT model well, having high discrimination, and having no DIF. Subsequently, using the optimal model based on the test-level model-fit check, the item, and theta parameters of the final items were re-estimated for the IRT via software R4.1.1, such as Mirt, Psych, Mokken, and so on the package.

The results show that Chinese freshmen' CEIS contains six items, the specific content as follows: host a radio program, create a piece of artistic and functional furniture, play a musical instrument, sketch a picture, perform on stage for a group of people, and write a novel.

Limitation and future research

We eliminated redundant, uninformative items from the CEIS-28 to develop a revised CEIS-6. The reduced length of the CEIS version could improve the future efficient measurement of creative expression interest. At the same time, it could also be used as a vocational assessment tool in the art field. Although this study develops a short CEIS among Chinese freshmen, our study still has some limitations:

• The researchers adopted a convenient snowball method to recruit samples. As a result, the number of women was much more than men. Therefore, we should pay more attention to the balance of subjects in future studies.

• The sample only contains Chinese freshmen, which is not conducive to promoting the results. Future research should consider the role of cultural differences and conduct cross-cultural research to test the scale of this study.

• This study only develops a short CEIS. The validity needs further testing in future research.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The study involved human participants and was reviewed and approved by the morality and ethics committee of the School of Psychology, Guizhou Normal University. The participants provided oral informed consent to participate in this study. Before conducting the questionnaire, each participant was informed of the individual privacy protection principle, and then they volunteered to participate in the survey.

Author contributions

PZ is responsible for data analysis, method selection, manuscript writing, and revision. XG is responsible for manuscript selection, data integration, and analysis. NZ is responsible for data analysis. ZL is responsible for the manuscript communication with the finalization of the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the National Social Science Foundation of China General Topics in Education-Localized Research and Application of Value-added Evaluation Models (BGA210060).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abdel-Khalek, A. M. (2006). Measuring happiness with a single-item scale. Social Behav. Personal. Int. J. 34, 139–150. doi: 10.2224/sbp.2006.34.2.139

CrossRef Full Text | Google Scholar

Acevedo-Mesa, A., Tendeiro, J. N., Roest, A., Rosmalen, J. G. M., and Monden, R. (2021). Improving the measurement of functional somatic symptoms with item response theory. Assessment 28, 1960–1970. doi: 10.1177/1073191120947153

PubMed Abstract | CrossRef Full Text | Google Scholar

Akaike, H. T. (1974). A new look at the statistical model identification. Automatic Control IEEE Trans. 19, 716–723. doi: 10.1109/TAC.1974.1100705

CrossRef Full Text | Google Scholar

Andrich, D. (1978). A rating formulation for ordered response categories. Psychometrika 43, 561–573. doi: 10.1007/BF02293814

PubMed Abstract | CrossRef Full Text | Google Scholar

Ark, L. A. v. d. (2007). Mokken scale analysis inR. J. Stat. Software 20, 1–19. doi: 10.18637/jss.v020.i11

CrossRef Full Text | Google Scholar

Bowling, N. A., Huang, J. L., Bragg, C. B., Khazon, S., Liu, M., and Blackmore, C. E. (2016). Who cares and who is careless? Insufficient effort responding as a reflection of respondent personality. J. Personal. Social Psychol. 111, 218–229. doi: 10.1037/pspp0000085

PubMed Abstract | CrossRef Full Text | Google Scholar

Burisch, M. (1997). Test length and validity revisited. Eur. J. Personal. 11, 303–315.

PubMed Abstract | Google Scholar

Chalmers, R. P. (2012). Mirt: a multidimensional item response theory package for theREnvironment. J. Stat. Software 48, 1–29. doi: 10.18637/jss.v048.i06

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, W., and Thissen, D. (1997). Local dependence indexes for item pairs using item response theory. J. Edu. Behav. Stat. 22, 265–289. doi: 10.3102/10769986022003265

CrossRef Full Text | Google Scholar

Cho, S., Drasgow, F., and Cao, M. (2015). An investigation of emotional intelligence measures using item response theory. Psychol Assess 27, 1241–1252. doi: 10.1037/pas0000132

PubMed Abstract | CrossRef Full Text | Google Scholar

Crane, P. K., Gibbons, L. E., Jolley, L., and Van, B. G. (2006). Difffferential item functioning analysis with ordinal logistic regression techniques. DIF Detect Dif Wthpar. Med. Care 44, 115–123. doi: 10.1097/01.mlr.0000245183.28384.ed

PubMed Abstract | CrossRef Full Text | Google Scholar

Drasgow, F., and Parsons, C. K. (1983). Application of unidimensional item response theory models to multidimensional data. Appl. Psychol. Measure. 7, 189–199. doi: 10.1177/014662168300700207

PubMed Abstract | CrossRef Full Text | Google Scholar

Embretson, S. E., and Reise, S. P. (2000). Item Response Theory for Psychologists. Mahwah, NJ: Erlbaum. doi: 10.1037/10519-153

CrossRef Full Text | Google Scholar

Flens, G., Smits, N., Terwee, C. B., Dekker, J., Huijbrechts, I., and de Beurs, E. (2017). Development of a computer adaptive test for depression based on the dutch-flemish version of the PROMIS item bank. Evaluat. Health Professions 40, 79–105. doi: 10.1177/0163278716684168

PubMed Abstract | CrossRef Full Text | Google Scholar

Gati, I. (1991). The structure of vocational interests. Psychol. Bull. 109, 309–324. doi: 10.1037/0033-2909.109.2.309

CrossRef Full Text | Google Scholar

Gibbons, R. D., and Hedeker, D. R. (1992). Full-information item bi-factor analysis. Psychometrika 57, 423–436. doi: 10.1007/BF02295430

CrossRef Full Text | Google Scholar

Gibson, A. M., and Bowling, N. A. (2020). The effects of questionnaire length and behavioral consequences on careless responding. Eur. J. Psychol. Assess. 36, 410–420. doi: 10.1027/1015-5759/a000526

CrossRef Full Text | Google Scholar

Goldammer, P., Annen, H., Stöckli, P. L., and Jonas, K. (2020). Careless responding in questionnaire measures: detection, impact, and remedies. Leadership Quarterly 31, 101384. doi: 10.1016/j.leaqua.2020.101384

CrossRef Full Text | Google Scholar

Guilford, J. P., Christensen, P. R., Bond, N. A., and Sutton, M. A. (1954). A factor analysis study of human interests. Psychol. Monographs General Appl. 68, 1–38. doi: 10.1037/h0093666

CrossRef Full Text | Google Scholar

Hambleton, R. K., and Swaminathan, H. (2014). Item Response Theory: Principles and Applications. Berlin: Springer Science and Business Media

Google Scholar

Hattie, J. (1985). Methodology review: assessing unidimensionality of tests and ltenls. Applied Psychol. Measure. 9, 139–164. doi: 10.1177/014662168500900204

CrossRef Full Text | Google Scholar

Herzog, A. R., and Bachman, J. G. (1981). Effects of questionnaire length on response quality. Public Opin. Quarterly 45, 549–559. doi: 10.1086/268687

CrossRef Full Text | Google Scholar

Hoff, K. A., Briley, D. A., Wee, C. J. M., and Rounds, J. (2018). Normative changes in interests from adolescence to adulthood: a meta-analysis of longitudinal studies. Psychol. Bull. 144, 426–451. doi: 10.1037/bul0000140

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoff, K. A., Chu, C., Einarsdóttir, S., Briley, D. A., Hanna, A., and Rounds, J. (2021). Adolescent vocational interests predict early career success: two 12-year longitudinal studies. Appl. Psychol. 71, 49–75. doi: 10.1111/apps.12311

CrossRef Full Text | Google Scholar

Hoff, K. A., Song, Q. C., Einarsdottir, S., Briley, D. A., and Rounds, J. (2020a). Developmental structure of personality and interests: a four-wave, 8-year longitudinal study. J. Personal. Social Psychol. 118, 1044–1064. doi: 10.1037/pspp0000228

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoff, K. A., Song, Q. C., Wee, C. J. M., Phan, W. M. J., and Rounds, J. (2020b). Interest fit and job satisfaction: a systematic review and meta-analysis. J. Vocational Behav. 123, 103503. doi: 10.1016/j.jvb.2020.103503

CrossRef Full Text | Google Scholar

Holland, J. L. (1959). A theory of vocational choice. J. Counsel. Psychol. 6, 35–45. doi: 10.1037/h0040767

CrossRef Full Text | Google Scholar

Jackson, D. N. (1977). Manual for the Jackson Vocational Interest Survey. Port Huron, MI: Research Psychologists Press.

Google Scholar

James, R., and Su, R. (2014). The nature and power of interests. Current Direct. Psychol. Sci. 23, 98–103. doi: 10.1177/0963721414522812

CrossRef Full Text | Google Scholar

Kraut, A. I., Wolfson, A. D., and Rothenberg, A. (1975). Some effects of position on opinion survey items. J. Appl. Psychol. 60, 774–776. doi: 10.1037/0021-9010.60.6.774

CrossRef Full Text | Google Scholar

Leung, S. A., Zhou, S., Ho, E. Y.-F., Li, X., Ho, K. P., and Tracey, T. J. G. (2014). The use of interest and competence scores to predict educational choices of Chinese high school students. J. Vocat. Behav. 84, 385–394. doi: 10.1016/j.jvb.2014.02.010

CrossRef Full Text | Google Scholar

Li, J. J., Reise, S. P., Tuscano, A. C., Mikami, A. Y., and Lee, S. S. (2015). Item response theory analysis of ADHD symptoms in children with and without ADHD. Assessment 23, 655–671. doi: 10.1177/1073191115591595

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, C., and Rounds, J. (2003). evaluting the structure of vocational interest. Acta Psychol. Sinica 35, 411–418.

Google Scholar

Loevinger, J. (1948). The technic of homogeneous tests compared with some aspects of scale analysis and factor analysis. Psychol Bull. 45, 507–529. doi: 10.1037/h0055827

PubMed Abstract | CrossRef Full Text | Google Scholar

Low, K. S. D., Yoon, M., Roberts, B. W., and Rounds, J. (2005). The stability of vocational interests from early adolescence to middle adulthood: a quantitative review of longitudinal studies. Psychol. Bull. 131, 713–737. doi: 10.1037/0033-2909.131.5.713

PubMed Abstract | CrossRef Full Text | Google Scholar

Maria, O., and David, T. (2003). Further investigation of the performance of S - X2: an item fit index for use with dichotomous item response theory models. Appl. Psychol. Measure. 27, 289–298. doi: 10.1177/0146621603027004004

CrossRef Full Text | Google Scholar

Meade, A. W., and Craig, S. B. (2012). Identifying careless responses in survey data. Psychol Methods 17, 437–455. doi: 10.1037/a0028085

PubMed Abstract | CrossRef Full Text | Google Scholar

Millsap, R. E., and Everson, H. T. (1993). Methodology review: statistical approaches for assessing measurement bias. Appl. Psychol. Measure. 17, 297–334. doi: 10.1177/014662169301700401

CrossRef Full Text | Google Scholar

Mokken, R. J. (1971). A Theory and Procedure of Scale Analysis. The Hague, Netherlands: Mouton. doi: 10.1515/9783110813203

CrossRef Full Text | Google Scholar

Muraki, E. (1992). A generalized partial credit model: application of an EM algorithm. Appl. Psychol. Measure. 16, 159–176. doi: 10.1177/014662169201600206

CrossRef Full Text | Google Scholar

Nye, C. D., Prasad, J., and Rounds, J. (2021). The effects of vocational interests on motivation, satisfaction, and academic performance: test of a mediated model. J. Vocat. Behav. 127, 103583. doi: 10.1016/j.jvb.2021.103583

CrossRef Full Text | Google Scholar

Nye, C. D., Su, R., Rounds, J., and Drasgow, F. (2017). Interest congruence and performance: revisiting recent meta-analytic findings. J. Vocat. Behav. 98, 138–151. doi: 10.1016/j.jvb.2016.11.002

CrossRef Full Text | Google Scholar

Orlando Edelen, M. O., Thissen, D., Teresi, J. A., Kleinman, M., and Ocepek-Welikson, K. (2006). Identification of differential item functioning using item response theory and the likelihood-based model comparison approach. Application to the mini-mental state examination. Med Care 44, S134–142. doi: 10.1097/01.mlr.0000245251.83359.8c

PubMed Abstract | CrossRef Full Text | Google Scholar

Orlando, T. (2000). Likelihood-based item-fit indices for dichotomous item response theory models. Appl. Psychol. Measure. 24, 50–64. doi: 10.1177/01466216000241003

CrossRef Full Text | Google Scholar

Osburn, H. G. (2000). Coefficient alpha and related internal consistency reliability coefficients. Psychol. Methods 5, 343–355. doi: 10.1037/1082-989X.5.3.343

PubMed Abstract | CrossRef Full Text | Google Scholar

Postmes, T., Haslam, S. A., and Jans, L. (2013). A single-item measure of social identification: reliability, validity, and utility. British J. Social Psychol. 52, 597–617. doi: 10.1111/bjso.12006

PubMed Abstract | CrossRef Full Text | Google Scholar

Reckase, M. D. (1979). Unifactor latent trait models applied to multifactor tests: results and implications. J. Edu. Stat. 4, 207–230. doi: 10.3102/10769986004003207

CrossRef Full Text | Google Scholar

Reeve, B. B. (2003). Item response theory modeling in health outcomes measurement. Expert Rev. Pharmacoecon Outcomes Res. 3, 131–145. doi: 10.1586/14737167.3.2.131

PubMed Abstract | CrossRef Full Text | Google Scholar

Reeve, B. B., Hays, R. D., Bjorner, J. B., Cook, K. F., Crane, P. K., Teresi, J. A., et al. (2007). Psychometric evaluation and calibration of health-related quality of life item banks: plans for the Patient-Reported Outcomes Measurement Information System (PROMIS). Med Care 45, S22–31. doi: 10.1097/01.mlr.0000250483.85507.04

PubMed Abstract | CrossRef Full Text | Google Scholar

Reise, S. P., Scheines, R., Widaman, K. F., and Haviland, M. G. (2013). Multidimensionality and structural coefficient bias in structural equation modeling. Edu. Psychol. Measure. 73, 5–26. doi: 10.1177/0013164412449831

CrossRef Full Text | Google Scholar

Robins, R. W., Hendin, H. M., and Trzesniewski, K. H. (2001). Measuring global self-esteem: construct validation of a single-item measure and the rosenberg self-esteem scale. Personal. Social Psychol. Bull. 27, 151–161. doi: 10.1177/0146167201272002

CrossRef Full Text | Google Scholar

Rounds, J. B., and Dawis, R. V. (1979). Factor analysis of strong vocational interest blank items. J. Appl. Psychol. 64, 132–143. doi: 10.1037/0021-9010.64.2.132

CrossRef Full Text | Google Scholar

Russell, J. A., Weiss, A., and Mendelsohn, G. A. (1989). Affect grid: a single-item scale of pleasure and arousal. J. Personal. Social Psychol. 57, 493–502. doi: 10.1037/0022-3514.57.3.493

CrossRef Full Text | Google Scholar

Samejima, F. (1969). Estimation of latent ability using a response pattern of graded scores. Psychometrika 34, 1–97. doi: 10.1007/BF03372160

PubMed Abstract | CrossRef Full Text | Google Scholar

Schwarz, G. (1978). Estimating the dimension of a model. Annals Stat. 6, 461–464. doi: 10.1214/aos/1176344136

CrossRef Full Text | Google Scholar

Steinberg, L., and Thissen, D. (1996). Uses of item response theory and the testlet concept in the measurement of psychopathology. Psychological Methods 1, 81–97. doi: 10.1037/1082-989X.1.1.81

CrossRef Full Text | Google Scholar

Stoll, G., Rieger, S., Nagengast, B., Trautwein, U., and Rounds, J. (2021). Stability and change in vocational interests after graduation from high school: a six-wave longitudinal study. J. Personal. Social Psychol. 120, 1091–1116. doi: 10.1037/pspp0000359

PubMed Abstract | CrossRef Full Text | Google Scholar

Su, R. (2020). The three faces of interests: an integrative review of interest research in vocational, organizational, and educational psychology. J. Vocat. Behavior 116, 103240. doi: 10.1016/j.jvb.2018.10.016

CrossRef Full Text | Google Scholar

Su, R., Tay, L., Liao, H.-Y., Zhang, Q., and Rounds, J. (2019a). Toward a dimensional model of vocational interests. J. Appl. Psychol. 104, 690–714. doi: 10.1037/apl0000373

PubMed Abstract | CrossRef Full Text | Google Scholar

Su, R., Zhang, Q., Liu, Y., and Tay, L. (2019b). Modeling congruence in organizational research with latent moderated structural equations. J. Appl. Psychol. 104, 1404–1433. doi: 10.1037/apl0000411

PubMed Abstract | CrossRef Full Text | Google Scholar

Sunderland, M., Afzali, M. H., Batterham, P. J., Calear, A. L., Carragher, N., Hobbs, M., et al. (2020). Comparing scores from full length, short form, and adaptive tests of the social interaction anxiety and social phobia scales. Assessment 27, 518–532. doi: 10.1177/1073191119832657

PubMed Abstract | CrossRef Full Text | Google Scholar

Tan, Q., Cai, Y., Li, Q., Zhang, Y., and Tu, D. (2018). Development and validation of an item bank for depression screening in the chinese population using computer adaptive testing: a simulation study. Front Psychol. 9, 1225. doi: 10.3389/fpsyg.2018.01225

PubMed Abstract | CrossRef Full Text | Google Scholar

Tay, L., Drasgow, F., Rounds, J., and Williams, B. A. (2009). Fitting measurement models to vocational interest data: are dominance models ideal? J. Appl. Psychol. 94, 1287–1304. doi: 10.1037/a0015899

PubMed Abstract | CrossRef Full Text | Google Scholar

Thurstone, L. L. (1931). A multiple factor study of vocational interests. Personnel J. 3, 198–205.

Google Scholar

Van Iddekinge, C. H., Roth, P. L., Putka, D. J., and Lanivich, S. E. (2011). Are you interested? A meta-analysis of relations between vocational interests and employee performance and turnover. J. Appl. Psychol. 96, 1167–1194. doi: 10.1037/a0024343

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, L., and Wei, X. (2019). Single-item measures: queries, responses and suggestions. Adv. Psychol. Sci. 27, 1194. doi: 10.3724/SP.J.1042.2019.01194

CrossRef Full Text | Google Scholar

Zhao, Z., and Kang, Y. (1988). Relationship between the length of questionnaire and response styles. Acta Psychologica Sinica 1, 15–19.

Google Scholar

Zhong, X., Li, M., and Li, L. (2021). Preventing and detecting insufficient effort survey responding. Adv. Psychol. Sci. 29, 225. doi: 10.3724/SP.J.1042.2021.00225

CrossRef Full Text | Google Scholar

Keywords: short scale, Creative Expression Interest Scale, item response theory, personality, self-efficiency

Citation: Zhao PJ, Gao XL, Zhao N and Luo ZS (2022) Development of the short Creative Expression Interest Scale based on item response theory. Front. Psychol. 13:955176. doi: 10.3389/fpsyg.2022.955176

Received: 28 May 2022; Accepted: 08 August 2022;
Published: 22 September 2022.

Edited by:

Peida Zhan, Zhejiang Normal University, China

Reviewed by:

Qianqian Pan, Nanyang Technological University, Singapore
YuTing Han, Peking University, China

Copyright © 2022 Zhao, Gao, Zhao and Luo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhao Sheng Luo, luozs@jxnu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.