- 1School of Physical Education, Southwest University, Chongqing, China
- 2School of Mathematics and Statistics, Southwest University, Chongqing, China
- 3Institute of Motor Quotient, Southwest University, Chongqing, China
- 4Department of Kinesiology and Community Health, University of Illinois at Urbana-Champaign, Urbana, IL, United States
- 5International College of Football, Tongji University, Shanghai, China
- 6The Branch of the Collaborative Innovation Center of Assessment Toward Basic Education Quality, Southwest University, Chongqing, China
The Canadian Agility and Movement Skill Assessment (CAMSA) was recently widely used to assess fundamental motor skills in children. Although the CAMSA is reported to be reliable and valid, its measurement properties are not clear. This study aimed to examine the measurement properties of the CAMSA in a sample of Chinese children using Rasch analysis. The study sample was from 1,094 children aged 9–12 years in Zunyi City, Guizhou Province. Descriptive data were analyzed using SPSS 24.0 software, and the dichotomous data were analyzed by Winsteps version 4.5.4 and Facets 3.67.1 software performing Rasch analysis. The present study investigated CAMSA measurement characteristics by Rasch analysis, including the reliability of the rating instrument, unidimensionality, item-fit statistics, and differential item functioning (DIF). Inter-rater reliability and retest reliability showed that the CAMSA had a good internal consistency. Rasch analysis indicated that the CAMSA was unidimensional, locally independent, and had a good item-fit-statistic. Additionally, the CAMSA displayed a good fit for the item separation index (12.50 > 2.0), as well as for item reliability (0.99 > 0.90). However, the item difficulty of the CAMSA did not fit well with personal ability, and a significant DIF was found across genders. In the Chinese children sample test, the CAMSA demonstrated appropriate goodness-of-fit validity and rater reliability. Thus, future research will explore item difficulty and person ability fit, as well as DIF across genders.
Introduction
Children's fundamental motor skills (FMS) have long been described as a cornerstone of their physical activity, and they are typically classified into movement skills (e.g., running, sliding), object control skills (e.g., catching, kicking), and stability skills (e.g., balance) (1, 2). Proficiency in these skills has an important implication for children's healthy development (3, 4), yet numerous studies have indicated that the global rate of children's mastery is low (5). The development of FMS in childhood is crucial for the development of individuals as they develop through life. Therefore, the ideal FMS assessment tool should be provided during childhood, not only for diagnosing levels of motor skill development but also for targeting children's development in relation to motor skills.
Currently, numerous assessment tools are available internationally to assess children's FMS, such as TGMD (6, 7), MABC (8, 9), KTK (10, 11), and BOT (12, 13). A common feature of these assessment tools is that each movement is measured independently, and actions are not connected. These motor skill assessments accurately measure the completion of movements; however, the “real sport situation” is ignored, and the measurement results may deviate from an actual context (14). Contrary to these traditionally popular assessment tools, the Canadian Agility and Movement Skill Assessment (CAMSA) is the first international closed-loop motor skills assessment tool based on a series of combined movements (15). This assessment model is more closely matched to “real sport situations.” The term “real sport situation” refers to the fact that the movements in the assessment are coherent, continuous, and highly similar to the practical movement situation (16, 17). Additionally, the test results of this real sports can better reflect what is happening during children's movements. Initially, the CAMSA was developed to measure children's fundamental, complex, and integrated motor skills, with the primary purpose of diagnosing the level of motor development and identifying the risk of motor disorders in children (15). Since then, the CAMSA has been widely and concurrently used as a critical component of the Canadian Assessment of Physical Literacy (CAPL) (18). The Children's Physical Literacy Assessment application showed that the CAMSA assessment results were similar to real sport situations (19).
Numerous studies from Canada (20), Australia (2), the Netherlands (21), the United Kingdom (22), South Africa (23), and China (19) have explored the measurement properties of the CAMSA instrument. However, validity of the CAMSA test has only been reported in three studies to date (except the CAMSA development team) (2, 19, 20). Of these, Lander et al. showed that CAMSA skill scores had good concurrent validity (rs = 0.68) and inter-rater retest reliability (ICC = 0.85) in an Australian study of early adolescent girls (2). Another study from Canada validated the reliability and validity of the PLAYfun using the CAMSA as a validity criterion (20). The PLAYfun is an instrument with similar functions to the CAMSA, and it is also used to measure children's motor proficiency. A study by Stearns et al. revealed moderate to large correlations between PLAYfun and CAMSA (r = 0.47–0.60) (20). Furthermore, another study from a sample of Chinese male children explored the validity of the CAMSA timing test in comparison to three commonly used agility tests (19).
In the original study, the validity of the CAMSA was determined using the expert Delphi method. ANOVA was used to examine age and gender differences, and paired t-tests were conducted to examine differences between the effects of footwear vs. no footwear, as well as indoors vs. outdoors (15). Despite some studies further confirming the validity of the CAMSA skills assessment instrument (2, 20), processing data from the assessment results are inadequate. In particular, as a dichotomous data variable for the results of each CAMSA item, Rasch analysis is an effective method for processing this category of data (14, 24, 25). However, there are currently no studies that have applied Rasch analysis to establish the validity of the CAMSA. Additionally, although the validity of the CAMSA has been demonstrated in both Canadian and Australian children, there is no reported evidence of its validity among Chinese children (14). Therefore, this present study aimed to validate the measurement properties of the CAMSA skills test instrument for Chinese children using Rasch analysis.
Methods
Participants
Clauser and Mazor stated that the sample size should not be too small if DIF analysis were to be conducted, with more than 500 participants at least (26). However, Mara and Angoff argued that, when the sample size was too large, the slightest difference in the test would be highly significant, increasing the probability of type I errors (27, 28). Therefore, the proposed collection of a sample of approximately 1,000 children aged 9–12 years is appropriate for this study. In addition, considering the gender balance of the participants, it would be sufficient for the gender ratio of children in each age group to be approximately equal. Therefore, at least 250 children should be included in each age group. Finally, a total of 1,094 children were recruited from six elementary schools in Zunyi, Guizhou Province, between October 8, 2019, and November 30, 2019. All children included should be physically non-disabled (no physical disabilities) and devoid of congenital disorders (e.g., heart disease).
Instruments
The CAMSA is a movement capability assessment tool developed by Longmuir et al. for children aged 8–12 years (15). The CAMSA is used to assess children's fundamental movement skills and assess their capability to combine simple and complex movements. The CAMSA consists of seven movement items: two-foot jumping (2 points), sliding (3 points), catching (1 point), throwing (2 points), skipping (2 points), one-foot hop (2 points), and kicking (2 points) (Figure 1). The scoring points for each item are scored on a one-point scale (0–1). The skill score of the CAMSA is the total number of correctly completed skill movements and the total score ranges from 0 to 14 (15, 18). More details on CAMSA movement scoring can be found in the Canadian Assessment of Physical Literacy, Second Edition (CAPL-2, https://www.capl-ecsfp.ca) (29).
Anthropometric measurements were performed using the standard protocol of the “National Student Physical Fitness Standards” (NSPFS, 2014 revised version) (Ministry of Education of the People's Republic of China, 2014). The GJH1211 electronic tester was used to measure the height and weight of the participants, and the children were required to be barefoot and wear light clothing for the measurements. The test scale values were 0.1 cm for height and 0.1 kg for weight. The participants' body mass index (BMI) was calculated using BMI = height/weight2 (m/kg2). The BMI scoring criteria developed by the NSPFS were as follows. For boys: 9 years old (overweight 19.5–22.1, obesity ≥22.2, low weight ≤13.8); 10 years old (overweight 20.2–22.6, obesity ≥22.7, low weight ≤14.1); 11 years old (overweight 21.5–24.1, obesity ≥24.2, low weight ≤14.3); 12 years old (overweight 21.9–24.5, obesity ≥24.6, low weight ≤14.6). For girls: 9 years old (overweight 18.7–21.1, obesity ≥21.2, low weight ≤13.5); 10 years old (overweight 19.5–22.0, obesity ≥22.1, low weight ≤13.6); 11 years old (overweight 20.6–22.9, obesity ≥23.0, low weight ≤13.7); 12 years old (overweight 20.9–23.6, obesity ≥ 23.7, low body weight ≤14.1) (30).
Procedures
This study was approved by the Ethics Committee of the Institute of Motor Quotient, Southwest University (IRB No: SWUIMQ20190516). The study protocol was guided by the guidelines of the International Declaration of Helsinki. All participants had obtained written permission from their parents/guardians in the study.
A total of 1,927 children from six schools went home with a CAMSA test presentation, a parent/guardian consent form, and a student demographic questionnaire. One week later, 1,366 signed parental consent forms and demographic questionnaires were received. Investigators screened the demographic information forms, and 61 children were excluded (21 children had physical defects and 40 children were assessed for rater consistency). Finally, the number of participants that actually completed the entire test and were included in this study came to 1,094.
The six raters underwent rigorous training in CAMSA testing and proficiently mastered the administration process and movement commands of the CAMSA assessment (15). Before formal testing, 40 students (five per age group, male and female) were randomly selected from one school and scored by six raters to ensure consistency in the assessment scale. Two raters were allocated to a team, with one team as primary and the other two as secondary in the scoring test. The three scoring teams took turns to be primary and secondary. The entire test was videotaped. After 1 week of the testing interval, six raters were asked to rate the video again. The data from the on-site test were used for inter-rater reliability testing, and the data from the two tests were examined for retest reliability. The formal test was only administered when the scores of the six raters met the consistency test criteria.
The children watched two demonstrations of the test movements before testing, in accordance with the CAMSA manual. The raters explained the movements and guidance words to the children during the video demonstrations. The children were given two practice trials and two scored trials. Scores on the best of the two tests were used as final scores. Testing was conducted outside and was suspended in the event of rain. Generally, the tests were performed in physical education classes or during extracurricular activities. One physical education teacher was scheduled to assist in the administration of the children, and the assessors were only responsible for the scoring and timing. During the assessment process, two raters alternated in rating and timing in accordance with the subgroups.
Statistical Analysis
Descriptive statistical analyses were performed using SPSS 24.0 software to capture the demographic characteristics of the participants, including gender, age, and BMI. Rasch analysis was conducted to verify the construct validity of the CAMSA using WINSTEPS Version 4.5.4 software (31). Multi-faceted Rasch analysis was performed by using Facets Version 3.67.1 software (32, 33). To examine the measurement characteristics of CAMSA, a three-step procedure was performed following the CAMSA development guidelines.
First, inter-rater reliability and retest reliability were examined. Inter-rater reliability was evaluated using the multi-faceted Rasch model (MFRM). The MFRM is widely used to examine consistency among multiple raters. Zhu and Cole recommend a fit index infit and outfit (MnSq) criterion between 0.7 and 1.3 for motor skill assessment (34). We referred to Facets guidelines and adopted MnSq values between 0.7 and 1.3 for acceptable criteria in this study (35, 36). The intraclass correlation coefficient (ICC) was used to examine the rater's retest reliability (37). Wikstrom recommends the following criteria for ICC: scores of 0.9–1.0 for excellent; 0.80–0.89 for good; 0.7–0.79 for fair; and below 0.69 for poor (38).
Second, the Rasch residuals were tested for unidimensionality using principal component analysis (PCA). The PCA of the residuals was acceptable when the eigenvalue of the first factor extracted from the residuals was <2.0. Before PCA, the Rasch measurement dimension analysis involved examining point-measurement (PTMEA) biserial correlations and fit statistics (39). The measurement dimensions were confirmed to be devoid of negative PTMEA biserial correlations, and the fit statistics were stable, with no sudden high or low fits.
The difficulty independence assumption was tested using latent parallel analysis (LPA). Once the CAMSA showed a single dominant measurement structure, we performed item-level analyses using the Rasch model. Rasch analysis mainly includes item-fit statistics and differential item functioning (DIF) (37). The two goodness-of-fit statistics were used to examine the fit of the item model, namely infit and outfit (MnSq) (40). In the Rasch model, items are weighted, along with a linear logistic function, in accordance with their level of difficulty. The ratio of observed variance to expected variance will be 1.0 if an item fits this linear function exactly. Mean square values significantly more than 1.0 indicate model underfit, in contrast to values <1.0, which indicate model overfit. Wright and Linuck showed that the acceptable range of MnSq values is between 0.5 and 1.5 (41). Subsequently, some scholars recommend that a standard of MnSq values between 0.6 and 1.4 is better (42–44). Additionally, Bond and Fox argue that MnSq values between 0.7 and 1.3 are more accurate (45). Therefore, MnSq values between 0.7 and 1.3 were used in this study.
In addition, item reliability was assessed in terms of “separation” (G), which was considered to be the ratio of the true distribution of measurements to their measurement error (46). Item separation indices more significant than 2.0 are considered to be good. A related indicator is the reliability of these separation indices, with coefficients ranging from 0 to 1; a coefficient of 0.80 is considered good and 0.90 is considered excellent (46).
Results
Demographic Characteristics
Table 1 Presents the demographic characteristics of the participants. A total of 49.8% of male and 50.2% of female subjects completed all tests. The BMIs of all subjects showed that 81.6% were normal, 17.1% were overweight or obese, and 1.3% were underweight. The family status of the subjects showed that 16.8% of the children had lived without their parents for a long time, while only one parent accompanied 23.8% of the participants and both parents accompanied 59.4%. A total of 60.3% of the subjects were from urban areas and 39.7% were from rural areas.
Inter-rater and Retest Reliability
Table 2 presents the results of the inter-rater reliability and retest reliability. The inter-rater reliability results showed that R4 is the strictest and R1 is the loosest. The infit MnSq values of the raters ranged between 0.99 and 1.05, which fit well with the acceptable standard interval of 0.7–1.3. Item separation coefficients of zero and far less than two indicated that inter-rater differences could not be effectively distinguished. In other words, there was no significant variability among raters. The ICC results showed that the retest reliability of the six raters ranged from 0.979 to 0.987, indicating that raters were skilled in the scoring rules.
Rasch Analysis
The results for the measurement dimensions were confirmed to be without negative correlation of PTMEA (r = 0.14–0.46), and the fit statistics were stable (infit MnSq = 0.90–1.12), with no abrupt high or low fit (Table 3). The PCA of the residuals indicated that the eigenvalue of the first factor extracted from the residuals was 1.7 <2, indicating that the model was consistent with unidimensionality. The correlation between the residuals of the items was not significant (r = 0.01–0.21). Therefore, the local independence assumption was not violated for any item.
The infit and outfit MnSq values for all items were within the standard interval of 0.7–1.3, except for item CS01, which exhibited marginal misfit (outfit MnSq = 1.34). Item CS07 was the most difficult (2.73 logits), while item CS02 was the easiest (−2.56 logits). The results indicated that the item separation index was strong (G = 12.50), being well-above the criterion of 2.0. Item separation reliability was also excellent (r = 0.99).
Figure 2 located both items (difficulty levels) and persons (distribution of person ability) on the same continuum of fundamental movement skill. The figure indicated that the subjects' fundamental movement skill levels ranged from −1.58 to 4.51 logits, and the range of item response difficulty (−2.56 to 2.73 logits) was less than the subjects' fundamental movement skill levels; however, the range of item response difficulty covered 90.7% (n = 991) of the subjects. In addition, only 0.9% (n = 10) of the subjects scored in the highest scoring zone (raw score = 14), and 0.6% (n = 2) scored in the lowest zone (raw score = 3). Thus, there was no significant floor or ceiling effect. Also, Figure 2 showed that some test items have similar levels of difficulty (e.g., items CS03, CS11, and CS14), but most items target different levels of fundamental motor skill.
Figure 2. The map of person-item response difficulty locations. Each “#” in the PERSON column is 9 PERSON; each “.” is 1 to 8. +M, item mean; M, person mean; S, 1 SD from the mean; T, 2 SD from the mean.
The DIF analysis revealed four items falling outside the criteria across the age dimension (Table 4). In particular, three items were found between the ages of 11 and 12 (CS06 DIF contrast = −0.62, p = 0.03; CS07 DIF contrast = −0.53, p = 0.01; CS11 DIF contrast = −0.55, p = 0.03) (14). One item was identified between the ages of 10 and 11 (CS01 DIF contrast = −0.59, p = 0.03). The DIF test demonstrated that the four items CS01, CS06, CS07, and CS11, were responded to differently by subjects of different ages (Table 4). For the CAMSA, this means that the measurement invariance of the items varies across age groups. In addition, the DIF analysis found that ten items did not fit the criteria across genders, indicating that there were measurable differences in the difficulty of the ten items between genders (Table 5).
Discussion
The CAMSA is a fundamental movement skills assessment tool for children, but its measurement properties have yet to be thoroughly investigated. The purpose of this study was to investigate the measurement properties of the CAMSA using a Rasch model. This study was the first to use the Rasch model to investigate the CAMSA measurement properties, including item unidimensionality, local independence, item-fit, and differential item functioning (DIF) (47, 48). Overall, the CAMSA displayed adequate inter-rater reliability, retest reliability, internal consistency, and structural validity.
As expected, a multi-faceted Rasch model to test the consistency of the six raters showed good inter-rater reliability for CAMSA. The rater re-rated after a 1-week interval also demonstrated a positive rater retest reliability. Compared to the retest reliability of rater skills reported by CAMSA developers (ICC = 0.69) (15) and applications in Australia (ICC = 0.85) (2), the rater levels in this study (ICC = 0.979–0.987) were superior to the former. The rater reliability test demonstrated the necessity and validity of the raters' training and provided strong evidence that the CAMSA scoring rules are clear and valid. The establishment of inter-rater reliability and retest reliability assured the quality of CAMSA data collection.
Rasch analysis demonstrated that the CAMSA test is unidimensional, with no items violating the local independence assumption, and it is generally a hierarchical and well-developed assessment instrument (46). Analysis of the fit statistics indicated that the items were well-fitted, except for one item (CS01) outside the fitting border in the outfit MnSq. However, some studies suggested that infit values were more stable than outfit values in the item-fit statistics (46). The item CS01 infit MnSq of 1.12 is a trustworthy fit value. Therefore, the overall fit statistics of the CAMSA are regarded as good.
The person-item Wright map (Figure 2) shows that item difficulty did not match well with person-ability. With 38.4% of persons between 1 and 2.5 logits, no matching assessment items were found in this interval. In contrast, in the interval of −1 and 0 logits, eight items (57.1%) distinguished only 8.3% of the person-ability. Overall, the test results indicated that the item difficulty of the CAMSA test was low and the person ability was high (49). Therefore, the item difficulty of the CAMSA test struggles to differentiate individual abilities precisely (50). In general, for test results with low item difficulty and high personnel competency, adjusting item difficulty or test sequence is used to calibrate the assessment tool. Since the CAMSA is a closed-ended test instrument, adjusting the action sequence may change the properties of the original test structure. As such, a more appropriate method may be to increase the difficulty of the test items. In particular, in item-intensive locations (−1 to 0 logits), adjusting item difficulty may be a better strategy.
The differential item functioning analysis showed that four items (9.5%) existed DIF by the age factor. However, ten items (71.4%) were significantly DIF by the gender factor. The DIF items that existed by the age factor were mainly between the ages of 11 and 12, while no items between the ages of 9 and 10 years existed significantly in DIF. In terms of the age factor, the number of DIF items was greater in the older group than in the younger group, indicating that the low difficulty of the items may have led to a decrease in the differentiation of the older group, and thus to the DIF. Furthermore, the DIF was present in 71.4% of the items in the gender factor. The difference in movement test results across genders is inevitable in terms of physical and physiological interpretation. However, this also suggests that we should be cautious in interpreting differences between genders when using CAMSA test results.
First, although the validity of the CAMSA skills test fitting was good, poor matching of personal ability to item difficulty may result in differences in CAMSA discrimination. Second, the differential item functioning across genders prompted us to be cautious about comparing the results between men and women. Third, although the CAMSA is a well-tested assessment of “real sport” motor skills, there are still many agility motor skills (e.g., bilateral coordination, dynamic balance) that may not be assessed. Therefore, we need to improve the validity of the CAMSA measurements by increasing movement difficulty, adjusting movement scoring criteria, and enhancing the match between the difficulty of the instrument's movements and personal ability. This is a task for future studies. Furthermore, we recommend and encourage researchers and elementary physical education teachers to use the CAMSA to replace the traditional assessment tool of fundamental movement skills.
Conclusions
The present study is the first to analyze the measurement properties of the CAMSA using the Rasch model. The CAMSA, as a closed-loop measure of fundamental movement skills in children, demonstrated good unidimensionality, local assumption independence, and item-fit statistics. The inter-rater reliability and retest reliability revealed that the CAMSA was internally consistent. However, there were significant differences in its person–item fit matching and across genders in relation to differential item functioning.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
The studies involving human participants were reviewed and approved by the Scientific and Ethics Committee of Institute of Motor Quotient, Southwest University (IRB NO. SWUIMQ20190605). Written informed consent to participate in this study was provided by the participants' legal guardian/next of kin.
Author Contributions
JC and LY: data collection. JC and HY: data analysis. JC, JW, and NS: conception and design. JC, LY, HY, and JW: writing the manuscript and revision. All authors contributed to the article and approved the submitted version.
Funding
This study was funded by Humanities and Social Sciences Project of Ministry of Education (17YJC890020), National Social Science Fund (15CTY011), Postgraduate Education Research Project of Tongji University (2021GL15), the Fundamental Research Funds for the Central Universities of Southwest University (SWU1709240), the Education and Teaching Reform Research Project of Southwest University (2021JY071), and Chongqing Educational Science Planning Project (2020-GX-257).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We sincerely thank all the volunteers who participated in this study. We are grateful to Changyun Li, director of the Quality Education Center of Zunyi Education Bureau, for his great support of our data collection. We would also like to thank the reviewers for their comments and advices.
References
1. Gallahue DL, Ozmun JC, Goodway JD. Understanding Motor Development: Infants, Children, Adolescents, Adults. 7th ed. New York, NY, McGraw-Hill (2006).
2. Lander N, Morgan P J, Salmon J. The reliability and validity of an authentic motor skill assessment tool for early adolescent girls in an Australian school setting. J Sci Med Sport. (2017) 6:590–4. doi: 10.1016/j.jsams.2016.11.007
3. Navarro-Patón R, Arufe-Giráldez V, Sanmiguel-Rodríguez A, Mecías-Calvo M. Differences on motor competence in 4-year-old boys and girls regarding the quarter of birth: is there a relative age effect? Children. (2021) 2:141. doi: 10.3390/children8020141
4. Robinson LE, Stodden DF, Barnett LM, Lopes VP, Logan SW, Rodrigues LP, et al. Motor competence and its effect on positive developmental trajectories of health. Sports Med. (2015) 9:1273–84. doi: 10.1007/s40279-015-0351-6
5. Hardy LL, Barnett L, Espinel P, Okely AD. Thirteen-year trends in child and adolescent fundamental movement skills: 1997-2010. Med Sci Sports Exerc. (2013) 10:1965–70. doi: 10.1249/MSS.0b013e318295a9fc
8. Henderson SE, Sugden DA. Movement Assessment Battery for Children. Kent: Psychological Corporation (1992).
9. Henderson S, Sugden D, Barnett A. The Movement Assessment Battery for Children-2. London, UK: Pearson Education, Inc. (2007).
11. Kiphard EJ, Shilling F. Körperkoordinationtest für Kinder 2, überarbeitete und ergänzte Auflage.Weinheim: Beltz test (2007).
12. Bruininks RH. Bruininks Oseretsky Test of Motor Proficiency. Circle pines, MN: American Guidance Service (1978).
13. Bruininks RH. Bruininks-oseretsky Test of Motor Proficiency: BOT-2. Minneapolis, MN: NCS Pearson/AGS (2005).
14. Chang J, Li Y, Song H, Yong L, Luo L, Zhang Z, et al. Assessment of validity of children's movement skill quotient (CMSQ) based on the physical education classroom environment. Biomed Res Int. (2020) 20:8938763. doi: 10.1155/2020/8938763
15. Longmuir PE, Boyer C, Lloyd M, Borghese MM, Knight E, Saunders TJ, et al. Canadian agility and movement skill assessment (CAMSA): validity, objectivity, and reliability evidence for children 8-12 years of age. J Sport Health Sci. (2017) 2:231–40. doi: 10.1016/j.jshs.2015.11.004
16. Hoeboer J, De Vries S, Krijger-Hombergen M, Wormhoudt R, Drent A, Krabben K, et al. Validity of an Athletic Skills Track among 6- to 12-year-old children. J Sports Sci. (2016) 21:2095–105. doi: 10.1080/02640414.2016.1151920
17. Hoeboer J, Krijger-Hombergen M, Savelsbergh G, De Vries S. Reliability and concurrent validity of a motor skill competence test among 4- to 12-year old children. J Sports Sci. (2018) 14:1607–13. doi: 10.1080/02640414.2017.1406296
18. Longmuir PE, Boyer C, Lloyd M, Yang Y, Boiarskaia E, Zhu W, et al. The Canadian assessment of physical literacy: methods for children in grades 4 to 6 (8 to 12 years). BMC Public Health. (2015) 15:767. doi: 10.1186/s12889-015-2106-6
19. Cao Y, Zhang C, Guo R, Zhang D, Wang S. Performances of the canadian agility and movement skill assessment (CAMSA), and validity of timing components in comparison with three commonly used agility tests in Chinese boys: an exploratory study. PeerJ. (2020) 8:e8784. doi: 10.7717/peerj.8784
20. Stearns JA, Wohlers B, McHugh TLF, Kuzik N, Spence JC. Reliability and validity of the PLAY fun tool with children and youth in Northern Canada. Meas Phys Educ Exerc Sci. (2019) 1:47–57. doi: 10.1080/1091367X.2018.1500368
21. de Jong NB, Elzinga-Plomp A, Hulzebos EH, Poppe R, Nijhof SL, van Geelen S. Coping with paediatric illness: child's play? Exploring the effectiveness of a play- and sports-based cognitive behavioural programme for children with chronic health conditions. Clin Child Psychol Psychiatry. (2020) 3:565–78. doi: 10.1177/1359104520918327
22. Grainger F, Innerd A, Graham M, Wright M. Integrated strength and fundamental movement skill training in children: a pilot study. Children. (2020) 10:161. doi: 10.3390/children7100161
23. Pienaar AE, Gericke C, Plessis WD. Competency in object control skills at an early age benefit future movement application: longitudinal data from the NW-CHILD study. Int J Environ Res Public Health. (2021) 4:1648. doi: 10.3390/ijerph18041648
25. Chang J, Song N. Measurement properties of agility and movement skill assessment in children: a Rasch analysis. Med Sci Sports Exerc. (2020) 7S:60. doi: 10.1249/01.mss.0000670668.01731.b5
26. Clauser BE, Mazor KM. Using statistical procedures to identify differentially functioning test items. An NCME Instructional Module. Educ Meas Issues Pract. (1998) 1:31–44. doi: 10.1111/j.1745-3992.1998.tb00619.x
27. Dorans NJ, Holland PW. DIF detection and description: Mantel-Haenszel and standardization 1, 2. ETS Res Rep Ser. (1992) 1:i-40. doi: 10.1002/j.2333-8504.1992.tb01440.x
28. Angoff WH. Perspectives on differential item functioning methodology. In: Holland PW, Wainer H, editors. Differential Item Functioning. Hillsdale, NJ: Lawrence Erlbaum (1993). p. 3–23.
29. Tremblay MS, Longmuir PE. Conceptual critique of Canada's physical literacy assessment instruments also misses the mark. Meas Phys Educ Exerc Sci. (2017) 3:174–6. doi: 10.1080/1091367X.2017.1333002
30. National Physical Fitness Standards for Students Rating Scale. (2014). Available online at: http://www.csh.moe.gov.cn/wtzx/zl/20141226/2c909e854a8490a4014a8498e6730009.html (accessed June 1, 2021).
32. Park MS, Chung CY, Lee KM, Sung KH, Choi IH, Cho TJ, et al. Rasch analysis of the pediatric outcomes data collection instrument in 720 patients with cerebral palsy. J Pediatr Orthop. (2012) 4:423–31. doi: 10.1097/BPO.0b013e31824b2a1f
33. Weigle SC. Using FACETS to model rater training effects. Lang Test. (1998) 2:263–87. doi: 10.1177/026553229801500205
34. Zhu W, Cole EL. Many-faceted Rasch calibration of a gross motor instrument. Res Q Exerc Sport. (1996) 1:24–34. doi: 10.1080/02701367.1996.10607922
35. ten Klooster PM, Taal E, van de Laar MA. Rasch analysis of the Dutch health assessment questionnaire disability index and the health assessment questionnaire II in patients with rheumatoid arthritis. Arthritis Rheum. (2008) 12:1721–8. doi: 10.1002/art.24065
36. Elizur D. Facets of work values: a structural analysis of work outcomes. J Appl Psychol. (1984) 3:379–89. doi: 10.1037/0021-9010.69.3.379
37. Lee JH, Hong I, Park JH, Shin JH. Validation of Yonsei-bilateral activity test (Y-BAT) -bilateral upper extremity inventory using Rasch analysis. OTJR. (2020) 4:277–86. doi: 10.1177/1539449220920732
38. Wikstrom EA. Validity and reliability of Nintendo Wii Fit balance scores. J Athl Train. (2012) 3:306–13. doi: 10.4085/1062-6050-47.3.16
39. Linacre J. Detecting multidimensionality: which residual data-type works best? J Outcome Meas. (1998) 2:266–83.
40. Gellis ZD. Assessment of a brief CES-D measure for depression in homebound medically ill older adults. J Gerontol Soc Work. (2010) 4:289–303. doi: 10.1080/01634371003741417
42. Bond TG, Fox CM. Applying the Rasch Model: Fundamental Measurement in the Human Sciences. 2nd ed. Mahwah, NJ: Lawrence Erlbaum Associates (2007).
43. Krumlinde-Sundholm L, Holmefur M, Kottorp A, Eliasson AC. The assisting hand assessment: current evidence of validity, reliability, and responsiveness to change. Dev Med Child Neurol. (2007) 4:259–64. doi: 10.1111/j.1469-8749.2007.00259.x
44. Chien CW, Bond TG. Measurement properties of fine motor scale of Peabody developmental motor scales-second edition: a Rasch analysis. Am J Phys Med Rehabil. (2009) 5:376–86. doi: 10.1097/PHM.0b013e318198a7c9
45. Bond TG, Fox CM. Applying the Rasch Model: Fundamental Measurement in the Human Sciences. Hove East Sussex, UK: Psychology Press (2013).
46. Bond TG, Fox CM. Applying the Rasch Model: Fundamental Measurement in the Human Sciences. Mahwah, NJ: Lawrence Erlbaum Associates (2001).
47. Haley SM, Ni P, Jette AM, Tao W, Moed R, Meyers D, et al. Replenishing a computerized adaptive test of patient-reported daily activity functioning. Qual Life Res. (2009) 4:461–71. doi: 10.1007/s11136-009-9463-5
48. Lee M, Peterson JJ, Dixon A. Rasch calibration of physical activity self-efficacy and social support scale for persons with intellectual disabilities. Res Dev Disabil. (2010) 4:903–13. doi: 10.1016/j.ridd.2010.02.010
Keywords: agility, motor skills, assessment, Rasch analysis, CAMSA
Citation: Chang J, Yong L, Yan H, Wang J and Song N (2021) Measurement Properties of Canadian Agility and Movement Skill Assessment for Children Aged 9–12 Years Using Rasch Analysis. Front. Public Health 9:745449. doi: 10.3389/fpubh.2021.745449
Received: 22 July 2021; Accepted: 15 November 2021;
Published: 06 December 2021.
Edited by:
Patrizia Tortella, Free University of Bozen-Bolzano, ItalyReviewed by:
Ivana I. Kavecan, University of Novi Sad, SerbiaFrancesco Sgrò, Kore University of Enna, Italy
Copyright © 2021 Chang, Yong, Yan, Wang and Song. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Naiqing Song, songnq@swu.edu.cn; Jibing Wang, benwangking@126.com
†These authors have contributed equally to this work and share first authorship