Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 18 January 2023
Sec. Organizational Psychology

Game based assessments of cognitive ability in recruitment: Validity, fairness and test-taking experience

Franziska Leutner,
Franziska Leutner1,2*Sonia-Cristina CodreanuSonia-Cristina Codreanu3Suzanne BrinkSuzanne Brink2Theodoros BitsakisTheodoros Bitsakis2
  • 1Institute of Management Studies, Goldsmiths, University of London, London, United Kingdom
  • 2HireVue, Inc, Salt Lake City, UT, United States
  • 3King’s College London, London, United Kingdom

Gamification and machine learning are emergent technologies in recruitment, promising to improve the user experience and fairness of assessments. We test this by validating a game based assessment of cognitive ability with a machine learning based scoring algorithm optimised for validity and fairness. We use applied data from 11,574 assessment completions. The assessment has convergent validity (r = 0.5) and test–retest reliability (r = 0.68). It maintains fairness in a separate sample of 3,107 job applicants, showing that fairness-optimised machine learning can improve outcome parity issues with cognitive ability tests in recruitment settings. We show that there are no significant gender differences in test taking anxiety resulting from the games, and that anxiety does not directly predict game performance, supporting the notion that game based assessments help with test taking anxiety. Interactions between anxiety, gender and performance are explored. Feedback from 4,778 job applicants reveals a Net Promoter score of 58, indicating more applicants support than dislike the assessment, and that games deliver a positive applicant experience in practise. Satisfaction with the format is high, but applicants raise face validity concerns over the abstract games. We encourage the use of gamification and machine learning to improve the fairness and user experience of psychometric tests.

1. Introduction

Cognitive ability tests have been used as popular selection and recruitment tools for decades, with good reason: The predictive validity of cognitive ability for job performance is well documented. A meta analytic review of over 20,000 studies combining 5 million participants shows an average correlation of r = 0.5 between cognitive validity and job performance (Kuncel et al., 2010). Cognitive ability is related to a variety of performance outcomes, including supervisor rated performance (r = 0.42–0.57, p < 0.001; Judge et al., 1999) and extrinsic career success (r = 0.53, p < 0.001; Higgins et al., 2007). In addition, cognitive ability tests are amongst the most effective selection methods available: Schmidt and Hunter’s review of meta-analytic studies on 19 selection methods reports a predictive validity of intelligence of r = 0.51 for job performance, rivalled only by work sample tests (r = 0.54) and structured interviews (r = 0.51) (Schmidt and Hunter, 1998). In addition, cognitive ability predicts a wide range of real world outcomes such as educational attainment, socioeconomic status, and career success (Schmidt and Hunter, 2004).

In recent years, gamification of cognitive tests has become a popular method for improving the test taker experience of psychometric tests (Raghavan et al., 2020). Gamification entails the introduction of game mechanisms as motivational affordances designed to result in psychological outcomes such as motivation, attitude and enjoyment (Hamari et al., 2014). In the context of recruitment, gamification should also include a level of real world connection and job relatedness (Korn et al., 2017). Game mechanisms include adaptive levels, progression, immediate feedback, and intermittent goals (Tremblay et al., 2010; Landers and Callan, 2011; Palmer et al., 2012). In the psychometric assessment for recruitment context, gamification also implies optimisation for mobile and online use (Landers and Callan, 2011). Gamification is studied in vastly different areas from entertainment and video games to gamified work tasks, but the most common context is serious games for learning (Hamari et al., 2014) or, within organisational psychology, gamification for employee training or customer engagement (Tippins, 2015).

Descriptions of gamified cognitive ability tests for use as psychometric assessments in recruitment are rare in the literature. However, several studies describe gamified measures of cognitive abilities that approximate performance on established, non-gamified tasks. For example, a gamified four-dimensional spatial task predicts performance on several cognitive ability tasks including a working memory span task (r = 0.62, p < 0.05), a quantitative reasoning task (r = 0.30, p < 0.05), and the Raven’s progressive matrices, a popular abstract reasoning test (r = 0.37, p < 0.05) (Atkins et al., 2014). In a similar vein, gamified versions of working memory and processing speed tasks correlate with the Raven’s Progressive Matrices short form between r = 0.29 (p < 0.01) and r = 0.45 (p < 0.05) (McPherson and Burns, 2008). Cognitive tasks implemented in the computer game Minecraft have high convergent validity with both Raven’s Standard Progressive Matrices and the Vandenberg & Kuse Mental Rotations Test (r = 0.72, p < 0.002 in Peters et al., 1995, 2021; Raven, 2000). Gameplay on 12 short video games selected specifically to test components of intelligence correlated strongly with cognitive ability assessed across 11 established cognitive ability tests (r = 0.93, p < 0.01; Quiroga et al., 2015).

Additional studies describe game based cognitive ability tests developed for specific populations such as cognitive testing and training of children (Verhaegh et al., 2013), detection of cognitive decline in elders (Jimison et al., 2004, 2007), entertainment games (Gamberini et al., 2010), or edutainment games to train cognitive ability (Dandashi et al., 2015). Some evidence for the predictive validity of gamified tasks is provided by two studies on school children. Performance on gamified versions of working memory and processing speed tasks correlates similarly with school performance and performance on the Raven’s matrices (McPherson and Burns, 2008). A game-like cognitive task differentiates between students with typical and low mathematics achievement (AUC = 0.897; Luft et al., 2013).

1.1. The gamification advantage

Entertainment games deliver a range of behavioural and affective outcomes (Connolly et al., 2012). These include immersion (Jennett et al., 2008) and experiences of flow (Chen, 2007; Nacke et al., 2009), which in turn enhance enjoyment (Weibel et al., 2008). The goal of gamification is to produce the same subjective experiences as games do (Huotari and Hamari, 2012), and to use game mechanisms to create enjoyment and playability (Landers and Callan, 2011). Indeed, the majority of studies that describe gamification observe positive effects on players’ subjective experiences, including on motivation, attitude and enjoyment, indicating that gamification is successful at producing some of the same affective outcomes as entertainment games (Hamari et al., 2014). For example, gamified tasks increase attention (Klein et al., 2017) and cognitive test performance (Lumsden et al., 2016). Market research questionnaires that are gamified increase enjoyment compared to those that are not (Guin et al., 2012). In addition to game mechanics, framing a task as a game alone increases interest and enjoyment (Lieberoth, 2015). Introducing gamification elements into HR tasks such as onboarding increases enjoyment relative to non-gamified tasks (Heimburger et al., 2020).

In the context of assessment, gamification might also help reduce test taking anxiety, which could in turn improve performance. Game based assessments might help reduce the effect of moderators and mediators of differences in test performance, namely test anxiety and stereotype threat. Test anxiety is an unpleasant emotional state experienced in testing situations (Dusek, 1980). Gamification provides users with a less anxiety-inducing environment by unobtrusively logging behaviour in a “fun” way, allowing the participant to fully immerse in the task and making it feel less threatening (McPherson and Burns, 2008; Kato and de Klerk, 2017). This decrease in anxiety can lead to increased performance compared to traditional pen-and-paper tests. For example, participants completing a gamified version of a multimedia systems knowledge test have decreased anxiety and increased performance compared to those completing a traditional paper test (Mavridis and Tsiatsos, 2017). Furthermore, gamification reduces performance differences caused by stereotype threat, with no gender performance differences observed in a high threat gamified logic test (Albuquerque et al., 2017).

1.2. Fairness and diversity problems in cognitive ability tests

In application for recruitment and selection, cognitive ability tests are prone to generating adverse impact for protected applicant groups (Hunter and Hunter, 1984). Compared to 16 of the most common selection methods, general mental ability tests generate the highest group differences between ethnicities (Ployhart and Holtz, 2008). These group differences disadvantage minority applicants in particular, thereby limiting the benefits of using cognitive ability tests as a pre-selection tool for several reasons: First, the use of potentially biassed selection methods raises ethical and human rights concerns (Yam and Skorburg, 2021). Second, although cognitive ability is related to job performance, diversity is an important hiring outcome associated with firm financial performance (Erhardt et al., 2003; Hunt et al., 2018). Any benefits of hiring for job performance might be reduced if they in turn reduce the diversity of an organisation. Third, violating adverse impact regulations that mandate thresholds for acceptable group differences in selection processes raises legal concerns for employers (U.S. Equal Employment Opportunity Commission, 2021).

One method used by employers and test providers to mitigate adverse impact of cognitive ability tests is to combine them with personality tests. This method was first proposed by Hunter and Hunter (1984). However, the method is not always effective in reducing adverse impact (Ryan et al., 1998; Avis et al., 2002). “Culture-fair” cognitive ability tests attempt to remove the impact of language, literacy and cultural values around rapid performance by removing any items that require explicit use of language, reading, or time constraints and focusing on visuospatial or abstract reasoning instead (Jensen, 1980; Arvey and Faley, 1988; Anastasi and Urbina, 1997). This, again, is not always effective, with culture-fair tests like the Raven’s Progressive Matrices producing group differences at the threshold of adverse impact in a real world selection process (LeBlanc and Chawla, 2003).

1.3. Machine learning based psychometric assessments

In traditional cognitive ability tests, scores are based on counting correct and incorrect answers to a number of items, typically answered within a specified time. Machine learning based scoring instead employs statistical models like regression as the scoring algorithm: Gameplay data predicts scores on an established cognitive ability test. The resulting prediction model is the scoring algorithm e.g., (see Atkins et al., 2014; Quiroga et al., 2015). This type of scoring is commonly used in the emerging field of computational psychometrics, for example to translate a person’s digital footprint into a psychometric profile (Roth et al., 2013; Kosinski et al., 2016), or to measure personality based on language use (Schwartz et al., 2013; Lambiotte and Kosinski, 2014; Park et al., 2015), video interviews (Hickman et al., 2021), or gameplay (Leutner et al., 2020). The digitalised format of game based assessments lends itself to machine learning based scoring because it allows for the collection of several gameplay behaviours such as levels completed or times spent on a single task, rather than simply recording correct or incorrect answers.

Machine learning based scoring algorithms for cognitive ability games have a distinct advantage in the applied selection context: They can be optimised for concurrent validity and, at the same time, for minimising adverse impact, a process described by several machine learning based assessment vendors (Raghavan et al., 2020). This is typically achieved by selecting features that predict the target outcome and deleting or down-weighting features that are also associated with protected group membership like age, gender, or ethnicity (Bogen and Rieke, 2018; Diamond et al., 2020; Leutner et al., 2021; Team PredictiveHire, 2021). The ability to optimise scoring models for fairness gives providers more flexibility in reducing the adverse impact of their assessment procuts, a process that in the past involved costly and resource intensive redesigning of assessments (Raghavan et al., 2020).

A common critique of machine learning models is that they are “black-boxes.” Opaque algorithms obscure what is being assessed and how it might discriminate against different applicants, impacting their human right to work (Yam and Skorburg, 2021). Yam and Skorburg (2021) describe three areas of algorithm opacity: First, the technical literacy of hiring managers who may wrongly interpret scores. This risk is minimised when machine learning based scoring is used to assess existing psychometric constructs like cognitive ability, and when assessments are compared to established measures. Second, a lack of transparency from test publishers that allows for biassed practises. This risk is minimised by peer review that allows for discussion and critique within the scientific community. Finally, representational characteristics of the machine learning models used to score assessments. This risk is reduced by using interpretable models such as regression or certain decision trees, and interpretable features, such as gameplay performance on face and content valid tasks.

1.4. Current limitations

There is a lack of academic studies describing game based assessments of cognitive ability as well as machine learning based assessments for the selection and recruitment context. Although these assessments are described in the literature for learning, development, education, or research contexts, the applied challenges in recruitment and selection are unique in terms of their requirements for transparency, accuracy, and fairness. Recruitment decisions impact the lives of individuals, and recruitment algorithms, or machine learning based assessments, affect applicants’ human rights (Yam and Skorburg, 2021). Crucially, the European Union’s proposed artificial intelligence regulation classes recruitment and selection algorithms as high risk systems, meaning they require regular conformity assessments and must comply with standards of transparency, validity, data management and more (European Commission, 2021).

There is a lack of data and studies on the diversity benefits of game based assessments and machine learning based scoring. Studies are needed to demonstrate whether scoring algorithms optimised for fairness are robust when adapted to different samples (Raghavan et al., 2020). Similarly, game based formats could help reduce the adverse impact of cognitive ability tests. Given studies showing that anxiety negatively impacts test performance, and that gamification reduces anxiety in the serious games and education contexts, studies are needed to explore the link between game based formats, anxiety and assessment scores in recruitment contexts (Richards et al., 2000; Thames et al., 2015; Kiili and Ketamo, 2017). Finally, user experience benefits are a main driver for the adoption of game based assessments (Tippins, 2015). Subjective user experience improvements, however, are largely context dependent (Hamari et al., 2014) and there is a lack of studies on applicant reactions to game based assessments in a job selection context to validate that gamification benefits translate to job applicants.

2. Materials and methods

We use data from 11,574 completed game based cognitive ability assessments, of which 10,924 come from job applicants and 650 from panellists (HireVue, 2021). Across four studies we address limitations in the current literature by describing the development, validity, fairness, and user experience of a machine learning scored, game based assessment of cognitive ability in the recruitment and selection context. In validating the game based assessment, we evaluate the potential of both gamification and machine learning based scoring to alleviate long standing problems with group differences in cognitive ability tests. We evaluate the applicant user experience with game based cognitive ability tests in a real world job application context. We are not aware of existing studies providing a comprehensive overview of a game based assessment used in the applied selection context.

2.1. Study 1: Developing and validating a machine learning based scoring algorithm for a game based assessment: Convergent validity, test–retest reliability

Using a sample of 3,609 panellists and real world job applicants, we develop a machine learning based scoring algorithm for recruitment and selection, based on two short games designed to measure cognitive ability. The algorithm is optimised for convergent validity with an established cognitive ability test as well as minimising group differences on age, gender, and ethnicity (ICAR; The International Cognitive Ability Resource Team, 2014). Gameplay features and scoring model weights are described to provide transparency, and we evaluate test–retest reliability.

2.2. Study 2: Testing adverse impact and fairness in an independent sample of 3,107 job applicants

In order to address a lack of literature on the effectiveness of machine learning based models optimised for fairness, we analyse the performance of a further 3,107 real life job applicants on the game based assessment to determine whether it leads to adverse impact on age, gender and ethnicity. We include adverse impact metrics that address outcome parity as they are typically used in psychometric test evaluation: Adverse Impact Ratio, Cohen’s D, and two Standard Deviations.

2.3. Study 3: The user experience: Test taking anxiety in game based assessments

We explore potential benefits of the game based assessment format on the user experience, and in turn assessment performance and fairness. Given that test taking anxiety can impact gender differences on cognitive ability tests, and given that gamification promises to reduce anxiety by increasing engagement (Osborne, 2007; Kato and de Klerk, 2017), we test the effect of anxiety and gender on game based assessment performance. The experimental study includes data from 85 panellists.

2.4. Study 4: In application: The user experience of job applicants

Given the lack of research on the user experience of gamification in the context of job applications, we review feedback from 4,778 real world job applicants who completed game based cognitive ability assessments. We discuss the implications for the claim that game based assessments deliver a user friendly experience in the job application context.

3. Studies

3.1. Study 1: Developing and validating a machine learning based scoring algorithm for a game based assessment

3.1.1. Data

Two samples are used to train the scoring algorithm: The first is a scoring sample consisting of 565 panellists recruited on Prolific Academic and compensated for their participation. Participants completed the cognitive ability games as well as traditional cognitive ability tests. The second is a mitigation sample consisting of 3,044 job applicants to entry and graduate level positions advertised in the EMEA region, who were part of a group of 6,151 applications. Participants did not need to be native English speakers. Job applicants who asked for accommodations or alternative assessments are not included in the sample. The remaining 3,107 applications are retained for use in Study 2. The two samples were split approximately equally along gender, age and ethnicity to ensure a good representation of each group in each sample. The majority of applicants are leaving university to join the job market at the time of application. They complete the game based assessment as part of their application process, and also provide demographic information. The game based assessment is the first assessment step in the recruitment process, followed by further assessments and interviewing and is completed at a convenient time online. See Table 1 for descriptives.

TABLE 1
www.frontiersin.org

Table 1. Demographics for samples used in studies one and two.

3.1.2. Measures

3.1.2.1. Cognitive ability games

For the machine learning based scoring model we use gameplay data from two games developed to assess cognitive ability: Shapedance and Numerosity (described in Table 2). Gameplay tasks are modelled after traditional cognitive ability tests, and to assess aspects of cognitive ability as specified by the Cattell–Horn–Carroll theory of intelligence (McGrew, 2009). Similar to other cognitive ability tests, the playtime of each game is fixed. Several gameplay behaviour features are tracked for each game.

m a x L e v e l = max l e v e l r e a c h e d b y t h e t e s t t a k e r i n a g a m e [ 0 , 100 ]
f i n a l L e v e l = f i n a l l e v e l r e a c h e d b y t h e t e s t t a k e r i n a g a m e [ 0 , 100 ]
w i n R a t i o = n u m b e r o f s u c c e s s f u l l e v e l s ÷ n u m b e r o f m a t c h e s [ 0 , 1 ]
l e v e l s W o n = n u m b e r o f l e v e l s w o n
l e v e l s L o s t = n u m b e r o f l e v e l s l o s t
t o t a l L e v e l s = t o t a l n u m b e r o f l e v e l s p l a y e d t h r o u g h o u t t h e g a m e ( l e v e l s W o n + l e v e l s L o s t )
TABLE 2
www.frontiersin.org

Table 2. Game descriptions, duration, and example screens.

m a x L e v e l and f i n a l L e v e l measure how far a player advances in relation to the maximum level reachable in the game (level 100 for all games). W i n R a t i o measures how many levels are completed correctly: if the player makes no mistakes at any level, w i n R a t i o = 1 . However, a player can have w i n R a t i o = 1 , but may not perform the task quickly enough to play many levels in the allocated time.

3.1.2.2. Cognitive ability questionnaire

Cognitive ability is assessed by combining the 16-item International Cognitive Ability Resource Sample Test (ICAR; The International Cognitive Ability Resource Team, 2014) and four items from the Cognitive Reflection Test (CRT; Frederick, 2005; Toplak et al., 2014). ICAR is a public-domain assessment tool of cognitive ability, widely used in psychological and social scientific research and recommended for use in high-stakes testing, including job selection and clinical purposes. ICAR has four item types: Three-Dimensional Rotation items (four items); Letter and Number Series (four items); Matrix Reasoning (four items); and Verbal Reasoning logic questions (four items). The CRT measures deliberative/rational thinking, the tendency to override a response alternative that is incorrect, and to engage in further reflection that leads to the correct response. The CRT predicts performance on rational thinking tasks even when cognitive ability and thinking dispositions are taken into account (Frederick, 2005; Toplak et al., 2014). CRT items were included to provide a broader assessment of cognitive ability.

3.1.3. Results

3.1.3.1. Game based assessment scoring algorithm

To generate the scoring algorithm, gameplay behaviour features from cognitive ability games are entered into a prediction model as predictors of scores on the cognitive ability questionnaire. A variety of machine learning models with 10-fold cross validation are tested as scoring methods including Classical Regression with regularisation, Random Forests, Naive Bayes, and Support Vector Machines (Bergstra and Bengio, 2012). Ridge Regression performs best and is chosen as the scoring model. Additionally, ridge regression models have high explainability and are easily interpretable.

To minimise subgroup differences, a novel Bias Penalisation method is used to further optimise the Ridge Regression model (see Rottman et al., 2021, manuscript submitted for publication1). This method models the diversity-validity trade-off by adding a bias penalty term to the model optimization during the fitting process. The result is a model that minimises subgroup differences whilst maintaining high predictive accuracy. Data from both scoring and mitigation samples is used to determine penalisation within the models. The algorithm thus uses data from both samples to generate a scoring algorithm that is optimised for predicting cognitive ability as well as minimising bias.

Overall model performance is estimated by Pearson correlation between observed and predicted cognitive ability scores to allow for easy interpretability. Correlations are reported based on the aggregate of out-of-sample cross validation prediction across all folds.

3.1.3.2. Descriptive statistics

Descriptive statistics for the questionnaire and game based cognitive ability assessments are presented in Table 3.

TABLE 3
www.frontiersin.org

Table 3. Descriptive statistics for the cognitive ability questionnaire, cognitive game based assessment, and cognitive ability games features in the scoring and validation samples as well as correlations between game behaviour and cognitive ability scores.

3.1.3.3. Model performance and concurrent validity

The machine learning based scoring algorithm achieves a concurrent validity of r = 0.5 (95% CI from Fisher transformation = 0.43–0.56) with cognitive ability as measured by the cognitive ability questionnaire. See Figure 1 for an illustration of the relationship between the Game Based Assessment scores and cognitive ability scores.

FIGURE 1
www.frontiersin.org

Figure 1. Observed questionnaire based and predicted game based cognitive ability scores using the ridge regression model with bias penalisation (Pearson’s r = 0.50).

3.1.3.4. Explainability and feature importance

Feature relative importance is computed for the scoring algorithm to determine which gameplay behaviours contribute more or less to the Game Based Assessment score (see Figure 2). Features from both cognitive ability games are represented in the scoring algorithm, with the final level reached on the Numerosity game and the total levels played on the Shapedance game being the two most important features in the scoring algorithm. Additionally, Table 3 shows the correlation between each game behaviour and cognitive ability scores on the questionnaire as well as the game based measures. Numerosity_finalLevel correlates strongly with both questionnaire and game based cognitive ability scores (r = 0.44 and r = 0.58, respectively). Shapedance_totalLevels correlates strongly with game based and moderately with questionnaire based cognitive ability (r = −0.19 and r = −0.63, respectively).

FIGURE 2
www.frontiersin.org

Figure 2. Feature relative importance for the machine learning scoring model.

3.1.3.5. Test–retest reliability

A subset of 102 participants from the scoring sample completed the game based assessment again within 4 months of their first completion (mean age 34 with SD = 10.20, 53% male, 27% Asian, 20% Black, 11% Hispanic, 42% White). Test–retest reliability is computed as the correlation between assessment scores at first and second completion and is moderate to high with Pearson’s r = 0.68 (p < 0.001) (see Figure 3).

FIGURE 3
www.frontiersin.org

Figure 3. Test–retest reliability and convergent validity for the game based cognitive ability assessment.

3.1.4. Discussion

The convergent validity of r = 0.50 achieved for the game based cognitive ability test with using Bias Penalisation compares to validities described in the literature: When correlating game based with single or separate established cognitive ability measures obtain r values of around r = 0.30 to 0.45 (e.g., McPherson and Burns, 2008; Atkins et al., 2014). Though studies using structural equation modelling to correlate latent factors derived from several traditional tests with those derived from a range of games report higher correlations ranging from r = 0.72 to 0.93 (Quiroga et al., 2015; Peters et al., 2021).

Study two tests whether the scoring algorithm with Bias Penalisation performs well in practise when used on a dataset of job applicants that were not known during model training. This is important given a lack in the literature of studies describing how bias mitigation methods might behave vis a vis previously unseen samples (Raghavan et al., 2020).

3.2. Study 2: Testing adverse impact and fairness in an independent sample of 3,107 job applicants

We test the generalisability of the adverse impact performance of our scoring algorithm generated in Study 1. This is done by evaluating the adverse impact for a set of participants not contained in the scoring algorithm dataset. Participants are scored using the scoring algorithm generated in Study one. Adverse Impact metrics are then computed for the participants and evaluated.

3.2.1. Data

We use data from 3,107 further job applicants to graduate and entry level roles, as described in Study one. All applicants complete the game based assessment with Shapedance and Numerosity that we describe in Study one.

3.2.2. Method

Adverse impact and group differences are tested using several metrics: The Uniform Guidelines on Employee Selection Procedures state that a selection rate for any race, sex, or ethnic group which is less than 4/5ths of the rate for the group with the highest passing rate will generally be regarded as evidence of adverse impact (Uniform Guidelines on Employee Selection Procedures, 1978; American Educational Research Association, American Psychological Association, and National Council on Measurement in Education, 1999; Society for Industrial and Organizational Psychology, Inc, 2003). Whilst the 4/5th rule presents a quick measure, the guidelines as well as professional standards recommend statistical measures also be used to establish whether adverse impact is present. We evaluate validity, reliability and adverse impact in line with these psychometric test guidelines and professional testing standards and include tests of proportion or passing rate differences (e.g., Fisher’s exact), and statistical significance tests of group differences (e.g., two standard deviations, and/or Cohen’s D).

3.2.3. Results

See Table 4 for practical and statistical measures of Adverse Impact for the game based assessment of cognitive ability described in Study one. The different measures show no adverse impact on any of the evaluated groups. Although the two standard deviations measure exceeds accepted thresholds for female applicants, this test is overpowered when sample sizes exceed 100, which is the case here. Fisher’s exact test is marginally significant, but both the Adverse Impact Ratio and Cohen’s D show acceptable levels of group differences between male and female applicants.

TABLE 4
www.frontiersin.org

Table 4. Adverse impact analysis of the game based cognitive ability assessment.

3.2.4. Discussion

The adverse impact metrics illustrate the game based cognitive ability assessment falls within acceptable ranges of group differences, even when tested on applicant data that was not used as part of the scoring model generation. That is, the scoring model generates acceptable ranges of group differences when used on new data. This indicates that the game based format and the machine learning based model with Bias Penalisation might have a positive impact on fairness in cognitive ability assessments. However, additional samples are needed to validate the generalisability of the Bias Penalisation to other recruitment contexts, such as other job positions, regions or job levels. In Study three, we look at game mechanisms that might contribute to reducing group differences in game based cognitive ability tests by reducing anxiety.

3.3. Study 3: The user experience: Test taking anxiety in game based assessments

A possible avenue to reduce the adverse impact of cognitive ability is reducing test taking anxiety. Test taking anxiety refers to cognitive and behavioural responses accompanying concern about possible negative consequences arising from an evaluative situation (Zeidner, 1998). Test taking anxiety causes poor academic performance (Hembree, 1988; Zeidner, 1998; Cassady and Johnson, 2002), and women tend to experience higher levels of test anxiety {Devine et al., 2012 (maths test); von der Embse et al., 2018 [meta analysis including classroom testing, Grade Point Average (GPA), IQ and standardised exams (e.g., SAT)]}. Game-based assessments may offer a way to combat test anxiety, and its adverse effect on performance. Individuals have positive experiences interacting with games used for serious purposes (Shen et al., 2009). Simple game-like artefacts increase participant enjoyment and interest in activities (Lieberoth, 2015). This study explores whether the gamified cognitive ability assessments result in test taking anxiety, and whether this anxiety impacts performance and does so differently for men and women.

The analysis first tests gender differences in game performance and test taking anxiety experienced as a result of playing the games. Second we use multiple regression to test the effect of test taking anxiety and gender on game performance, as well as the interaction between test taking anxiety and gender on game performance.

3.3.1. Data

The study includes 85 participants recruited from Prolific Academic and compensated for their participation, 52 (61%) are women and 33 (39%) men with a mean age of 34 (SD = 11.45). The sample is predominantly white (86%) and educated beyond highschool (70%). The dataset excludes thirteen additional participants that were removed due to poor response variability, missing answers, or poor game performance indicating no valid attempt.

3.3.2. Measures

3.3.2.1. Test taking anxiety

Test taking anxiety is the difference between an individual’s pre assessment test anxiety and their post assessment test anxiety, with higher test taking anxiety scores indicating lower test taking anxiety after the assessment compared to before. Test anxiety is measured using adapted versions of the Test Anxiety Inventory, answered on a four point Likert scale, from strongly disagree to strongly agree (TAI; Spielberger et al., 1980). TAI measures symptoms of test anxiety experienced before, during, and after testing situations. It has good test–retest and internal-consistency reliability, and acceptable concurrent, construct and discriminant validity (Spielberger et al., 1980; Szafranski et al., 2012; Ali and Mohsin, 2013). Pre assessment test anxiety is measured as a baseline with the five item Test Anxiety Inventory Short Form (TAI-5; Taylor and Deane, 2002). Post assessment test anxiety is measured using an adapted thirteen item version of the TAI (Spielberger et al., 1980), where items referencing future outcomes related to test performance and test feedback are removed, and items are re-worded to measure present test anxiety resulting from the game based assessments. The two inventories are comparable on patterns of correlation, means and standard deviations and show good internal consistency with Chronbach’s Alpha TAI-5 = 0.89; TAI = 0.93 (see Table 5 for descriptives and Table 6 for correlations).

TABLE 5
www.frontiersin.org

Table 5. Means, standard deviations (SD) and group differences between men and women on test taking anxiety and game performance.

TABLE 6
www.frontiersin.org

Table 6. Correlations between test taking anxiety and game performance.

3.3.2.2. Numerosity, pathfinder, shapedance, and singularity game scores

Individual game scores are calculated by multiplying the participants max level achieved by their win ratio.

3.3.2.3. Games total score

Sum of the standardised individual game scores for each of the four games, describing overall game performance.

3.3.3. Results

No significant differences between men and women are observed for performance on the games total score and on individual games. Although women report significantly higher baseline test anxiety before playing the games than men, there are no significant gender differences on test taking anxiety experienced as a result of playing the games (see Table 5).

There is a significant correlation between test taking anxiety and game performance for the games total score as well as performance on the Shapedance game, such that those who experience less test taking anxiety after playing the games perform better (see Table 6). However, test taking anxiety does not significantly predict game performance: multiple regressions are computed predicting the games total score from test taking anxiety, gender and the interaction of test taking anxiety and gender. The interaction term is included to assess whether test taking anxiety affects men and womens’ performance differently. There are no significant main effects of gender or test taking anxiety on game performance. However, there is a significant interaction effect of gender and test taking anxiety on the games total score. When computing the model for each game individually, the relationship is significant for the Numerosity score only, but not for the remaining game scores. See Table 7 for results of multiple regression models. Interaction plots reveal that men perform better on the games total and Numerosity scores when they experience less test taking anxiety (see Figure 4). Test taking anxiety has no significant effect on women’s performance.

TABLE 7
www.frontiersin.org

Table 7. Multiple regression predicting game performance from gender, test taking anxiety, and the interaction of gender and test taking anxiety.

FIGURE 4
www.frontiersin.org

Figure 4. Game performance by anxiety change and gender. A higher test taking anxiety score indicates lower test taking anxiety experienced after playing the games than baseline test taking anxiety measured before playing the games.

3.3.4. Discussion

Women experience higher test taking anxiety at baseline than men, in line with previous studies (Núñez-Peña et al., 2016). However, there is no significant difference in test taking anxiety when adjusting for baseline test taking anxiety, indicating that the games evoke similar levels of anxiety in women and men. As in previous studies showing that test anxiety negatively impacts academic performance, and performance on cognitive ability tests, test taking anxiety correlates with lower game performance in this study (Hembree, 1988; Reeve and Bonaccio, 2008). However, this does not appear to negatively impact women: no gender differences in game performance are observed on the total Games Score or the single games. We also find that test taking anxiety affects the performance of men negatively, in particular on the Numerosity game, but not that of women. This result may be explained with the sex-linked anxiety coping theory (McCarthy and Goffin, 2005). This theory posits that differences in coping strategies mean that although women experience higher levels of test anxiety, they also engage in more coping behaviours (e.g., positive self-talk) than their male counterparts. As a result, the relationship between test anxiety and test performance is weaker for women than men, affecting men’s performance negatively but not women’s.

3.4. Study 4: In application: The user experience of job applicants

3.4.1. Data

In this study, we use a dataset of 4,778 job applicants to Graduate roles in a large consulting firm in Eastern Europe. Applicants are typically recent graduates from university and are local to any of the Eastern European countries in scope, which includes countries like Poland, Czech Republic and Slovakia. Demographic data on these applicants is not available due to data protection legislation.

3.4.2. Method

All applicants complete a game-based assessment designed to assess cognitive ability as part of their job application. Applicants apply to technical Graduate roles in the company (e.g., IT roles) and are presented with a cognitive game package that includes three games: Numerosity, Singularity and Shapedance. For applicants in some countries, the game package is combined with an English language test or a business specific technical test.

We collect both quantitative and qualitative data from these applicants. Quantitative data is available from 4,778 candidates, who answer nine questions related to overall satisfaction with the assessment, ease of use, available support and resources for completing the assessment, relevance, and length of the assessment (Table 8 lists all nine areas). The candidates record their answers on a 5-point Likert scale (1 = Highly Dissatisfied; 5 = Highly Satisfied). The candidates also answer the standard Net Promoter Score (NPS) question: “Based on the experience you just had, how likely are you to recommend [Company name] to a friend or colleague?” on a 1–10 scale.

TABLE 8
www.frontiersin.org

Table 8. Descriptive statistics of the user experience questionnaire.

We also collect qualitative comments from candidates using open-ended questions about the assessment experience and whether they have any recommendations or suggestions for improvement. We only include comments in the English language, which amount to 685 usable, substantive comments. These comments are analysed via Thematic Analysing (Braun and Clarke, 2006).

3.4.3. Results

The NPS obtained is 58. The NPS scale ranges from-100 to +100, where the percentage of detractors (answering zero to six out of ten) is subtracted from the percentage of promoters (answering nine or ten out of ten). An NPS of 58 indicates a good user experience, comparing favourably to average NPS scores achieved in several industries and product areas (average NPS score 43 in professional services, 35 for technology companies, and 43 for consumer goods companies; Gitlin, 2021).

Table 8 presents the results for the user experience questionnaire. The analysis of the user experience questionnaire also indicates the candidates are overall satisfied with the assessment, they perceive sufficient support and resources, enough preparation time, they are satisfied with the response they had prepared, and they are also satisfied with the answering time provided. The assessment aspects where the applicants show the lowest satisfaction are related to the length of the assessment (i.e., too many questions), the perceived relevance of the assessment, and the degree to which the assessment enabled the applicants to showcase their skills.

The qualitative analyses reveal a number of prevalent themes within candidates’ comments reflecting a positive experience, including positive attitudes (“fun,” “candidate friendly,” “good support,” “more games,” “this is the future”) and overall satisfaction with the format (“everything is perfect as is,” “no further suggestions as everything is good”). At the same time, the analysis reveals several areas of potential improvement for the game-based assessments, including a desire for more practise time to prepare for the assessment and better instructions, and a desire for more technical/job-related content within the assessment.

3.4.4. Discussion

User experience questionnaires obtained from real life job applicants indicate that game based assessments deliver a favourable experience for most candidates. Areas of improvement highlighted both in qualitative and quantitative feedback appear to relate to the abstract nature of cognitive ability tests rather than the game based format: “Relevant assessment” and “showcase skills” ranked lowest in terms of satisfaction, and verbal comments reflected a lack of understanding for the job relevance of the test. This highlights the need to frame assessments in relation to the relevant job competencies they measure, something that may be even more important with game based assessments that have a less serious test-like feel.

4. Discussion

With this collection of studies we seek to advance the literature on game based as well as machine learning based assessments, and to highlight the potential of both methodologies to improve the fairness and thereby effectiveness of cognitive ability assessments. We describe and validate a machine learning based scoring algorithm optimised for fairness with good convergent validity (r = 0.50), comparable to values described in previous studies ranging from r = 0.30 to.93 (McPherson and Burns, 2008; Atkins et al., 2014; Quiroga et al., 2015; Peters et al., 2021). Notably, this is the case when using the novel Bias Penalisation method that optimises the scoring algorithm to minimise group differences at the same time as maximising convergent validity (Rottman et al., 2021, manuscript submitted for publication1). This penalisation might have reduced convergent validity, but results suggest validities are comparable to those reported elsewhere without employing penalisation. The resulting assessment provides a valid estimate of cognitive ability that should be accurate enough for the purpose of pre-hire assessments, and at the same time reduces group differences to a level accepted in application. For example, in our illustration in Study 2, adverse impact metrics are within acceptable ranges for selecting out the bottom 33% of applicants, a cut score that is useful for hiring managers in practise.

4.1. Generalisability of fairness optimised machine learning scoring algorithm

Like mitigation methods described by assessment providers, the Bias Penalisation is successful at producing a fair assessment in terms of outcome parity (Bogen and Rieke, 2018; Diamond et al., 2020; Leutner et al., 2021; Team PredictiveHire, 2021). However, to our knowledge, Study two is the first to describe the generalisability of a fairness optimised machine learning algorithm for a game based assessments in the scientific literature. The scoring algorithm produced levels of group difference within accepted ranges in a separate sample, providing a first indication that the dual purpose optimization may offer a successful methodology for reducing group differences across samples. However, the two samples are similar in that they were generated from a pool of applications in the same geographical region and with the same job roles and levels represented in both samples. If this finding can be replicated with more heterogeneous samples the described methodology might offer an option for reducing the adverse impact of cognitive ability tests and aligning their use with diversity goals. Generalizability is important in the applied context, as cognitive ability assessments are often used in high volume recruitment by international organisations that use assessments across regions, job levels and functions (Raghavan et al., 2020).

4.2. Anxiety and game performance

We found no significant direct prediction of test anxiety for game performance, and no significant gender differences in anxiety produced by the games. This provides additional empirical evidence for claims that gamification can reduce test taking anxiety, as described in previous literature (Mavridis and Tsiatsos, 2017), as well as being a positive indicator for the gender fairness of the games. For women, higher anxiety also did not lead to lower game performance, although this was the case for men on one of the four games tested. The significant relationship for men is of less immediate practical concern given that men tend to be the baseline group with the highest passing rate both on cognitive ability tests as well as the most likely group to be recruited. However, it might highlight potential discrimination for those with test taking anxiety in an assessment context. More studies are needed to replicate this finding and uncover the role of different game mechanics in reducing anxiety, or weakening the link between anxiety and performance.

4.3. Applicant reactions

We review applicant feedback showing a good user experience of cognitive ability games in the real world job application context with an NPS score of 58. This is in line with previous studies suggesting the favourable user experience of game based assessments (Guin et al., 2012). However, this data is, to our knowledge, the first describing user experience in the recruitment context, and provides some validation for claims that gamification is beneficial in the job application process. The feedback suggests that game mechanics that induce enjoyment and flow in other areas of application do indeed also produce a favourable game experience in the job selection context. Feedback contained concerns around the perceived applicability of cognitive ability tasks to job performance, an issue commonly raised with cognitive ability tests (Smither et al., 1993; Hausknecht et al., 2004). The face validity aspect of the user experience in game based assessments is not prominently covered in the literature on gamification, but is of particular interest for the recruitment context: Low face validity has a negative effect on applicant perceptions as well as test performance, although the effect on test performance can be mitigated by high test taker motivation, which is likely to occur in job applications (Chan, 1997). Study four illustrates the importance of evaluating gamification in the applied recruitment context.

4.4. Impact of technology on cognitive ability testing

With regards to the use of machine learning and game based assessment for the measurement of cognitive ability, this collection of studies raises two broader points: First, that technology can be used to enhance current cognitive ability tests and that its adaptation in practise and research is warranted. And second that the use of these technologies inevitably changes, at least to some degree, what cognitive ability tests measure, and thereby how cognitive ability relates to real life outcomes of interest, including job performance.

Regarding the first point, we argue in this paper that the reduction of group differences and improved user experience are the two technological benefits most relevant in the applied selection context. Game mechanisms are predominantly discussed for their user experience benefits (Landers and Callan, 2011; Huotari and Hamari, 2012). With study three we demonstrate the point that the user experience invoked through gamification has a direct impact on fairness outcomes by dis-or encouraging the participation and performance of applicants from different protected groups. Research on variations in group differences as a result of assessment format exists (Chan and Schmitt, 1997), but has not been explored in the game based assessment space to our knowledge.

Equally, machine learning scoring algorithms are typically discussed as a means to improve fairness (Diamond et al., 2020; Raghavan et al., 2020; Team PredictiveHire, 2021), but they also offer user experience benefits, for example through shorter assessment times (Leutner et al., 2020). Scoring model optimisation methodologies like described in Studies one and two are relatively common in application but rarely described in the academic literature. We hope to provide an applied view of the potential benefits of these scoring methodologies in reducing group differences. They are particularly encouraging as they do not require changing the assessment content itself, a significant practical advantage (Raghavan et al., 2020). However, optimising for outcome parity on known group differences is a narrow view of fairness that does not solve all equality problems in cognitive ability testing. Importantly, optimization can only work on known protected groups. In practise, these are groups that are routinely recorded during the application process such as age, gender, and ethnicity, and that also have large volumes within each group, excluding, for example, smaller ethnic minority groups. Even when mitigated through optimised scoring algorithms, group differences in cognitive ability tests exist. Cognitive ability testing remains a fairness concern and must be carefully balanced with diversity goals when used in application. Factors that influence group differences on cognitive ability tests need to be further explored in the context of applied selection in order to provide a favourable testing environment that is set up for fairness. This includes not only the tests themselves, their framing and face validity, but also the wider recruitment context like job advertisements, employer image, and organisational values and make up.

Regarding the second point, the convergent validity we observe in Study one is high, but variance in gameplay behaviour remains unexplained. New technologies might help deliver fairer and more engaging assessments, but this is only useful in practise if they measure skills and abilities that are relevant at work: The justification for using cognitive ability tests despite the group differences they produce is their consistent relationship with job performance (Schmidt and Hunter, 1998; Kuncel et al., 2010). By losing convergent validity with traditional tests that have high criterion validity, game based assessments might produce a weaker link with job performance. This needs to be tested by relating games to criterion outcomes, and exploring whether unexplained variance relates to job performance. However, even without criterion data, convergent validity as well as content validity, interpreted together with decades of research into the predictive validity of cognitive ability for a range of job roles and level, provide a good foundation to justify the use of game based assessments of cognitive ability in practise. This justification is further strengthened if a game based assessment delivers better outcome parity compared to traditional tests.

4.5. Limitations and directions for future work and practise

As a validation and demonstration of game based assessments for the applied selection, this paper has some limitations with regards to concurrent validity: The impact of bias penalisation and gamified format changes on predictive validity for job performance are unclear. There is a clear gap in the literature here that needs to be addressed in future studies. Additionally, each study presents its own limitations. Studies one and two demonstrate that fairness optimised scoring algorithms work on data unseen during scoring model generation. However, further samples from different cultures, languages, employers, and job levels should be explored in order to test the generalisability of the optimised algorithm. This should be done in a structured way to test how optimization responds to data that is different on each of the above characteristics. Tests should additionally include protected groups unknown to the algorithm, such as smaller minority ethnic or neurodiverse groups. Adverse Impact evaluation needs to be an ongoing project for any cognitive ability assessment used in application.

Study three presents clear limitations due to its small sample size and restriction to gender, with ethnic and other protected groups excluded from the analysis. More research is needed to understand the test taking experience of different groups, and how anxiety, stereotype threat, or other characteristics might affect performance on game based assessments. This will help test developers design assessments that produce fewer group differences. The study was also conducted on panellists rather than real world applicants. Given the impact of high versus low stakes settings on test taking behaviour, this research should preferably be conducted in the applied selection context. If possible, test providers’ user experience questionnaires should include measures of test taking anxiety, face validity and other relevant aspects. We want to see more research published on game based assessments used in selection processes.

Study four uses feedback collected from real job applicants, which presents several limitations for analysing and interpreting results. First, some applicants provided feedback directly after playing the games, whilst others took additional tests before providing feedback. This will have affected their responses. Second, applicants did not take any traditional, non-gamified tests and a comparison of their user experience between gamified and non-gamified tests was therefore not possible. We use the Net Promoter score to provide a standardised metric of user experience compared to user experiences in other industries. However, an industry average Net Promoter score is, to our knowledge, not available for psychometric assessments or job applications. We encourage the use of Net Promoter scores in future studies on game based assessments as well as for assessment providers.

5. Conclusion

We find that game based assessments of cognitive ability work. They accurately measure cognitive ability, showing good convergent validity with traditional cognitive ability tests, Furthermore, we find that the use of gamification and machine learning based scoring with bias penalisation might help solve the biggest practical issue with cognitive ability tests, their bias, by minimising group differences. This benefit seems to be retained in real world selection applications, at least in the context we are able to evaluate. Anxiety did not impact performance on game based cognitive ability tests, providing some evidence that gamification is beneficial for test taking anxiety. Additionally, we find that the user experience of real world job applicants who take game based assessments of cognitive ability is positive, further supporting their use in applied settings. Gamification did not eliminate face validity concerns of applicants that are typical in cognitive tasks. This is of high practical importance and must be addressed.

Overall, our data show that game based formats and machine learning based scoring deliver key benefits of practical importance, namely increased fairness and user experience. They also offer avenues for the advancement of psychometric test design, for example by allowing the optimisation of scoring keys to fairness outcomes. These new technologies should therefore be investigated and further developed in psychometric testing. Further investigation into their use is encouraged.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by the Ethics Committee of the Division of Psychology and Language Sciences, Department of Experimental Psychology, University College London. The patients/participants provided their written informed consent to participate in this study.

Author contributions

FL, S-CC, and TB contributed to the data analysis and write up of Studies 1–4, with TB leading on data analysis. SB and FL conducted Study 4. All authors contributed to the article and approved the submitted version.

Acknowledgments

We thank the Data Science, Industrial Organizational Psychology, and Product teams at HireVue, as well as Clemens Aichholzer, Luca Boschetti, and Maurizio Attisani for their support and their work developing the game based assessments described in this article.

Conflict of interest

Authors FL, SB, and TB were employed at HireVue at the time of writing.

The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^Rottman, C., Gardner, C., Liff, J., Mondragon, N., and Zuloaga, L. (2021). New Strategies for Addressing the Diversity-Validity Dilemma with Big Data. J. Applied Psychology.

References

Albuquerque, J., Bittencourt, I. I., Coelho, J. A., and Silva, A. P. (2017). Does gender stereotype threat in gamified educational environments cause anxiety? An experimental study. Comput. Educ. 115, 161–170. doi: 10.1016/j.compedu.2017.08.005

CrossRef Full Text | Google Scholar

Ali, M. S., and Mohsin, M. N. (2013). Test anxiety inventory (TAI): factor analysis and psychometric properties. J. Hum. Soc. Sci. 8, 73–81. doi: 10.9790/0837-0817381

CrossRef Full Text | Google Scholar

American Educational Research Association, American Psychological Association, and National Council on Measurement in Education (1999). Standards for Educational and Psychological Testing. Washington, DC: American Educational Research Association.

Google Scholar

Anastasi, A., and Urbina, S. (1997). Psychological Testing Prentice. 7th edn. Hall/Pearson Education.

Google Scholar

Arvey, R. D., and Faley, R. H. (1988). Fairness in Selecting Employees. 2nd edn. Addison Wesley Publishing Company.

Google Scholar

Atkins, S. M., Sprenger, A. M., Colflesh, G. J., Briner, T. L., Buchanan, J. B., Chavis, S. E., et al. (2014). Measuring working memory is all fun and games. Exp. Psychol. 61, 417–438. doi: 10.1027/1618-3169/a000262

PubMed Abstract | CrossRef Full Text | Google Scholar

Avis, J. M., Kudisch, J. D., and Fortunato, V. J. (2002). Examining the incremental validity and adverse impact of cognitive ability and conscientiousness on job performance. J. Bus. Psychol. 17, 87–105. doi: 10.1023/A:1016200317002

CrossRef Full Text | Google Scholar

Bergstra, J., and Bengio, Y. (2012). Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305. doi: 10.5555/2503308.2188395

CrossRef Full Text | Google Scholar

Bogen, M., and Rieke, A. (2018). Help wanted: an examination of hiring algorithms, equity, and bias. Upturn. Available at: https://apo.org.au/node/210071

Google Scholar

Braun, V., and Clarke, V. (2006). Using thematic analysis in psychology. Qual. Res. Psychol. 3, 77–101. doi: 10.1191/1478088706qp063oa

CrossRef Full Text | Google Scholar

Cassady, J. C., and Johnson, R. E. (2002). Cognitive test anxiety and academic performance. Contemp. Educ. Psychol. 27, 270–295. doi: 10.1006/ceps.2001.1094

CrossRef Full Text | Google Scholar

Chan, D., and Schmitt, N. (1997). Video-based versus paper-and-pencil method of assessment in situational judgment tests: Subgroup differences in test performance and face validity perceptions. J. Applied Psychology 82, 143–159. doi: 10.1037/0021-9010.82.1.143

CrossRef Full Text | Google Scholar

Chan, D. (1997). Racial subgroup differences in predictive validity perceptions on personality and cognitive ability tests. J. Appl. Psychol. 82, 311–320. doi: 10.1037/0021-9010.82.2.311

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, J. (2007). Flow in games (and everything else). Commun. ACM 50, 31–34. doi: 10.1145/1232743.1232769

CrossRef Full Text | Google Scholar

Connolly, T. M., Boyle, E. A., MacArthur, E., Hainey, T., and Boyle, J. M. (2012). A systematic literature review of empirical evidence on computer games and serious games. Comput. Educ. 59, 661–686. doi: 10.1016/j.compedu.2012.03.004

CrossRef Full Text | Google Scholar

Dandashi, A., Karkar, A. G., Saad, S., Barhoumi, Z., Al-Jaam, J., and El Saddik, A. (2015). Enhancing the cognitive and learning skills of children with intellectual disability through physical activity and edutainment games. Int. J. Distrib. Sens. Netw. 11:165165. doi: 10.1155/2015/165165

CrossRef Full Text | Google Scholar

Devine, A., Fawcett, K., Szűcs, D., and Dowker, A. (2012). Gender differences in mathematics anxiety and the relation to mathematics performance while controlling for test anxiety. Behav. Brain Funct. 8, 1–9. doi: 10.1186/1744-9081-8-33

CrossRef Full Text | Google Scholar

Diamond, D., Weisberger, D., Baker, L., Ward, M., Murali, V., Knaide, J., et al. (2020). audit-AI. GitHub. Available at: https://github.com/pymetrics/audit-ai

Google Scholar

Dusek, J. B. (1980). “The development of test anxiety in children” in Test Anxiety: Theory, Research and Applications (pp. 87–110). ed. I. G. Sarason (Hillsdale, NJ: Lawrence Erlbaum)

Google Scholar

Erhardt, N. L., Werbel, J. D., and Shrader, C. B. (2003). Board of director diversity and firm financial performance. Corp. Gov. 11, 102–111. doi: 10.1111/1467-8683.00011

CrossRef Full Text | Google Scholar

European Commission (2021). Regulation of the European Parliament and of the Council: Laying Down Harmonised Rules on Artificial Intelligence (Artificial Intelligence Act) and Amending Certain Union Legislative Acts. Available at: https://eur-lex.europa.eu/legal-content/EN/TXT/?qid=1623335154975&uri=CELEX%3A52021PC0206

Google Scholar

Frederick, S. (2005). Cognitive reflection and decision making. J. Econ. Perspect. 19, 25–42. doi: 10.1257/089533005775196732

CrossRef Full Text | Google Scholar

Gamberini, L., Cardullo, S., Seraglia, B., and Bordin, A. (2010). Neuropsychological testing through a nintendo wii® console. Annu. Rev. Cyberther. Telemed. 8, 29–33. doi: 10.3233/978-1-60750-561-7-29

CrossRef Full Text | Google Scholar

Gitlin, J. (2021). What is a good net promoter score? And how does it vary across industries? SurveyMonkey. Available at: https://www.surveymonkey.co.uk/curiosity/what-is-a-good-net-promoter-score/

Google Scholar

Guin, T. D. L., Baker, R., Mechling, J., and Ruyle, E. (2012). Myths and realities of respondent engagement in online surveys. Int. J. Mark. Res. 54, 613–633. doi: 10.2501/IJMR-54-5-613-633

CrossRef Full Text | Google Scholar

Hamari, J., Koivisto, J., and Sarsa, H. (2014). Does gamification work?- a literature review of empirical studies on gamification. In Proceeding 2014 47th Hawaii International Conference on System Sciences, 3025–3034. Waikoloa, HI: IEEE. doi: 10.1109/HICSS.2014.377

CrossRef Full Text | Google Scholar

Hausknecht, J. P., Day, D. V., and Thomas, S. C. (2004). Applicant reactions to selection procedures: an updated model and meta-analysis. Pers. Psychol. 57, 639–683. doi: 10.1111/j.1744-6570.2004.00003.x

CrossRef Full Text | Google Scholar

Heimburger, L., Buchweitz, L., Gouveia, R., and Korn, O. (2020). “Gamifying onboarding: how to increase both engagement and integration of new employees” in Advances in Social and Occupational Ergonomics. AHFE 2019. Advances in Intelligent Systems and Computing. eds. R. Goossens and A. Murata, vol. 970 (Cham: Springer)

Google Scholar

Hembree, R. (1988). Correlates, causes, effects, and treatment of test anxiety. Rev. Educ. Res. 58, 47–77. doi: 10.3102/00346543058001047

CrossRef Full Text | Google Scholar

Hickman, L., Bosch, N., Ng, V., Saef, R., Tay, L., and Woo, S. E. (2021). Automated video interview personality assessments: reliability, validity, and generalizability investigations. J. Appl. Psychol. 107, 1323–1351. doi: 10.1037/apl0000695

PubMed Abstract | CrossRef Full Text | Google Scholar

Higgins, D. M., Peterson, J. B., Pihl, R. O., and Lee, A. G. (2007). Prefrontal cognitive ability, intelligence, big five personality, and the prediction of advanced academic and workplace performance. J. Pers. Soc. Psychol. 93, 298–319. doi: 10.1037/0022-3514.93.2.298

PubMed Abstract | CrossRef Full Text | Google Scholar

HireVue. (2021). HireVue for Candidates (4.6.3) [Mobile App]. Google Play Store. Available at: https://play.google.com/store/apps/details?id=com.hirevue.candidate.generic&hl=en_GB&gl=US

Google Scholar

Hunt, V., Prince, S., Dixon-Fyle, S., and Yee, L. (2018). Delivering through diversity. McKinsey Company. 1–39. Available at: https://www.immigrationresearch.org/system/files/Delivering-through-diversity_full-report.compressed-min.pdf

Google Scholar

Hunter, J. E., and Hunter, R. F. (1984). Validity and utility of alternative predictors of job performance. Psychol. Bull. 96, 72–98. doi: 10.1037/0033-2909.96.1.72

CrossRef Full Text | Google Scholar

Huotari, K., and Hamari, J. (2012). Defining gamification: a service marketing perspective. In Proceedings of the 16th International Academic Mind Trek Conference, 17–22. New York, NY: ACM

Google Scholar

Jennett, C., Cox, A. L., Cairns, P., Dhoparee, S., Epps, A., Tijs, T., et al. (2008). Measuring and defining the experience of immersion in games. Int. J. Hum. Comput. Stud. 66, 641–661. doi: 10.1016/j.ijhcs.2008.04.004

CrossRef Full Text | Google Scholar

Jensen, A. R. (1980). Bias in Mental Testing. New York: Free Press.

Google Scholar

Jimison, H. B., Pavel, M., Bissell, P., and McKanna, J. (2007). A framework for cognitive monitoring using computer game interactions. Stud Health Technol Inform. 129,1073–1077.

PubMed Abstract | Google Scholar

Jimison, H., Pavel, M., McKanna, J., and Pavel, J. (2004). Unobtrusive monitoring of computer interactions to detect cognitive status in elders. IEEE Trans. Inf. Technol. Biomed. 8, 248–252. doi: 10.1109/TITB.2004.835539

PubMed Abstract | CrossRef Full Text | Google Scholar

Judge, T. A., Higgins, C. A., Thoresen, C. J., and Barrick, M. R. (1999). The big five personality traits, general mental ability, and career success across the life span. Pers. Psychol. 52, 621–652. doi: 10.1111/j.1744-6570.1999.tb00174.x

CrossRef Full Text | Google Scholar

Kato, P., and de Klerk, S. (2017). Serious games for assessment: Welcome to the jungle. J. Applied Testing Technology 18, 1–6.

Google Scholar

Kiili, K., and Ketamo, H. (2017). Evaluating cognitive and affective outcomes of a digital game-based math test. IEEE Trans. Learn. Technol. 11, 255–263. doi: 10.1109/TLT.2017.2687458

CrossRef Full Text | Google Scholar

Klein, R. M., Hassan, T., Wilson, G., Ishigami, Y., and Mulle, J. (2017). The AttentionTrip: a game-like tool for measuring the networks of attention. J. Neurosci. Methods 289, 99–109. doi: 10.1016/j.jneumeth.2017.07.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Korn, O., Brenner, F., Börsig, J., Lalli, F., Mattmüller, M., and Müller, A. (2017). Defining Recrutainment: a model and a survey on the gamification of recruiting and human resources. In International Conference on Applied Human Factors and Ergonomics (pp. 37–49). Springer, Cham.

Google Scholar

Kosinski, M., Wang, Y., Lakkaraju, H., and Leskovec, J. (2016). Mining big data to extract patterns and predict real-life outcomes. Psychol. Methods 21, 493–506. doi: 10.1037/met0000105

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuncel, N. R., Ones, D. S., and Sackett, P. R. (2010). Individual differences as predictors of work, educational, and broad life outcomes. Personal. Individ. Differ. 49, 331–336. doi: 10.1016/j.paid.2010.03.042

CrossRef Full Text | Google Scholar

Lambiotte, R., and Kosinski, M. (2014). Tracking the digital footprints of personality. Proc. IEEE 102, 1934–1939. doi: 10.1109/JPROC.2014.2359054

CrossRef Full Text | Google Scholar

Landers, R. N., and Callan, R. C. (2011). “Casual social games as serious games: the psychology of gamification in undergraduate education and employee training” in Serious Games and Edutainment Applications. eds. M. Ma and A. Oikonomou and L. Jain (London: Springer), 399–423. doi: 10.1007/978-1-4471-2161-9_20

CrossRef Full Text | Google Scholar

LeBlanc, P. A. H. M. M., and Chawla, A. (2003). Cognitive ability testing and employment selection: does test content relate to adverse impact? Appl. HRM Res. 7, 41–48.

Google Scholar

Leutner, F., Codreanu, S. C., Liff, J., and Mondragon, N. (2020). The potential of game-and video-based assessments for social attributes: examples from practice. J. Manag. Psychol. 36, 533–547. doi: 10.1108/JMP-01-2020-0023

CrossRef Full Text | Google Scholar

Leutner, F., Liff, J., Zuloaga, L., and Mondragon, N. (2021). HireVue’s assessment science. HireVue. Available at: https://webapi.hirevue.com/wp-content/uploads/2021/03/HireVue-Assessment-Science-whitepaper-2021.pdf?_ga=2.239366848.21559469.1628169017-383464143.1628169017

Google Scholar

Lieberoth, A. (2015). Shallow gamification: testing psychological effects of framing an activity as a game. Games Cult. 10, 229–248. doi: 10.1177/1555412014559978

CrossRef Full Text | Google Scholar

Luft, C. D. B., Gomes, J. S., Priori, D., and Takase, E. (2013). Using online cognitive tasks to predict mathematics low school achievement. Comput. Educ. 67, 219–228. doi: 10.1016/j.compedu.2013.04.001

CrossRef Full Text | Google Scholar

Lumsden, J., Skinner, A., Woods, A. T., Lawrence, N. S., and Munafò, M. (2016). The effects of gamelike features and test location on cognitive test performance and participant enjoyment. PeerJ 4:e2184. doi: 10.7717/peerj.2184

PubMed Abstract | CrossRef Full Text | Google Scholar

Mavridis, A., and Tsiatsos, T. (2017). Game-based assessment: investigating the impact on test anxiety and exam performance. J. Comput. Assist. Learn. 33, 137–150. doi: 10.1111/jcal.12170

CrossRef Full Text | Google Scholar

McCarthy, J. M., and Goffin, R. D. (2005). Selection test anxiety: exploring tension and fear of failure across the sexes in simulated selection scenarios. Int. J. Sel. Assess. 13, 282–295. doi: 10.1111/j.1468-2389.2005.00325.x

CrossRef Full Text | Google Scholar

McGrew, K. S. (2009). CHC theory and the human cognitive abilities project: standing on the shoulders of the giants of psychometric intelligence research. Intelligence 37, 1–10. doi: 10.1016/j.intell.2008.08.004

CrossRef Full Text | Google Scholar

McPherson, J., and Burns, N. R. (2008). Assessing the validity of computer-game-like tests of processing speed and working memory. Behav. Res. Methods 40, 969–981. doi: 10.3758/BRM.40.4.969

PubMed Abstract | CrossRef Full Text | Google Scholar

Nacke, L. E., Nacke, A., and Lindley, C. A. (2009). Brain training for silver gamers: effects of age and game form on effectiveness, efficiency, self-assessment, and gameplay experience. Cyber Psychol. Behav. 12, 493–499. doi: 10.1089/cpb.2009.0013

PubMed Abstract | CrossRef Full Text | Google Scholar

Núñez-Peña, M. I., Suárez-Pellicioni, M., and Bono, R. (2016). Gender differences in test anxiety and their impact on higher education students’ academic achievement. Proc. Soc. Behav. Sci. 228, 154–160. doi: 10.1016/j.sbspro.2016.07.023

CrossRef Full Text | Google Scholar

Osborne, J. W. (2007). Linking stereotype threat and anxiety. Educ. Psychol. 27, 135–154. doi: 10.1080/01443410601069929

CrossRef Full Text | Google Scholar

Palmer, D., Lunceford, S., and Patton, A. J. (2012). The engagement economy: how gamification is reshaping businesses. Deloitte Rev. 11, 52–69.

Google Scholar

Park, G., Schwartz, H. A., Eichstaedt, J. C., Kern, M. L., Kosinski, M., Stillwell, D. J., et al. (2015). Automatic personality assessment through social media language. J. Pers. Soc. Psychol. 108, 934–952. doi: 10.1037/pspp0000020

PubMed Abstract | CrossRef Full Text | Google Scholar

Peters, H., Kyngdon, A., and Stillwell, D. (2021). Construction and validation of a game-based intelligence assessment in minecraft. Comput. Hum. Behav. 119:106701. doi: 10.1016/j.chb.2021.106701

CrossRef Full Text | Google Scholar

Peters, M., Laeng, B., Latham, K., Jackson, M., Zaiyouna, R., and Richardson, C. (1995). A Redrawn Vandenberg and Kuse Mental Rotations Test - Different Versions and Factors That Affect Performance. Brain and Cognition 28, 39–58. doi: 10.1006/brcg.1995.1032

PubMed Abstract | CrossRef Full Text | Google Scholar

Ployhart, R. E., and Holtz, B. C. (2008). The diversity–validity dilemma: strategies for reducing racioethnic and sex subgroup differences and adverse impact in selection. Pers. Psychol. 61, 153–172. doi: 10.1111/j.1744-6570.2008.00109.x

CrossRef Full Text | Google Scholar

Quiroga, M. Á., Escorial, S., Román, F. J., Morillo, D., Jarabo, A., Privado, J., et al. (2015). Can we reliably measure the general factor of intelligence (g) through commercial video games? Yes, we can! Intelligence 53, 1–7. doi: 10.1016/j.intell.2015.08.004

CrossRef Full Text | Google Scholar

Raghavan, M., Barocas, S., Kleinberg, J., and Levy, K. (2020). Mitigating bias in algorithmic hiring: evaluating claims and practices. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 469–481. New York, NY: ACM. doi: 10.1145/3351095.3372828

CrossRef Full Text | Google Scholar

Raven, J. (2000). The Raven’s progressive matrices: change and stability over culture and time. Cogn Psychol. 41, 1–48. doi: 10.1006/cogp.1999.0735

CrossRef Full Text | Google Scholar

Reeve, C. L., and Bonaccio, S. (2008). Does test anxiety induce measurement bias in cognitive ability tests? Intelligence 36, 526–538. doi: 10.1016/j.intell.2007.11.003

CrossRef Full Text | Google Scholar

Richards, A., French, C. C., Keogh, E., and Carter, C. (2000). Test-anxiety, inferential reasoning and working memory load. Anxiety Stress Coping 13, 87–109. doi: 10.1080/10615800008248335

CrossRef Full Text | Google Scholar

Roth, P. L., Bobko, P., Van Iddekinge, C. H., and Thatcher, J. B. (2013). Social media in employee-selection-related decisions: a research agenda for uncharted territory. J. Manag. 42, 269–298. doi: 10.1177/0149206313503018

CrossRef Full Text | Google Scholar

Ryan, A. M., Ployhart, R. E., and Friedel, L. A. (1998). Using personality testing to reduce adverse impact: a cautionary note. J. Appl. Psychol. 83, 298–307. doi: 10.1037/0021-9010.83.2.298

CrossRef Full Text | Google Scholar

Schmidt, F. L., and Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology: practical and theoretical implications of 85 years of research findings. Psychol. Bull. 124, 262–274. doi: 10.1037/0033-2909.124.2.262

CrossRef Full Text | Google Scholar

Schmidt, F. L., and Hunter, J. (2004). General mental ability in the world of work: occupational attainment and job performance. J. Pers. Soc. Psychol. 86, 162–173. doi: 10.1037/0022-3514.86.1.162

PubMed Abstract | CrossRef Full Text | Google Scholar

Schwartz, H. A., Eichstaedt, J. C., Kern, M. L., Dziurzynski, L., Ramones, S. M., Agrawal, M., et al. (2013). Personality, gender, and age in the language of social media: the open-vocabulary approach. PLoS One 8:e73791. doi: 10.1371/journal.pone.0073791

PubMed Abstract | CrossRef Full Text | Google Scholar

Shen, C., Wang, H., and Ritterfeld, U. (2009). “Serious Games and Seriously Fun Games: Can They be One and the Same?” in Serious Games. Routledge, 1st edn. 70–84.

Google Scholar

Smither, J. W., Reilly, R. R., Millsap, R. E., Kenneth Pearlman AT&T, and Stoffey, R. W. (1993). Applicant reactions to selection procedures. Pers. Psychol. 46, 49–76. doi: 10.1111/j.1744-6570.1993.tb00867.x

CrossRef Full Text | Google Scholar

Society for Industrial and Organizational Psychology, Inc (2003). Principles for the Validation and Use of Personnel Selection Procedures, 4th Edn. College Park, MD: SIOP.

Google Scholar

Spielberger, C. D., Gonzalez, H. P., Taylor, C. J., Anton, E. D., Algaze, B., Ross, G. R., et al. (1980). Manual for the Test Anxiety Inventory (“Test Attitude Inventory”). Redwood City, CA: Consulting Psychologists Press.

Google Scholar

Szafranski, D. D., Barrera, T. L., and Norton, P. J. (2012). Test anxiety inventory: 30 years later. Anxiety Stress Coping 25, 667–677. doi: 10.1080/10615806.2012.663490

PubMed Abstract | CrossRef Full Text | Google Scholar

Taylor, J., and Deane, F. P. (2002). Development of a short form of the test anxiety inventory (TAI). J. Gen. Psychol. 129, 127–136. doi: 10.1080/00221300209603133

PubMed Abstract | CrossRef Full Text | Google Scholar

Team PredictiveHire (2021). Known biases can be removed with the right data and testing. Predictive Hire. Available at: https://www.predictivehire.com/blog/testing-bias-reducing-bias/

Google Scholar

Thames, A. D., Panos, S. E., Arentoft, A., Byrd, D. A., Hinkin, C. H., and Arbid, N. (2015). Mild test anxiety influences neurocognitive performance among African Americans and European Americans: identifying interfering and facilitating sources. Cult. Divers. Ethn. Minor. Psychol. 21, 105–113. doi: 10.1037/a0037530

PubMed Abstract | CrossRef Full Text | Google Scholar

The International Cognitive Ability Resource Team. (2014). The international cognitive ability resource. Available at: https://icar-project.com/

Google Scholar

Tippins, N. T. (2015). Technology and assessment in selection. Annu. Rev. Organ. Psychol. Organ. Behav. 2, 551–582. doi: 10.1146/annurev-orgpsych-031413-091317

CrossRef Full Text | Google Scholar

Toplak, M. E., West, R. F., and Stanovich, K. E. (2014). Assessing miserly information processing: an expansion of the cognitive reflection test. Think. Reason. 20, 147–168. doi: 10.1080/13546783.2013.844729

CrossRef Full Text | Google Scholar

Tremblay, J., Bouchard, B., and Bouzouane, A. (2010). Adaptive game mechanics for learning purposes-making serious games playable and fun. In CSEDU 2, 465–470. doi: 10.5220/0002855604650470

CrossRef Full Text | Google Scholar

U.S. Equal Employment Opportunity Commission. (2021) Prohibited employment policies/practices. Available at: https://www.eeoc.gov/prohibited-employment-policiespractices

Google Scholar

Uniform Guidelines on Employee Selection Procedures (1978). Federal Register, vol. 43, 38290–3831.

Google Scholar

Verhaegh, J., Fontijn, W. F. J., Aarts, E. H. L., and Resing, W. C. M. (2013). In-game assessment and training of nonverbal cognitive skills using TagTiles. Pers. Ubiquit. Comput. 17, 1637–1646. doi: 10.1007/s00779-012-0527-0

CrossRef Full Text | Google Scholar

von der Embse, N., Jester, D., Roy, D., and Post, J. (2018). Test anxiety effects, predictors, and correlates: a 30-year meta-analytic review. J. Affect. Disord. 227, 483–493. doi: 10.1016/j.jad.2017.11.048

PubMed Abstract | CrossRef Full Text | Google Scholar

Weibel, D., Wissmath, B., Habegger, S., Steiner, Y., and Groner, R. (2008). Playing online games against computer-vs. human-controlled opponents: effects on presence, flow, and enjoyment. Comput. Hum. Behav. 24, 2274–2291. doi: 10.1016/j.chb.2007.11.002

CrossRef Full Text | Google Scholar

Yam, J., and Skorburg, J. A. (2021). From human resources to human rights: impact assessments for hiring algorithms. Ethics Inf. Technol. 23, 611–623. doi: 10.1007/s10676-021-09599-7

CrossRef Full Text | Google Scholar

Zeidner, M. (1998). Test Anxiety: The State of the Art. New York: Plenum.

Google Scholar

Keywords: game based assessment, cognitive ability, adverse impact, user experience, outcome parity, fairness, psychometric tests, recruitment

Citation: Leutner F, Codreanu S-C, Brink S and Bitsakis T (2023) Game based assessments of cognitive ability in recruitment: Validity, fairness and test-taking experience. Front. Psychol. 13:942662. doi: 10.3389/fpsyg.2022.942662

Received: 12 May 2022; Accepted: 20 December 2022;
Published: 18 January 2023.

Edited by:

Luke Treglown, University College London, United Kingdom

Reviewed by:

Oliver Korn, Offenburg University of Applied Sciences, Germany
Lynsay Shepherd, Abertay University, United Kingdom

Copyright © 2023 Leutner, Codreanu, Brink and Bitsakis. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Franziska Leutner, yes f.leutner@gold.ac.uk

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.