- 1Department of English, Indiana University of Pennsylvania,Indiana, PA, United States
- 2Department of English, Duke Kunshan University, Kunshan, Jiangsu, China
- 3Department of Ecology and Evolutionary Biology, STEM Program Evaluation and Research Lab (STEM-PERL), Yale University, New Haven, CT, United States
- 4Department of Biology, Montclair State University, Montclair, NJ, United States
- 5Unidad de Investigaciones en Salud Pública, Facultad de Ciencias Químico Biológicas, Universidad Autónoma de Sinaloa, Ciudad Universitaria, Culiacán, SI, Mexico
- 6Department of Biology, Illinois Wesleyan University, Bloomington, IL, United States
- 7Department of Biological Sciences, University of North Texas, Denton, TX, United States
- 8Department of Biology, Howard University, Washington, DC, United States
- 9Department of Biological Sciences, University of Southern California, Los Angeles, CA, United States
- 10Department of Biological Sciences, The University of North Carolina at Charlotte, Charlotte, NC, United States
- 11Department of Biological Sciences, University of Pittsburgh, Pittsburgh, PA, United States
- 12Biology Department, Madison Area Technical College, Madison, WI, United States
- 13Department of Biology, Virginia State University, Petersburg, VA, United States
- 14Department of Biology, Doane University, Crete, NE, United States
- 15Department of Biology, Truckee Meadows Community College, Reno, NV, United States
- 16Department of Natural Sciences and Mathematics, Webster University, St. Louis, MO, United States
- 17Department of Biology and Biotechnology, Worcester Polytechnic Institute, Worcester, MA, United States
- 18Department of Biological Sciences, University of North Carolina at Charlotte, Charlotte, NC, United States
- 19Department of Biology, College of Charleston, Charleston, SC, United States
- 20Department of Biological Sciences, University of Maryland, Baltimore, MD, United States
- 21School of Biological Sciences, University of Nebraska-Lincoln, Lincoln, NE, United States
- 22School of Life Sciences, University of Hawaii at Manoa, Honolulu, HI, United States
- 23Department of Biology, University of West Florida, Pensacola, FL, United States
- 24Agricultural and Biological Engineering Department, Purdue University, West Lafayette, IN, United States
- 25Department of Biology, Wartburg College, Waverly, IA, United States
- 26Department of Natural Sciences, Mitchell Community College, Statesville, NC, United States
- 27Department of Biology, University of Detroit Mercy, Detroit, MI, United States
- 28Department of Biomedical Sciences, Rocky Vista University—Montana College of Osteopathic Medicine, Billings, MT, United States
- 29School of Natural Sciences, Indiana University Southeast, New Albany, IN, United States
- 30Science Department, Dominican University New York, Orangeburg, NY, United States
- 31Department of Biology, Queens University of Charlotte, Charlotte, NC, United States
- 32Department of Biological Sciences, Nova Southeastern University, Fort Lauderdale, FL, United States
- 33Department of Biology, Indian River State College, Fort Pierce, FL, United States
- 34Department of Biology, Marist College, Poughkeepsie, NY, United States
- 35Biology Department, New Mexico Institute of Mining and Technology, Socorro, NM, United States
- 36Departamento de Microbiología Básica, Facultad de Ciencias Bioquímicas y Farmacéuticas-UNR, Rosario, Santa Fe, Argentina
- 37Department of Natural Sciences, Saint Leo University, St. Leo, FL, United States
- 38Department of Biology, Southern Connecticut State University, New Haven, CT, United States
- 39Department of Biological Sciences, Tarleton State University, Stephenville, TX, United States
- 40Department of Microbiology, University of Lagos, Lagos, Nigeria
- 41Department of Biology, Columbia State Community College, Columbia, TN, United States
- 42Department of Science, Neumann University, Aston Township, PA, United States
- 43Molecular, Cellular, and Developmental Biology, Universtiy of Colorado, Boulder, Boulder, CO, United States
- 44Department of Biology—School of Sciences, University of Louisiana at Monroe, Monroe, LA, United States
- 45Department of Biology, Johns Hopkins University, Baltimore, MD, United States
- 46Biology Department, Oregon Coast Community College, Newport, OR, United States
- 47Department of Science, Durham Tech, Durham, NC, United States
- 48Department of Microbiology, Immunology, and Molecular Genetics, University of California, Los Angeles, Los Angeles, CA, United States
- 49Biology Department, Winthrop University, Rock Hill, SC, United States
- 50Department of Chemistry and Physics, Western Carolina University, Cullowhee, NC, United States
- 51Department of Biology, Colorado State University-Pueblo, Pueblo, CO, United States
- 52Biology Department, South Texas College, McAllen, TX, United States
- 53Department of Biology, Tufts University, Medford, MA, United States
- 54Department of Biological and Chemical Sciences, New York Institute of Technology, Old Westbury, NY, United States
- 55Department of Biological Sciences and Geology, Queensbrough Community College, Queens, NY, United States
- 56Department of Biological Sciences, Idaho State University, Pocatello, ID, United States
- 57School of Biology, University of St Andrews, Biomedical Sciences Research Complex, St Andrews, United Kingdom
- 58Department of Biology, Gonzaga University, Spokane, WA, United States
- 59Science Division, Allegany College of Maryland, Cumberland, MD, United States
- 60School of Molecular Biosciences, Washington State University, Pullman, WA, United States
- 61Basic and Clinical Sciences, Albany College of Pharmacy and Health Sciences, Albany, NY, United States
- 62Department of Life Sciences, University of New Hampshire Manchester, Manchester, NH, United States
- 63Center for Biological Data Science, Virginia Commonwealth University, Richmond, VA, United States
- 64Department of Biology, Bryn Mawr College, Bryn Mawr, PA, United States
- 65Department of Biological Sciences, Lehigh University, Bethlehem, PA, United States
- 66Math and Natural Sciences, Collin College, Plano, TX, United States
- 67Biology Department, University of Wisconsin-River Falls, River Falls, WI, United States
- 68Department of Biology, Saint Joseph's University, Philadelphia, PA, United States
- 69Department of Biological Sciences, University of Mary Washington, Fredericksburg, VA, United States
- 70Department of Biology, Virginia Western Community College, Roanoke, VA, United States
- 71Life Sciences Department/STEM Division, Hartford Community College, Bel Air, MD, United States
- 72Department of Biology, Austin Peay State University, Clarksville, TN, United States
- 73Department of Science, Cabrini University, Radnor, PA, United States
- 74Integrated Science, Business and Technology, La Salle University, Philadelphia, PA, United States
- 75Division of Natural Sciences, Mount Saint Mary College, Newburgh, NY, United States
- 76Department of Biology, University of Evansville, Evansville, IN, United States
- 77Department of Biological Sciences, Florida Gulf Coast University, Fort Myers, FL, United States
- 78Department of Science and Mathematics, Northern State University, Aberdeen, SD, United States
- 79Molecular and Biomedical Sciences and Honors College, University of Maine, Orono, ME, United States
- 80Hicks Honors College, University of North Florida, Jacksonville, FL, United States
- 81Departamento de Microbiología Básica, Facultad de Ciencias Bioquímicas y Farmacéuticas-UNR, Rosario, Santa Fe, Argentina
- 82Natural Sciences, College of Coastal Georgia, One College Drive, Brunswick, GA, United States
- 83Biological Sciences, SUNY Old Westbury, Old Westbury, NY, United States
- 84Department of Biology, Thiel College, Greenville, PA, United States
- 85College of Health and Natural Sciences, Franklin Pierce University, Rindge, NH, United States
- 86STEM Academy, Salish Kootenai College, Pablo, MT, United States
- 87Department of Biology, Radford University, Radford, VA, United States
- 88Department of Biological Sciences, Benedictine University, Lisle, IL, United States
- 89Department of Plant Pathology, Entomology, and Microbiology, Iowa State University, Ames, IA, United States
- 90Department of Biological Sciences and Life Sciences, Bowling Green State University, Bowling Green, OH, United States
- 91Department of Molecular Biosciences, University of South Floria, Tampa, FL, United States
- 92Department of Biological Sciences, Webster University, St. Louis, MO, United States
- 93Department of Biology, Frostburg State University, Frostburg, MD, United States
- 94Institute for Advanced Medical Research and Training, College of Medicine, University of Ibadan, UCH Ibadan, Ibadan, Oyo, Nigeria
- 95Department of Biology, Ouachita Baptist Univeristy, Arkadelphia, AR, United States
- 96Department of Biology, Western Kentucky University, Bowling Green, KY, United States
- 97Department of Biology, University of Massachusetts, Amherst, MA, United States
- 98CEA, CNRS, Institute for Integrative Biology of the Cell (I2BC), Université Paris-Saclay, Gif-sur-Yvette, France
- 99Ottawa Institute of Systems Biology and Department of Biochemistry, Michobiology, and Immunulogy, University of Ottawa, Ottawa, KS, United States
- 100Biology Department, SUNY Plattsburgh, Plattsburgh, NY, United States
- 101Department of Natural Sciences, University of Houston-Downtown, Houston, TX, United States
- 102Department of Biology, Washington University in St. Louis, St. Louis, MO, United States
- 103Department of Science, Penn State Schuylkill, Schuylkill Haven, PA, United States
- 104Department of Biology, Salem State University, Salem, MA, United States
- 105School of Life Sciences, University of Nevada, Las Vegas, Las Vegas, NV, United States
- 106Department of Biological Sciences, Southern Maine Community College, South Portland, ME, United States
- 107School of Integrated Sciences, James Madison University, Harrisonburg, VA, United States
- 108Biology Department, College of St. Scholastica, Duluth, MN, United States
- 109Department of Biology, Carthage College, Kenosha, WI, United States
- 110Department of Biology, Northwestern College, Orange City, IA, United States
- 111Biology Department, University of Puerto Rico at Cayey, Cayey, Puerto Rico
- 112Department of Biology, Case Western Reserve University, Cleveland, OH, United States
- 113Science and Math, Durham Technical Community College, Durham, NC, United States
- 114Department of Biology and Chemistry, Alliance University, New York, NY, United States
- 115Biology Department, The College of St. Scholastica, Duluth, MN, United States
- 116Department of Biochemistry, Microbiology and Immunology, University of Otttawa, Ottawa, KS, United States
- 117Department of Natural Sciences, CCSNH-Concord's Community College, Concord, NH, United States
- 118Translational and Molecular Medicine, University of Ottawa, Ottawa, KS, United States
- 119Center for the Advancement of Science Leadership and Culture, Howard Hughes Medical Institute, Chevy Chase, MD, United States
- 120Department of Biology, James Madison University, Harrisonburg, VA, United States
Course-based research pedagogy involves positioning students as contributors to authentic research projects as part of an engaging educational experience that promotes their learning and persistence in science. To develop a model for assessing and grading students engaged in this type of learning experience, the assessment aims and practices of a community of experienced course-based research instructors were collected and analyzed. This approach defines four aims of course-based research assessment—(1) Assessing Laboratory Work and Scientific Thinking; (2) Evaluating Mastery of Concepts, Quantitative Thinking and Skills; (3) Appraising Forms of Scientific Communication; and (4) Metacognition of Learning—along with a set of practices for each aim. These aims and practices of assessment were then integrated with previously developed models of course-based research instruction to reveal an assessment program in which instructors provide extensive feedback to support productive student engagement in research while grading those aspects of research that are necessary for the student to succeed. Assessment conducted in this way delicately balances the need to facilitate students’ ongoing research with the requirement of a final grade without undercutting the important aims of a CRE education.
Introduction
Recent educational initiatives in STEM are facilitating wide-spread implementation of course-based research experiences (CRE) because they increase persistence for students across many demographics (Russell et al., 2007; Jordan et al., 2014; Hanauer et al., 2017; Hernandez et al., 2018). This educational approach is characterized by having students involved in conducting and contributing to authentic scientific research projects (Hanauer et al., 2006, 2012, 2016, 2017; President’s Council of Advisors on Science and Technology, 2012; Graham et al., 2013; Auchincloss et al., 2014; Hanauer and Dolan, 2014; Hernandez et al., 2018). Recent research on the pedagogical approach to teaching a CRE describes how this educational design transitions the ways in which instructors teach and the way in which the relationship between the instructor and the student is conceptualized and manifest (Hanauer et al., 2022). In particular, the hierarchy which is so prevalent in most educational settings is flattened slightly with the instructor and student working together on a shared research project (Hanauer et al., 2022). The expertise of the instructor is utilized in supporting a research process, the outcomes of which are not necessarily known (Auchincloss et al., 2014). For both instructor and student, the research is on-going and to a degree unpredictable. Timing for various outcomes may vary across students and projects, the type of interaction and expertise that the instructor has to provide may change and broadly the instructor and student need to be flexible in the ways in which they interact around the emerging scientific work. Hanauer et al. (2022) describe in detail the nature of this pedagogy and the ways in which instructors work with students in teaching a CRE.
While the pedagogical implementation of a CRE transitions the relations between instructor and student, the institutional requirement for a grade has not changed. Classroom grading is a significant and ubiquitous practice in STEM education in general and is a requirement whether the class is a CRE or not. The specific nature of a CRE raises several problems in relation to classroom grading. How does a teacher maintain the process of “shared” scientific research that is important beyond the classroom, if the instructor is “grading” the student on in-class tasks? When the nature of a class is not dictated by delimited content knowledge or a prescribed set of skills, what are the aims of assessment within a CRE? How does an instructor support and encourage a student during the challenges and potential failures of authentic science, if both student and instructor know that they need to assign a grade for the work being conducted? Broadly the problem of assessing and grading students in a CRE is that the CRE aims to provide a professional, authentic research experience in which the student feels that they are scientists. Grading seems quite artificial in this particular educational design.
Prior approaches to assessing a student’s scientific inquiry divide into two camps: analytic schemes and authentic task modeling. Early work used an analytic scheme to define the components of scientific inquiry and suggested methods for assessing each of the parts in isolation. For example, Zachos (2004) delineates the core capabilities of scientific inquiry to include coordinating theories, searching for underlying principles, being concerned with precision, identifying sources of error in measurement and proportional reasoning, and suggest these should be used in the design of a series of performance tasks. Wenning (2007) designed a multiple-choice test of the components of a scientific inquiry such as identifying a problem, formulating a hypothesis, generating a prediction, designing an experiment, collecting and organizing data, using statistical methods, and explaining results. Shavelson et al. (1998) proposed using a range of performance tasks to evaluate scientific inquiry abilities of students. In line with this analytic approach, Pelaez et al. (2017) specified a set of core experimentation competencies consisting of the categories—identify, question, plan, conduct, analyze, conclude, and communicate. Zelaya et al. (2022) categorize 14 survey style instruments and 16 evaluation rubrics in relation to this set of competencies specifying the degree of overlap between each tool and the specified competencies. Similarly, in an extensive review of the existing tools that can be used for the assessment of a CURE, Shortlidge and Brownell (2016) review 26 survey style tools that can be used to assess different aspects of the research experience such as critical thinking, views of science, project ownership, biological concepts, and experimental design. What many these approaches have in common is the idea that the grading of scientific inquiry can be externalized from the actual research that the student is doing; students are evaluated for a set of skills, competencies, dispositions, and abilities for future scientific research.
The second camp proposed modeling authentic activity. In principle, if a CRE involves authentic research which produces scientific findings useful for a scientific community and the student is seen as a researcher, it would be logical that the evaluation of the student’s work would be situated in the ways professional scientists are evaluated. However, practically, waiting for a paper to be published or a poster presented at a professional conference would be problematic both in relation to timing and the threshold level for successful student outcomes. Instead, Hanauer et al. (2009) proposed an approach termed Active Assessment which analyzes the professional research practices of a specific research project and then uses these as a way of generating a rubric for evaluating student work. Assessment is done on the student as they work through the scientific inquiry they are involved in. A similar approach has been proposed by Dolan and Weaver (2021). What characterizes this approach are the ideas that assessment and grading should be situated in the performance of a student while conducting research in the CRE and that this assessment should be based on professional performance.
However, while this second approach offers a conceptual basis of how assessment in a CRE could be conducted, it is not based on data from actual instructors teaching a CRE. The aim of this study is to look at how experienced instructors in a large-scale CRE program—the Science Education Alliance (SEA) program by the Howard Hughes Medical Institute (HHMI)—describe their processes of assessing their students engaged in course-based research. Working with this large community of experienced CRE instructors over a 2-year period, models of CRE assessment were developed. In addition, this current paper builds upon prior research on models of CRE instruction, which were similarly developed with this community of SEA instructors (Hanauer et al., 2022). The outcome of this study thus provides insight into how CREs can be assessed and graded while maintaining the pedagogical approach designed to provide an authentic research experience for students and enhance persistence.
Issues with assessment and grading
In a classic text, Walvoord and Anderson (1998) specify a series of basic roles that grading is expected to perform: (1) It should be a reliable measure of a student’s performance of required work; (2) It should be a means of communicating the quality of the student’s performance with parents, other faculty, the university, future institutions, and places of work; (3) It should be a source of motivation; (4) It should provide meaningful information for feedback to students and instructors to enhance learning; and (5) It can be a way of organizing class work. However, as seen in the scholarship, the implementation of grading is not unproblematic.
As documented over decades, there are questions as to whether grading always fulfills the stated aims above (Jaschik, 2009). Prior research has suggested that STEM faculty have the knowledge to create assessment tasks but often lack an understanding of how to validate these tasks (Hanauer and Bauerle, 2015). Some faculty problematically assume that the way they were graded is a basis for the grading of their own students leading to a persistence of outdated assessment practices (Boothroyd and McMorris, 1992). When considering what to assess and grade, there can be confusion between learning components tied to stated learning objectives of the course and other aspects of being a student such as punctuality, attendance, and participation (Hu, 2005). Additionally, there is little agreement between instructors as to which components should go into a grade with different instructors varying greatly in relation to how assessment is conducted (Cizek et al., 1996). Research has also shown that grades can vary in relation to variables such as instructors, departments, disciplines, and institutions (Lipnevich et al., 2020) and in relation to specific student characteristics such as physical attractiveness (Baron and Byrne, 2004) and ethnicity (Fajardo, 1985).
It is important to understand the central role grading plays in the lives of students. Grading can increase anxiety, fear, and lack of interest and hinder the ability to perform on subsequent tasks (Butler, 1988; Crooks, 1988; Pulfrey et al., 2011). There are alarming rates of attrition from STEM documented for students who identify as African American or Black, Latino, or Hispanic, and American Indian and Alaska Native (National Science Board, 2018; Asai, 2020; Whitcomb and Chandralekha, 2021) and low grades is one of the factors that leads to this outcome (Whitcomb and Chandralekha, 2021). The relationship between grading and persistence is situated in the effect of negative feedback on performance (such as a lower-than-expected grade) and the individual’s sense of self-efficacy in that field (Bandura, 1991, 2005). Students who identify as African American or Black, Latino or Hispanic, and American Indian and Alaska Native may enter the STEM fields with pre-existing fears and anxieties about their work resulting from stereotype threat (Hilts et al., 2018). Negative experiences with grading further exacerbate these feelings leading to a disbelief in their ability to continue in STEM and hence attrition from that course of study (Hilts et al., 2018; Whitcomb and Chandralekha, 2021). Recent research has shown that grading works in two parallel ways: lower grades limit the opportunities that are available to students and increase the negative psychological impact on students’ intent to persist in STEM (Hatfield et al., 2022). As such grading, if not conducted appropriately, could directly undermine the main aim of a CRE—increased persistence in STEM for all students.
Methodology
Overview
A multi-method, large-scale and multi-year research methodology was employed in this study. Data collection and analysis was conducted over a 2-year period in a series of designed stages with full participation from a large group of CRE instructors and a dedicated science education research team. The project developed in the following stages:
1. Survey: The initial stage of the study involved a qualitative and quantitative survey. The qualitative section asked about grading and assessment procedures used by instructors in their CRE courses and asked for a detailed explanation of the way these were used in their courses. The quantitative section used the psychometrically validated scales of the Faculty Self-Reported Assessment survey (Hanauer and Bauerle, 2015) to evaluate the knowledge level of the surveyed faculty. The aim of this first stage of the project was to collect descriptive data on the participants’ understanding of assessment and specific information on the way they conduct assessment and grading in their courses.
2. Analysis and large-scale community checking of assessment aims and practices: Data from the qualitative study were analyzed using a systematic content analysis process, and the quantitative data were analyzed using standard statistical procedures. The quantitative data was analyzed in terms of high-level assessment aims and specific grading and assessment practices. All analyses were summarized and then presented in a workshop setting to a cohort of 106 CRE instructors. In a small-focus group format, the aims and practices were presented and instructors provided written feedback on the validity of the analysis, the specification of the high-level aims, the specification of practices, and the assignment of the practices to assessment. Instructors responded within the workshop and were subsequently given an additional week to provide online responses to the questions posed. All data were collected using an online survey tool.
3. Analysis and community checking of models of assessment and grading: Data from the first stage of community checking were analyzed for modifications to the assessment aims and the assigned assessment and grading practices. Percentage of agreement with the aims and practices was calculated and modifications to the models were assigned. During this analysis there were no changes to the high-level aims, but several specific practices were added. Once the table of aims and practices had been finalized, the original survey commentary dealing with how assessment and grading were conducted was consulted. Using this commentary and the pedagogical models of CRE instruction (Hanauer et al., 2022), the aims and practices of assessment were integrated with the discussion of CRE instruction. Three integrated models were developed and presented to a dedicated group of 23 instructors for validation process. Instructors were asked to provide feedback on the quality and descriptive validity of the models, the specification of aims of assessment and the specific practices. Instructors provided feedback during the workshop and for a week after the workshop. All data were collected using an online survey tool.
4. Finalization of the models: Feedback from the workshop was analyzed for verification of the models and any required modifications that might be needed. Agreement with the models and their components were checked. Following this process, the models were finalized.
Participants
Participants for this study were elicited from the full set of instructors who teach in the SEA program. The SEA program is a large-scale, two semesters, program implemented at 190 institutions predominantly with Freshman and Sophomore students. This course is supported by the Howard Hughes Medical Institute and has scientific support from the Hatfull laboratory at the University of Pittsburgh. For the first stage of data collection, a survey request was sent to 330 SEA instructors. 105 faculty responded with 72 instructors providing full answers on the survey. Table 1 presents the instructor demographics. The SEA faculty respondents are predominantly White (≥58.1%) and women (≥49.5%). A range of academic ranks from instructor to full professor were represented in the sample. As seen in Table 1, the majority of respondents had at least 3 years of teaching in the program and above 6+ years of teaching postsecondary science. Respondents for the community checking of the model were drawn from the SEA faculty. For each stage, 100+ instructors participated. Demographic data were not collected on the participants at the two community checking sessions. As a community of CRE instructors, during the semester, the SEA has a weekly 1-h, Friday afternoon session providing scientific and educational instructor development. During the Fall 2022 semester, two sessions were conducted by the Lead Assessment Coordinator of the SEA (Dr. Hanauer) dedicated to the development of a meaningful assessment approach. The sessions involved a lecture approach of general principles of assessment including constructive alignment between objectives and instruments, active assessment instruments that could be used and ways of interpreting outcomes. Participation in these Friday sessions were voluntary. Approximately 50 faculty attended these two sessions.
Instruments
As described in the overview of the research process, data collection consisted of a qualitative and quantitative initial survey, followed by a large community checking survey and a final assessment model checking survey. A specific tool was developed for each of these stages. The original survey consisted of three sections:
1. Familiarity with assessment terms: The first set of items were from the psychometrically validated Faculty Self-Reported Assessment survey (Hanauer and Bauerle, 2015). The survey consists of 24 established terms relating to assessment, organized into two components—assessment program and instrument knowledge, and knowledge of assessment validation procedures. On a five-point scale of familiarity (1 = I have never heard this term before; 5 = I am completely familiar with this term and know what it means), faculty rated each of the terms in relation to their familiarity with the term. The FRAS is used to evaluate levels of experience and exposure of faculty to assessment instruments and procedures. See Table 2, for a full list of the assessment terms used.
2. Qualitative reporting of student assessment: The second set of items were qualitative and required the instructor to describe the way in which they assess students in the SEA program, to specify the types of assessment used (such as quiz, rubric… etc.), and to explain what each assessment is used for. Following the first question, faculty were asked to describe how they grade students and what goes into the final grade. Answers consisted of written responses.
3. Self-efficacy assessment scales: The third set of items consisted self-reported measures of confidence in completing different aspects of assessment. The 12 items were taken from the FRAS (Hanauer and Bauerle, 2015) and consisted of a set of statements about the ability to perform different aspects of the assessment process (see Table 3 for a full list of the statement). All statements were rated on an agreement scale (1 = Strongly Disagree, 5 = Strongly Agree).
In order to collect verbal responses during the community checking stage of this project, participants completed an online survey that was presented following a shared online session in which the analyses of the main aims of assessment and the associated practices were presented (see Table 3). The survey asked for a written response to the following questions relating to each of the specified aims and associated practices:
1. Does this assessment aim make sense to you? Please specify if you agree or disagree that this is an aim of your CRE assessment.
2. For this aim, do the practices listed above make sense to you? Please comment on any that do not.
3. For this aim, are there practices of assessment that are not listed? If so, please list these additional practices and describe what these practices are used to evaluate.
4. Are there aims of assessment beyond the 4 that are listed above? If so, please describe any additional aims of assessment below.
The final community checking procedure involved the presentation of the full models of assessment to the collected participants in a shared online session (see Figures 1–3). Following the presentation of the models, the participants were divided into groups and each group was assigned a model to discuss and respond to. Each model was reviewed by two groups, and all responses were collected using an online written survey with the following questions:
1. For each of the instructional models, have the appropriate assessment aims been specified?
2. For each of the instruction models, have the appropriate assessment practices been specified?
3. Overall, do the models present an accurate and useful description of grading practices in the SEA?
4. Please suggest any modifications and comments you have on the model.
Figure 1. The core components of a CRE assessment model: based on the qualitative analysis of faculty descriptions of their assessment and grading practices in a CRE, four central aims of assessment were defined: (1) Assess Laboratory Work and Scientific Thinking; (2) Evaluate Mastery of Concepts, Quantitative Thinking, and Skills; (3) Appraise Forms of Scientific Communication; and (4) Metacognition of Learning. Together these four aims and associated assessment and grading practices define the assessment program of a CRE.
Figure 2. Assessing being a scientist and generating data: this model has three distinct stages. The first stage relates to the assessment of implicit instruction and protocol training. The second stage relates to aspects of doing science in the laboratory and the final stage relates to scientific outputs. The model presents the aims and practices of assessment applied at each of these stages.
Figure 3. Assessing procedural knowledge: this model has three distinct stages. The first stage relates to content information. The second stage relates to protocol training and training a student to think like a scientist. The third stage relates to scientific outputs. The model presents the aims and practices of assessment applied at each of these stages.
Procedures
Data were collected in three stages. The initial stage consisted of an online survey that was distributed to all faculty of the SEA using the web-based platform Qualtrics. Following the informed consent, process responses to the qualitative and quantitative items were recorded. The second stage involved the collection of community checking data from SEA instructors. A dedicated online Zoom session was arranged for this during one of the monthly virtual faculty meetings organized through the SEA program. During a 1-h session, the analysis of the aims of assessment and the associated practices were presented to the faculty. In small groups (breakout rooms), each of the aims and its associated practices were discussed. Following the session, an online survey was sent to faculty to collect their level of agreement with the aims and practices that were presented. They were also asked to modify or add any aims or practices that had been missed in the presented analysis of the original survey. The third stage of community checking data analysis consisted of a second online session during the regular end-of-week faculty meeting. During a 1-h session, each of the assessment models was presented to the faculty who then discussed them in small groups (breakout rooms). A survey was sent to the faculty during the session to respond to the models and write their responses to the models. All data were collected in accordance with the guidelines of Indiana University of Pennsylvania IRB #21-214.
Analysis
The analysis of the data in this study was conducted in four related stages. The initial survey had both quantitative and qualitative data. The quantitative data was analyzed using established statistical descriptive methods. The qualitative verbal data consisted of a series of written statements relating to the practices used for assessment by the different instructors and the aims of using these practices. Using an emergent content analysis approach, each of the instructor statements was analyzed and coded. Two different initial code books were developed. One dealt with the list of practices used by the faculty; the second involved the explanation of why these practices were used and what the instructor was trying to assess. The data were coded by two trained applied linguistic researchers and following several iterations, a high level of agreement was reached on the practices and aims specified by the instructors. The second stage of this analysis of the verbal survey data consisted of combining the aims and practices codes. The specified practices across all of the instructors for each of the aims was tabulated. A frequency count of the number of faculty who specified each of the practices was conducted. The outcome of the first stage of analysis was a statistical description of the levels of knowledge and confidence of faculty on assessment issues and the specification of four main aims of assessment with associated assessment practices.
The second stage of analysis followed the presentation of the tabulated coded data from the original survey to participants. In this stage of community checking, faculty specified agreement (or disagreement) with the assessment aims and the set of associated practices. The verbal responses were analyzed by two applied linguistics researchers and modifications were made to the tabulated data. The degree of agreement with each of the aims and associated practices was counted. Any additional practices specified by faculty were added to the model. No new aims were specified and as such no changes were made. The table of assessment aims and practices was finalized.
Having established the aims of assessment and related practices, a third stage of analysis involved integrating the emergent assessment aims and practices with models of CRE instruction which had been previously defined for the SEA instructors (see Hanauer et al., 2022 for full details). A team of two researchers worked together to specify the points of interaction between the instructional and assessment components of CRE teaching. Using the qualitative data of the original models and the verbal statements of aims for the assessment data, integrated models of assessment were developed. Following several iterations, three assessment models corresponding to the instructional models were specified.
The final stage of analysis followed the presentation of the models of assessment to the community of SEA faculty. A team of two researchers went over the changes presented by faculty in relation to each of the models. Changes that were specified, such as the addition of specific practices into different models, were made. The outcome of this process was a series of three models that capture the aims and practices of assessment.
Results
Instructor familiarity and self-efficacy with assessment
To build models of CRE assessment based on qualitative reports from instructors in the SEA program, we first evaluated instructors’ knowledge of assessment terms and their confidence in implementing assessment tasks. For instructor knowledge of assessment, we utilized the Faculty Self-Reported Assessment Survey (FRAS; Hanauer and Bauerle, 2015)—a tool which measures two components of assessment knowledge: (1) knowledge of assessment programs and instruments and (2) knowledge of assessment validation. Internal consistency was calculated for the each of the FRAS components. Cronbach’s Alpha was 0.86 for the Knowledge of assessment programs and instruments components and 0.94 for the knowledge of assessment validation component. These levels suggest that each of the components is sufficiently consistent and hence reliable.
For the Program and Instrument component, instructors reported high levels of familiarity (Scale = 1–5, Grand Mean = 4.26, Std. = 0.55). All items were above 4 (high level of familiarity), except for the terms related to performance assessment. These latter terms, which include Alternative Assessment and Authentic Assessment, were nevertheless familiar to instructors (above 3). The Validation components of the survey, which addresses terms relating to the evaluation and quality control of assessment development, were also familiar to instructors (Grand Mean = 3.34, Std. = 0.35). This result is in line with prior studies of faculty knowledge of assessment terms (Hanauer and Bauerle, 2015). The results overall for the two dimensions suggest that instructors in this study have the required degree of assessment understanding to be reliable reporters of their assessment procedures and activities.
To augment the FRAS data, self-efficacy data were collected on instructors’ confidence in completing assessment related tasks. Internal consistency was calculated for the self-efficacy scale. Cronbach’s Alpha was 0.93 which shows that this scale is reliable As shown in Table 3, instructors reported high levels of confidence in their assessment abilities (Scale = 1–5, Grand Mean = 4.04, Std. = 0.65). The highest confidence was in relation to defining important components of their course and student learning outcomes, while the lowest levels of confidence were in relation to the ability to evaluate, analyze and report on their assessments. The confidence levels for the latter were still relatively high (just below 4) and reflect, to a certain extent, the same trend as seen using the FRAS instrument. Taking into consideration the results of the FRAS and self-efficacy tasks, instructors report moderate to high levels of assessment expertise and confidence, which suggest that these instructors have the required expertise to report and evaluate the aims, practices, and models of CRE assessment.
Aims and practices of CRE assessment
A fundamental goal of this study was to describe the aims and practices of experienced CRE instructors for assessing students in a CRE. As described in the methodology section, a list of aims and practices for assessment was elicited from the written survey data completed by instructors in the HHMI SEA program, which was then community-checked and modified. The faculty were asked to describe how they assess students in the SEA program what types of assessment used (such as quiz, rubric…etc.), and to explain what each assessment is used for. The aims specified by the faculty reflected components of pedagogical activity that came together while teaching a CRE. So, for example, assessing the physical work of lab was integrated with scientific thinking as a single aim. Broadly the aims reflected work in the laboratory, aspects of mastery, communication and student self-evaluation of their learning.
Four central aims of CRE assessment were defined. For each aim, there were a cluster of assessment practices that were employed to assess student learning, with different instructors utilizing different subsets of these practices. The aims of CRE assessment, the practices related to each of the aims, and the degree of agreement among faculty for each aim and set of practices are presented in Table 4 and described below:
1. Assess laboratory work and scientific thinking: The objective of this assessment aim was to assess a student’s readiness, in terms of their practices, thought patterns and ethics, to function as a researcher in the laboratory setting. As seen in Table 4, several different practices were related to this aim, which include (1) assessing student behaviors such as participation, attendance, citizenship, collaboration, safety and independence, and (2) assessing students’ scientific thinking based on their lab notebooks, data cards, independent research, conference participation, and informal discussion. During the community checking stage, 85.95% of the faculty specified that this category was an aim of their assessment program and that the assigned practices were appropriate.
2. Evaluate mastery of concepts, quantitative thinking, and skills: The objective of this assessment aim was to assess the underpinning knowledge and skills that students need in order to function successfully, as a researcher, in the CRE laboratory setting. The practices related to this assessment aim include (1) the checking of laboratory techniques and skills using practical exams and lab notebooks, (2) the evaluation of required scientific knowledge through exams, tests, quizzes, written reports, and articles, and (3) the assessment of quantitative knowledge. During the community checking stage, 80.99% of faculty specified that this category was an aim of their assessment program and that the assigned practices were appropriate.
3. Appraise forms of scientific communication: The objective of this assessment aim was to evaluate the ability of students to convey their research and attain scientific knowledge through the different forms of science communication. The practices related to this assessment include (1) oral abilities such as oral presentation, peer review, lab notebook meetings, scientific poster, and elevator speech, and (2) literacy abilities such as reading and writing a research paper, report writing, notebook writing, scientific paper reading, literature review, and poster creation. 63.64% of faculty specified that this category was part of their assessment program.
4. Metacognition of learning: The objective of this assessment aim was to assess the ability of students to regulate and oversee their own learning process. This aim is based on the assumption that being in control of your learning process improves the ability to learn. The practices related to this aim include reflection, discussion and an exit ticket. 76.85% of faculty specified that this category was part of their assessment program.
These four aims and associated practices define a program of assessment for CRE teaching. As depicted in Figure 1, the central aspect of an assessment program for a CRE is to evaluate the ability of a student to work and think in a scientific way. This central aspect is supported by two underpinning forms of knowledge: (1) mastery of concepts, quantitative thinking, and skills and (2) the ability to communicate science. Overseeing the whole process is metacognition, which allows the student to regulate and direct their learning process. Accordingly, information on the students’ functioning across all these areas are collected as part of the assessment program.
Models of assessment in a CRE
The assessment program presented in this study is implemented by instructors in conjunction with a program of CRE instruction that has been previously described (Hanauer et al., 2022). The assessment aims and practices described here can therefore be integrated with the aims and practices (or models) of CRE instruction. The stated aims of CRE instruction are (1) Facilitating the experience of being a scientist and generating data; (2) Developing procedural knowledge, that is the skills and knowledge required to function as a researcher; and (3) Fostering project ownership, which include the feelings of personal ownership and responsibility over their scientific research and education (Hanauer et al., 2022). These aims are directly in line with the broad aim of a CRE in providing a student with an authentic research experience (Dolan and Weaver, 2021). In the sections that follow, and using a constructive alignment approach (Biggs, 1996; Ambrose et al., 2010), the assessment aims and practices uncovered in this study are presented with the associated models of CRE instruction previously described.
Model 1: assessing being a scientist and generating data
Being a scientist and generating novel data is a core aspect of a CRE. As shown in Figure 2 and described below, the instructional approach to achieving this aim involves three stages of instruction:
a. Stage 1 involves preparing the student with the required knowledge and procedures in order to function as a researcher who can produce usable data for the scientific community. The pedagogy employed here includes the use of explicit instruction to provide students with the foundational knowledge to understand the science they are involved with and protocol training to make sure a student can perform the required scientific task.
Accordingly, assessment in this first stage of the model is aimed at Evaluating Mastery of Concepts and Quantitative Thinking. The assessment practices used here include both exams and in class quizzes, which are well suited for this purpose. Additionally, given that this foundational scientific knowledge must often be retrieved from various forms of scientific communication, including lecture, a research paper, a poster, and an informal discussion with an expert, the ability to use scientific communication for knowledge acquisition is also evaluated. Practices such as the evaluation of a literature search report or presentation at a journal club can provide information on how the student understands and uses different modes of scientific communication. Combined, the use of exams, quizzes, literature search reports, and journal club participation can provide a rich picture of the foundational knowledge of a student as they enter the process of doing authentic research.
To assess a student’s ability to use a range of specific protocol properly, instructors rely on practical exams and a student’s lab notebook, which are well established ways of checking whether a student understands and knows how to perform a specific procedure. Beyond these approaches, instructors reported that they used informal discussion, reflective writing, article writing, and the lab notebook meeting to evaluate formally and informally whether the students understand how to perform the different scientific tasks that are required of them. This combination of explicit teaching of scientific knowledge and procedures, with formal and informal assessment of these abilities, serves to create a basis for the second stage of this pedagogical model, described below.
b. Stage 2 involves supporting students to manage the process of implementing procedures in order to generate authentic data. A central aspect of this stage is that the student moves from a consumer to a producer of knowledge, and this involves a change in the students’ mindset concerning thinking processes, independence, perseverance, and the ability to collaborate with others. Importantly, as is the case with science, positive results are not guaranteed and students face the ambiguity of failed outcomes and unclear paths forward. It is for this reason that the pedagogy at this stage involves a range of different supportive measures on the part of the instructor. These include modeling scientific thinking, providing encouragement and enthusiasm, mentoring the student at different points and, most importantly, making sure that the students understand that the scientific process is one that is fraught with challenges that need to be overcome. A lot of instruction is provided at the time that a task or event occurs.
Assessment at this stage is covered by the aim of Assessing Laboratory Work and Scientific Thinking and the Metacognition of Learning. The scientific thinking of the student is primarily assessed through the discussion of the lab notebook, data and annotation cards, often during lab meetings. Importantly, as reported by faculty, a lot of this assessment is directed by informal discussion with the aim of providing direct feedback to the student so that they can perform the tasks that are required. This is very much a formative assessment approach with direct discussion with the student while they are working and in relation to the research they are doing. There are behaviors that faculty specify are important to track, such as participation, attendance, collaboration, lab citizenship and lab safety. These behaviors are a prerequisite for the research to move forward for the student and the research group as a whole. The use of assessment practices such as reflection and discussion allows the assessment of the degree of independence of the student, in addition to actually positioning the student as independent; the requirement of a reflection task, whether written in one’s lab notebook or verbally, situates the students as the researcher thinking through what they are doing. Overall, this stage involves extensive informal formative assessment of where the student is in the process from the practical, scientific and emotional aspects of doing science, combined with a more formal evaluation of the behaviors which underpin a productive and safe research environment.
c. The third and final stage of this pedagogical model involve the actual scientific output produced by the student researcher. A CRE is defined by the requirement that data are produced that is actually useful for a broader community of scientists. If the second stage of the assessment of this pedagogical model is characterized by informal, formative assessment approaches, this final stage is characterized primarily by formal summative assessment. At this stage, the student has produced scientific knowledge and is in the process of reporting this knowledge using established modes of scientific communication. The student is assessed in relation to the knowledge they have produced and the way they communicate it. As such, both the aims of Assessing Laboratory Work and Scientific Thinking and the Appraisal of Forms of Scientific Communication are utilized. The lab notebook, data card, annotation, conference presentation, oral presentation, and poster all involve a double summative assessment approach: an evaluation of the quality of the scientific work that has been produced and an evaluation of the ability of the student to communicate this knowledge using established written and verbal modes of scientific communication. This final stage provides the opportunity for evaluating the whole of the research experience that the student has been involved in.
To summarize, the instruction and assessment model of Being a Scientist and Generating Data has three distinct stages. The initial stage is designed to make sure that the student can perform the required tasks and understand the underlying science. Assessment at this stage is important as the learning involved in this stage is a prerequisite for the second stage of the model. During the second stage, while the student is functioning as a researcher, the primary focus of the assessment model is to provide feedback to the student and the required level of expertise advice and emotional support to allow the research to move forward. This stage is characterized by informal discussion and is primarily a formative assessment approach. The final stage is directed at evaluating the scientific outcomes and the student’s ability to communicate them. Assessment at this stage offers a direct understanding of the quality of the work that has been conducted, the degree to which the student understands the work, and the ability of the student to communicate it.
Model 2: assessing procedural knowledge
Being able to perform a range of scientific procedures is a central and underpinning aspect of being a scientist and a core feature of a CRE. Figure 3 presents a pedagogical and assessment model for teaching procedural knowledge. As seen in the previous model, protocols are an important precursor that enables an undergraduate student to conduct scientific research. In model 2, how students learn scientific procedures is further explicated from model 1. As can be seen in Figure 3, there are three stages to the development of procedural knowledge.
a. The first stage involves enhancing the students’ content knowledge concerning the science behind the protocol they are using and scientific context of the research they will be involved with. For a student to become an independent researcher, they need to be able to not just follow a set of procedures but also to understand the science that it relates to. The pedagogical practice involved here includes explicit instruction, discussion and reading of primary literature. From an assessment perspective, the evaluation of this underpinning content knowledge is conducted using established practices such as exams, tests, and quizzes. In addition, as reported by faculty, this material was informally discussed with students to gauge understanding of the context and role of the procedure.
b. In the second stage, students are taught how to implement the procedure and to think like a scientist. This involves using a protocol, scientifically thinking through the process of using a protocol, and appropriate documentation of the process of using a protocol. Scientific thinking at this stage includes interpretation of outcomes, problem solving, and deciding about next steps. In this way, learning a protocol is not only about being able to perform, analyze and document a procedure appropriately, but also involves the development of independence for the researcher. These two components are related in that if a student really has a full understanding of the procedure, they can also make decisions and function more autonomously. Such mastery is particularly critical in a CRE because the research being conducted is intended to support an ongoing authentic research program. As reported by faculty, there are both formal and informal assessments that facilitate this evaluation. Practical exams allow faculty to really check the performance of a particular procedure and their understanding. Lab notebook evaluation, lab meeting interactions and informal discussion about the work of a student as they perform certain tasks provides further evidence of the student’s mastery of the concepts and skills that are involved. These interactions are primarily formative and have the aim of providing feedback for the improvement of the student’s understanding of scientific procedures.
c. An additional level of assessment at this stage relates to the ability of students to document their research in the lab notebook, explain their research in a lab meeting and to converse with peers and instructors about what they are doing. These are all aspect of scientific communication, and assessment at this second stage of learning procedural knowledge includes the aims evaluating mastery of concepts and skills and of an appraisal of scientific communication. Since these are new forms of communication for many undergraduate students, instructors report using rubrics to evaluate and provide feedback on the quality of the communication.
d. The final stage of this model relates to the scientific outcomes of the students’ work. At this stage, assessment aims to evaluate the quality of the outcomes of these procedures and the level to which the student really understands what they have done. Evaluation here therefore combines the use of data cards, annotation outputs, lab notebooks, oral presentations, conference participation, and the student’s reflections on their own work. As reported by faculty, not all procedures are successful and students are not graded negatively for a failed experiment as long as the procedures, including the thinking involved, follows the scientific process. Thus, as reported by faculty, both the instructor and the student often work collaboratively to evaluate how well the student understands the different procedures they are learning to use.
Model 3: assessing the facilitation of project ownership
The educational practice of a CRE involves a desired transition of the student from being a more passive learner of knowledge to being an active producer of knowledge who is integrated into a larger community of researchers. This transition, in which the student has a sense of ownership over their work and responsibility over their research and learning, is an aim of CRE pedagogy and has important ramifications to being a student researcher (Hanauer et al., 2022). Furthermore, prior research has shown that the development of a sense of project ownership differentiates between an authentic research experience and a more traditional laboratory course. Figure 4 presents the pedagogical and assessment model of fostering project ownership. The model has three stages of development.
a. The first stage of fostering project ownership is developing in students a broad understanding and ability to perform a range of scientific protocols. This is because project ownership requires the belief and the ability to actually do science. It is an issue of self-efficacy and mastery of concepts and skills. As such, the first stage of assessment involves evaluating the degree of mastery a student has over a specific protocol. As opposed to prior models, this is enacted here through formative, informal discussions, which also serves to enhance that mastery.
b. The second stage of the model aims to develop the student’s sense of personal responsibility. Primary to this process is the promotion and encouragement of the student’s independence. This can involve both emotional supports, the provision of resources, and the allotment of time for the student to ponder the work that they are doing. As reported by faculty, not every question has to be or can be answered immediately. Allowing a student to think about their work and what they think should be done is an important aspect of a CRE education. Accordingly, a central component of the assessment model here is having the student reflect on their work. The task of assessment here thus expands beyond the instructor to student as well.
c. A different aspect of both fostering and assessing responsibility and ownership over one’s research involves a series of behaviors related to scientific work. Faculty report assessing lab citizenship, collaboration and lab safety protocols. Being responsible includes behaving in appropriate ways in the laboratory and as such these aspects of the students’ work are evaluated. Some faculty also reported that having the student propose projects that extend the ongoing classroom research project allowed them to assess the degree of independence of the student.
d. The final stage of the model involves situating the student-researcher within a broader scientific context. Talking with the student about future careers and educational opportunities, and providing encouragement and enthusiasm for the work the student is doing positions the student at the center of their own development. Project ownership involves pride in the research one is doing and seeing ways in which this work can be developed beyond the specific course. Once again, reflection plays a central role in assessing and facilitating this, and occurs as an informal and ongoing process.
e. In parallel, the outcomes of the research the student does is reported using established modes of scientific communication. A student is responsible for reporting their work using oral presentations, scientific posters, research papers, and reports. At this point, they will receive feedback on their work in both formal and informal ways. One important aspect of this reporting is the real-world evaluation of their output. Other peer student researchers may respond, in addition to faculty and scientists beyond the classroom. Having ownership over one’s research also includes an understanding that the work will be evaluated beyond the classroom grade and that the work itself is part of a far larger community of scientists. In this sense, the evaluation of the scientific output facilitates ownership of the research itself.
Figure 4. Assessing the facilitation of project ownership: this model has three distinct stages. The first stage relates to development of understanding concerning protocol usage. The second stage relates to the fostering of the student’s sense of personal responsibility. The third stage involves situating the student within the broader scientific context. The model presents the aims and practices of assessment applied at each of these stages.
Discussion
The main aim of this paper is to explore how assessment of students engaged in course-based research is implemented and aligned with the educational goals of this form of pedagogy. In terms of constructive alignment, the aims of any assessment program should reflect and support defined instructional objectives. Assessment of scientific inquiry, as is typically implemented in traditional labs, focus on mastery of the components of research (see Wenning, 2007 for an example). The aim of instruction and assessment within a traditional lab is to make sure that a defined procedure has been mastered by the student so that in some future course or scientific project, the student knows how to perform it. In the traditional lab, grading is evidence of qualification for the student’s ability to function in a future scientific activity. Failure, if it happens, is indeed failure and a reason for not progressing further.
In contrast, a CRE aims to provide the student with an authentic research experience in which they are contributors of research data that is useful for advancing science. As such, mastery is a necessary but not sufficient aim of assessment. As specified by instructors in this study, mastery of concepts, quantitative thinking and skills is important in order to conduct and understand a scientific process; but this is situated in relation to the actual performance of scientific research (also an aim of assessment), which involves an understanding of how to communicate science and ownership over one’s learning and research activity. Thus, from the perspective of what to assess, it is clear that assessment in a CRE needs a broader approach than the assessment program of traditional labs. In this study, four aims of assessment were defined by experienced CRE instructors: (1) Assessing Laboratory Work and Scientific Thinking; (2) Evaluating Mastery of Concepts, Quantitative Thinking and Skills; (3) Appraising Forms of Scientific Communication; and (4) Metacognition of Learning.
The alignment between these assessment aims and the aims of CRE instruction is further explicated here. Across the instructional aims of Facilitating Being a Scientist and Generating Data, Developing Procedural Knowledge, and Fostering Project Ownership, the four aims of assessment were seen to provide ways of collecting useful data that supports the progress of students toward these stated aims of CRE instruction. With regard to how assessment data are collected in a CRE, there are particular relationships between formal and informal assessment and the formative and summative approaches. Summative assessment with formalized tools tended to be at the beginning and end of a research process, in relation to first the development of required mastery of concept and skills and last the evaluation of scientific outputs, which are the products of the research. Mastery can be evaluated using tests and exams, while products can be evaluated using rubrics. In contrast, during the process of conducting the research project, the emphasis is on providing feedback to students to help support the ongoing work. This includes the use of a range of laboratory practices, such as lab notebook documentation and lab meetings. And while assessment data are collected, the response is often informal and formative with the aim of supporting the student to further their research.
Beyond collecting assessment data, there is also a particular way in which assessment, evaluation and grading manifest in a CRE setting. The terms of assessment, evaluation and grading are often used interchangeably. But these terms relate to different concepts. Assessment is primarily a data collection and interpretation task; evaluation is a judgment in relation to the data collected; and grading is a definitive decision expressed as a number or letter as to the final quality of the work of a student. The majority of institutions require grades for a CRE. But not all things that are assessed in a CRE need to be graded. In particular, informal discussion with students of the different aspects of the scientific tasks students are performing allows the instructor to provide supportive feedback that facilitates the scientific inquiry. This informal, formative assessment does not require a grade directly. At the same time, there is a role for assessing and grading the underpinning knowledge, behaviors (such as lab citizenship, attendance, participation, collaboration, and lab safety), and scientific outputs of the students. Thus, there is a two-tiered assessment and grading process in which, during the process of scientific inquiry, which is the majority of the course time, assessment data are collected but not graded; however, the knowledge, skills, behaviors, and outcomes are graded. Since the aim of the whole course is to give the student the experience of being a researcher and to produce scientific data, providing facilitative feedback based on assessment during the research process helps the student to complete the tasks in a meaningful way. The grading of the underpinning knowledge, skills, and behaviors also facilitates the work that is conducted in laboratory. Without appropriate mastery and behavior, the lab research will not be possible. Thus, once again, the form of assessment supports the progress of authentic research. As presented in this study, the way to grade a CRE is to differentiate the framing of the research that is conducted from the process of doing the research; provide extensive formative assessment in an informal manner throughout the research process; grade the underpinning components of knowledge, skill, and behavior; and provide a final grade which weights the quality of the work and the output that is produced. The aim should be for every student to be successful in the research process and assessment should facilitate this work.
The assessment and grading practices presented here are clearly facilitative of student learning. First, knowledge, skills, and behaviors are measured because they are foundational for students to productively engage in their research. Second, a large part of the assessment work is directly aimed at providing feedback without penalizing a student through grade assignment. There is extensive informal formative assessment that can be seen as a departure from assessment in more traditional labs and which approximates the type of facilitation that characterize mentor-mentee relationships in authentic research settings (e.g., in individual undergraduate research experiences, postbaccalaureate research opportunities, or during postgraduate research). This mentor-mentee relationship can build trust and counter stereotype threat to enhance persistence and learning. Additionally, an assessment program with extensive informal formative assessments leaves fewer instances when a student might be penalized by grading and suffer the negative psychological effects associated with lower grading. Third, the components of CRE assessment address a broad range of skills, beyond just mastery of procedures that a student needs as a scientist and a learner. In particular, included within the aims of CRE assessment are scientific communication and metacognition. Scientific communication is an important component of being a researcher, while metacognition not only provides information that can be used to evaluate where a student is and how they are thinking about their work, but also positions the student as an evaluator of their own work. In this case, the task of assessment itself directs the students toward better learning and might explain why CREs improve student learning despite the CRE content not always being directly aligned with lecture content (in comparison to traditional lab). We hypothesize that these various aspects of CRE assessment contribute to the positive outcomes observed for students across many demographics and when compared to the traditional lab.
As presented in the introduction, a CRE poses quite specific challenges in terms of assessment and grading. A primary concern relates to the need to maintain a professional shared research project with contributions from instructor and student, while still assessing and grading a student. As presented here, this delicate balancing act is facilitated by using assessment and grading thoughtfully and in a coordinated manner. If the instructor is providing extensive feedback that supports the work of the student and grades the aspects of science that are necessary for the student to succeed, the relationship with the student is different from a relationship in which the teacher is just grading a student. The assessment models presented here provide a framework to facilitate the aims of a CRE without undercutting the broader aims of promoting student learning and persistence in science, and can serve to inform assessment and grading practices in STEM, more generally.
Limitations
The data and analyses presented in this study emerged from a collective process with a large number of faculty who all implement CREs through the Science Education Alliance (SEA) program by HHMI. Organized as an inclusive Research and Education Community (iREC), faculty in the SEA program are supported by centralized programming to lead the instruction of research projects with a shared research agenda (Hanauer et al., 2017). This does have some ramifications that limit the generalizability of the current results. First, CREs with different research agendas and that require different procedures may change the ratios of formal and informal assessment and what is considered important for grading. Second, while the instructors do work at a wide range of institutions, they also work together in SEA. There is extensive interaction between instructors facilitated by yearly in-person faculty meetings, monthly science and education seminars, and on-line shared resources. This familiarity, interaction and shared course components can lead to a degree of homogeneity in relation to how procedures such as assessment and grading are conducted. As the SEA community facilitated the current data collection and analysis process, it can limit results by not including a much broader set of underlying CRE educational and scientific designs.
Conclusion
Course-based research experiences are increasingly implemented at institutions of higher learning because they offer a strategy to scale-up opportunities for students to engage in authentic research, which is strongly correlated with an increased persistence in science for a wide range of student populations (Russell et al., 2007; Jordan et al., 2014; Hanauer et al., 2017; Hernandez et al., 2018). However, given that CREs situate the research opportunity within the context of a course, it is critically important that the involvement of course grading does not negatively influence students’ belief in their abilities and willingness to persist in STEM (Hatfield et al., 2022). As seen in the reviews of the multiple instruments developed for the assessment of students in a CRE, the past tendency has been to conceptualize the goals of CRE as a set of skills, competencies, dispositions and abilities to be gained by students for their future engagement in research (Shortlidge and Brownell, 2016; Zelaya et al., 2022). The assessment of such externalized goals instead of the actual science and scientific process that is at the core of the CRE can lessen the value of the research students are engaged in and contradict their self-perception as researchers.
In contrast, the study presented here models how faculty actively teaching in a large CRE program have integrated assessment into their CRE pedagogy in a way that supports the actual research that is being conducted. In this way, assessment and grading are directly tied to the intended value and aim of a CRE in providing students with an opportunity to engage in research authentically. This is particularly critical because students’ sense of being a scientist is foundational to long-term persistence in the sciences and inappropriate assessment and grading practices could interfere with the positive social and educational values embedded in a CRE (Hanauer et al., 2017). The models of assessment presented here describe how assessment and grading can be conceptualized and implemented in a way that maintains the student’s authentic sense of being a researcher. The approach to assessment described in this paper, which emerged from an extensive interaction with a large community of faculty who actively teach a CRE, describes ways in which assessment can support the educational and social agenda of a CRE. We hope that this study will encourage other researchers working a wider range of CREs to study their own assessment and grading objectives and practices and consider the ways in which assessment can facilitate and not hinder the student’s research experience.
Data availability statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.
Ethics statement
The studies involving humans were approved by Indiana University of Pennsylvania, IRB #21-214. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
DHa: Conceptualization, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing. TZ: Formal analysis, Writing – review & editing. MGr: Conceptualization, Writing – review & editing. SA: Validation, Writing – review & editing. YA-S: Validation, Writing – review & editing. RA: Validation, Writing – review & editing. MAn: Validation, Writing – review & editing. MAy: Validation, Writing – review & editing. ME: Validation, Writing – review & editing. CBa: Validation, Writing – review & editing. TB: Validation, Writing – review & editing. MBe: Validation, Writing – review & editing. EB: Validation, Writing – review & editing. AB: Validation, Writing – review & editing. RB: Validation, Writing – review & editing. DB: Validation, Writing – review & editing. LB: Validation, Writing – review & editing. VB-K: Validation, Writing – review & editing. MBu: Validation, Writing – review & editing. SB: Validation, Writing – review & editing. KB: Validation, Writing – review & editing. CBy: Validation, Writing – review & editing. SCa: Validation, Writing – review & editing. CC: Validation, Writing – review & editing. RC: Validation, Writing – review & editing. H-MC: Validation, Writing – review & editing. KCl: Validation, Writing – review & editing. SCol: Validation, Writing – review & editing. DPa: Validation, Writing – review & editing. SCon: Validation, Writing – review & editing. BrC: Validation, Writing – review & editing. PC: Validation, Writing – review & editing. BeC: Validation, Writing – review & editing. JC-E: Validation, Writing – review & editing. KCr: Validation, Writing – review & editing. TD'E: Validation, Writing – review & editing. MD: Validation, Writing – review & editing. LDe: Validation, Writing – review & editing. LDi: Validation, Writing – review & editing. ID: Validation, Writing – review & editing. NE: Validation, Writing – review & editing. DE: Validation, Writing – review & editing. TE: Validation, Writing – review & editing. EE: Validation, Writing – review & editing. PF-S: Validation, Writing – review & editing. CF: Validation, Writing – review & editing. AFi: Validation, Writing – review & editing. EF: Validation, Writing – review & editing. MFi: Validation, Writing – review & editing. MFo: Validation, Writing – review & editing. AFr: Validation, Writing – review & editing. VF: Validation, Writing – review & editing. MGa: Validation, Writing – review & editing. AmG: Validation, Writing – review & editing. AtG: Validation, Writing – review & editing. HG: Validation, Writing – review & editing. RG: Validation, Writing – review & editing. BG: Validation, Writing – review & editing. UG: Validation, Writing – review & editing. AnG: Validation, Writing – review & editing. SG: Validation, Writing – review & editing. RH: Validation, Writing – review & editing. SH: Validation, Writing – review & editing. JH: Validation, Writing – review & editing. LH: Validation, Writing – review & editing. PJ: Validation, Writing – review & editing. KJ: Validation, Writing – review & editing. AJ: Validation, Writing – review & editing. MKa: Validation, Writing – review & editing. MKe: Validation, Writing – review & editing. BK: Validation, Writing – review & editing. KKl: Validation, Writing – review & editing. KKo: Validation, Writing – review & editing. MKu: Validation, Writing – review & editing. AL: Validation, Writing – review & editing. JL-S: Validation, Writing – review & editing. LL: Validation, Writing – review & editing. HL: Validation, Writing – review & editing. JMa: Validation, Writing – review & editing. SMa: Validation, Writing – review & editing. MMa: Validation, Writing – review & editing. VM: Validation, Writing – review & editing. SMc: Validation, Writing – review & editing. EM: Validation, Writing – review & editing. JMe: Validation, Writing – review & editing. SMi: Validation, Writing – review & editing. JMi: Validation, Writing – review & editing. SMo: Validation, Writing – review & editing. DM: Validation, Writing – review & editing. MMu: Writing – review & editing, Validation. HN: Validation, Writing – review & editing. FN-F: Validation, Writing – review & editing. JN: Validation, Writing – review & editing. IN: Validation, Writing – review & editing. MO'D: Validation, Writing – review & editing. SP: Validation, Writing – review & editing. AP: Validation, Writing – review & editing. JP-U: Validation, Writing – review & editing. TaP: Validation, Writing – review & editing. TiP: Validation, Writing – review & editing. NP: Validation, Writing – review & editing. VP: Validation, Writing – review & editing. RP: Validation, Writing – review & editing. MP: Validation, Writing – review & editing. DPu: Validation, Writing – review & editing. MR: Validation, Writing – review & editing. NR: Validation, Writing – review & editing. CR: Validation, Writing – review & editing. JR: Validation, Writing – review & editing. OR: Validation, Writing – review & editing. ARu: Validation, Writing – review & editing. ER: Validation, Writing – review & editing. ARy: Validation, Writing – review & editing. SS: Validation, Writing – review & editing. CSh: Validation, Writing – review & editing. MS: Validation, Writing – review & editing. AS: Validation, Writing – review & editing. CSt: Validation, Writing – review & editing. CSu: Validation, Writing – review & editing. BT: Validation, Writing – review & editing. LT: Validation, Writing – review & editing. KT: Validation, Writing – review & editing. MT: Validation, Writing – review & editing. DT: Validation, Writing – review & editing. ST: Validation, Writing – review & editing. JG: Validation, Writing – review & editing. MV: Validation, Writing – review & editing. EV: Validation, Writing – review & editing. RW: Validation, Writing – review & editing. CW: Validation, Writing – review & editing. VW: Validation, Writing – review & editing. MW: Validation, Writing – review & editing. JW: Validation, Writing – review & editing. DW: Validation, Writing – review & editing. KW: Validation, Writing – review & editing. BW: Validation, Writing – review & editing. EW: Validation, Writing – review & editing. WB: Project administration, Resources, Writing – review & editing. SCr: Validation, Writing – review & editing. DHe: Validation, Writing – review & editing. DJ-S: Validation, Writing – review & editing. GH: Project administration, Supervision, Validation, Writing – review & editing. DA: Funding acquisition, Project administration, Supervision, Validation, Writing – review & editing. VS: Funding acquisition, Project administration, Resources, Supervision, Validation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was funded by a grant awarded to DHe by the Howard Hughes Medical Institute (GT#12052).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Ambrose, S.A., Bridges, M.W., DiPietro, M, Lovett, M.C., and Norman, M.K. (2010). How Learning Works: Seven Research-Based Principles for Smart Teaching (1st Edn). San Francisco, CA: Jossey-Bass
Auchincloss, L. C., Laursen, S. L., Branchaw, J. L., Eagan, K., Graham, M., Hanauer, D. I., et al. (2014). Assessment of course-based undergraduate research experiences: a meeting report. CBE Life Sci. Educ. 13, 29–40. doi: 10.1187/cbe.14-01-0004
Bandura, A. (1991). Social cognitive theory of self-regulation. Organ. Behav. Hum. Decis. Process. 50, 248–287. doi: 10.1016/0749-5978(91)90022-L
Bandura, A. (2005). “The evolution of social cognitive theory” in Great Minds in Management. eds. K. G. Smith and M. A. Hitt (Oxford: Oxford University Press), 9–35.
Biggs, J. (1996). Enhancing teaching through constructive alignment. High. Educ. 32, 347–364. doi: 10.1007/BF00138871
Boothroyd, R. A., and McMorris, R. F. (1992). “What do teachers know about testing and how did they find out?” in Paper presented at the annual meeting of the National Council on Measurement in Education, San Francisco, CA.
Butler, R. (1988). Enhancing and undermining intrinsic motivation: the effects of task-involving and ego-involving evaluation on interest and performance. Br. J. Educ. Psychol. 58, 1–14. doi: 10.1111/j.2044-8279.1988.tb00874.x
Cizek, G. J., Fitzgerald, S. M., and Rachor, R. E. (1996). Teachers’ assessment practices: preparation, isolation, and the kitchen sink. Educ. Assess. 3, 159–179. doi: 10.1207/s15326977ea0302_3
Crooks, T. J. (1988). The impact of classroom evaluation practices on students. Rev. Educ. Res. 58, 438–481. doi: 10.3102/00346543058004438
Dolan, E. L., and Weaver, G. C. (2021). A Guide to Course-Based Undergraduate Research. Austin: MacMillan Learning.
Fajardo, D. M. (1985). Author race, essay quality, and reverse discrimination. J. Appl. Soc. Psychol. 15, 255–268. doi: 10.1111/j.1559-1816.1985.tb00900.x
Graham, M., Frederick, J., Byars-Winston, A., Hunter, A.-B., and Handelsman, J. (2013). Increasing persistence of college students in STEM. Science 341, 1455–1456. doi: 10.1126/science.1240487
Hanauer, D. I., and Bauerle, C. (2015). The Faculty Self-Reported Assessment Survey (FRAS): differentiating faculty knowledge and experience in assessment. CBE Life Sci. Educ. 14:ar17.
Hanauer, D. I., and Dolan, E. (2014). The project ownership survey: measuring differences in scientific inquiry experiences. CBE Life Sci. Educ. 13, 149–158. doi: 10.1187/cbe.13-06-0123
Hanauer, D. I., Frederick, J., Fotinakes, B., and Strobel, S. A. (2012). Linguistic analysis of project ownership for undergraduate research experiences. CBE Life Sci. Educ. 11, 378–385. doi: 10.1187/cbe.12-04-0043
Hanauer, D. I., Graham, M. J., Arnold, R. J., Ayuk, M. A., Balish, M. F., Beyer, A. R., et al. (2022). Instructional models for course-based research experience (CRE) teaching. CBE Life Sci. Educ. 21, 1–14. doi: 10.1187/cbe.21-03-0057
Hanauer, D. I., Graham, M. J., Betancur, L., Bobrownicki, A., Cresawn, S. G., Garlena, R. A., et al. (2017). An inclusive research education community (iREC): impact of the SEA-PHAGES program on research outcomes and student learning. Proc. Nat. Acad. Sci. U.S.A. 114, 13531–13536. doi: 10.1073/pnas.1718188115
Hanauer, D. I., Graham, M., and Hatfull, G. F. (2016). A measure of student persistence in the sciences (PITS). CBE Life Sci. Educ. 15:ar54. doi: 10.1187/cbe.15-09-0185
Hanauer, D. I., Hatful, G.F., and Jacobs-Sera, D. (2009). Active Assessment: Assessing Scientific Inquiry. New York: Springer
Hanauer, D. I., Jacobs-Sera, D., Pedulla, M. L., Cresawn, S. G., Hendrix, R. W., and Hatfull, G. F. (2006). Teaching scientific inquiry. Science 314, 1880–1881. doi: 10.1126/science.1136796
Hatfield, N., Brown, N., and Topaz, C. M. (2022). Do introductory courses disproportionately drive minoritized students out of STEM pathways? PNAS Nexus 1, 1–10. doi: 10.1093/pnasnexus/pgac167
Hernandez, P. R., Woodcock, A., Estrada, M., and Wesley Schultz, P. (2018). Undergraduate research experiences broaden diversity in the scientific workforce. BioScience, 68, 204–211.
Hilts, A., Part, R., and Bernacki, M. L. (2018). The roles of social influences on student competence, relatedness, achievement, and retention in STEM. Sci. Educ. 102, 744–770. doi: 10.1002/sce.21449
Hu, S. (2005). Beyond grade inflation: grading problems in higher education. ASHE High. Educ. Rep. 30, 1–99. doi: 10.1002/aehe.30061
Jaschik, S (2009). Imagining college without grades. Available at: www.insidehighered.com/news/2009/01/22/grades (Accessed February 20, 2023).
Jordan, T. C., Burnett, S. H., Carson, S., Caruso, S. M., Clase, K., DeJong, R. J., et al. (2014). A broadly implementable research course in phage discovery and genomics for first-year undergraduate students. MBio 5, e01051–e01013. doi: 10.1128/mBio.01051-13
Lipnevich, A. A., Guskey, T. R., Murano, D. M., and Smith, J. K. (2020). What do grades mean? Variation in grading criteria in American college and university courses. Assess. Educ. 27, 480–500. doi: 10.1080/0969594X.2020.1799190
National Science Board (2018). Science and engineering indicators 2018 (Tech. Rep. No.NSB-2018-1). National Science Foundation. Available at: https://www.nsf.gov/statistics/indicators/
Pelaez, N., Anderson, T. R., Gardner, S. M., Yin, Y., Abraham, J. K., Bartlett, E., et al. (2017). The basic competencies of biological experimentation: Concept-skill statements. PIBERG Instructional Innovation Materials. Paper 4.
President’s Council of Advisors on Science and Technology (2012). Engage to excel: Producing one million additional college graduates with degrees in science, technology, engineering, and mathematics. U.S. Government Office of Science and Technology, Washington, DC.
Pulfrey, C., Buchs, C., and Butera, F. (2011). Why grades engender performance-avoidance goals: the mediating role of autonomous motivation. J. Educ. Psychol. 103, 683–700. doi: 10.1037/a0023911
Russell, S. H., Hancock, M. P., and McCullough, T. (2007). The pipeline: benefits of undergraduate research experiences. Science 316, 548–549. doi: 10.1126/science.1140384
Shavelson, R., Solano-Flores, G., and Araceli Ruiz-Primo, M. (1998). Toward a science performance assessment technology. Eval. Program. Plan. 21, 171–184. doi: 10.1016/S0149-7189(98)00005-6
Shortlidge, E. E., and Brownell, S. E. (2016). How to assess your CURE: a practical guide for instructors of course-based undergraduate research experiences. J. Microbiol. Biol. Educ. 17, 399–408. doi: 10.1128/jmbe.v17i3.1103
Walvoord, B. E., and Anderson, V.J. (1998). Effective Grading: A Tool for Learning and Assessment. San Francisco: Jossey-Bass
Wenning, C. (2007). Assessing inquiry skills as a component of scientific literacy. J. Phys. Educ. 4, 21–24.
Whitcomb, K. M., and Chandralekha, S. (2021). Underrepresented minority students receive lower grades and have higher rates of attrition across STEM disciplines: a sign of inequity? Int. J. Sci. Educ. 43, 1054–1089. doi: 10.1080/09500693.2021.1900623
Zachos, P. (2004). Pendulum phenomena and the assessment of scientific inquiry capabilities. Sci. Educ. 13, 743–756. doi: 10.1007/s11191-004-6179-3
Zelaya, A. J., Blumer, L. S., and Beck, C. W. (2022). “Comparison of published assessments of biological experimentation as mapped to the ACE-bio competence areas” in Trends in Teaching Experimentation in the Life Sciences. eds. N. J. Pelaez, S. M. Gardner, and T. R. Anderson (Springer, Cham: Contributions from Biology Education Research)
Keywords: course-based research experience (CURE), science education, assessment, intergrated research and education community (iREC), grading
Citation: Hanauer DI, Zhang T, Graham MJ, Adams SD, Ahumada-Santos YP, Alvey RM, Antunes MS, Ayuk MA, Elena Báez-Flores M, Bancroft CT, Bates TC, Bechman MJ, Behr E, Beyer AR, Bortz RL, Bowder DM, Briggs LA, Brown-Kennerly V, Buckholt MA, Bullock SK, Butela KA, Byrum CA, Caruso SM, Chia CP, Chong RA, Chung H-M, Clase KL, Coleman ST, Parks Collins D, Conant SB, Condon BM, Connerly PL, Connors BJ, Cook-Easterwood JE, Crump KE, D’Elia T, Dennis MK, DeVeaux LC, Diacovich L, Duffy I, Edgington NP, Edwards DC, Egwuatu TOG, Eivazova ER, Fallest-Strobl PC, Fillman CL, Findley AM, Fisher E, Fisher MR, Fogarty MP, Freise AC, Frost VJ, Gainey MD, Garcia Costas AM, Garza AA, Gavin HE, Ghittoni R, Gibb B, Golebiewska UP, Grinath AS, Gurney SMR, Hare RF, Heninger SG, Hinz JM, Hughes LE, Jayachandran P, Johnson KC, Johnson AA, Kanther M, Kenna M, Kirkpatrick BL, Klyczek KK, Kohl KP, Kuchka M, LaPeruta AJ, Lee-Soety JY, Lewis LO, Lindberg HM, Madden JA, Markov SA, Mastropaolo MD, Mathur V, McClory SP, Merkhofer EC, Merkle JA, Michael SF, Mitchell JC, Molloy SD, Monti DL, Mussi MA, Nance H, Nieto-Fernandez FE, Nissen JC, Nsa IY, O’Donnell MG, Page ST, Panagakis A, Parra-Unda JR, Pelletier TA, Perez Morales TG, Peters NT, Phuntumart V, Pollenz RS, Preuss ML, Puthoff DP, Raifu MK, Reyna NS, Rinehart CA, Rocheleau JM, Rossier O, Rudner AD, Rueschhoff EE, Ryan A, Saha S, Shaffer CD, Smith MAV, Sprenkle AB, Strong CL, Sunnen CN, Tarbox BP, Temple L, Thoemke KR, Thomas MA, Tobiason DM, Tolsma SS, Garcia JT, Valentine MS, Vazquez E, Ward RE, Ward CM, Ware VC, Warner MH, Washington JM, Westholm DE, Wheaton KA, Wilkes BM, Williams EC, Biederman WH, Cresawn SG, Heller DM, Jacobs-Sera D, Hatfull GF, Asai DJ and Sivanathan V (2023) Models of classroom assessment for course-based research experiences. Front. Educ. 8:1279921. doi: 10.3389/feduc.2023.1279921
Edited by:
Dina Tavares, Polytechnic Institute of Leiria, PortugalReviewed by:
Marina Marchisio, University of Turin, ItalyChristopher W. Beck, Emory University, United States
Copyright © 2023 Hanauer, Zhang, Graham, Adams, Ahumada-Santos, Alvey, Antunes, Ayuk, Elena Báez-Flores, Bancroft, Bates, Bechman, Behr, Beyer, Bortz, Bowder, Briggs, Brown-Kennerly, Buckholt, Bullock, Butela, Byrum, Caruso, Chia, Chong, Chung, Clase, Coleman, Parks Collins, Conant, Condon, Connerly, Connors, Cook-Easterwood, Crump, D'Elia, Dennis, DeVeaux, Diacovich, Duffy, Edgington, Edwards, Egwuatu, Eivazova, Fallest-Strobl, Fillman, Findley, Fisher, Fisher, Fogarty, Freise, Frost, Gainey, Garcia Costas, Garza, Gavin, Ghittoni, Gibb, Golebiewska, Grinath, Gurney, Hare, Heninger, Hinz, Hughes, Jayachandran, Johnson, Johnson, Kanther, Kenna, Kirkpatrick, Klyczek, Kohl, Kuchka, LaPeruta, Lee-Soety, Lewis, Lindberg, Madden, Markov, Mastropaolo, Mathur, McClory, Merkhofer, Merkle, Michael, Mitchell, Molloy, Monti, Mussi, Nance, Nieto-Fernandez, Nissen, Nsa, O'Donnell, Page, Panagakis, Parra-Unda, Pelletier, Perez Morales, Peters, Phuntumart, Pollenz, Preuss, Puthoff, Raifu, Reyna, Rinehart, Rocheleau, Rossier, Rudner, Rueschhoff, Ryan, Saha, Shaffer, Smith, Sprenkle, Strong, Sunnen, Tarbox, Temple, Thoemke, Thomas, Tobiason, Tolsma, Garcia, Valentine, Vazquez, Ward, Ward, Ware, Warner, Washington, Westholm, Wheaton, Wilkes, Williams, Biederman, Cresawn, Heller, Jacobs-Sera, Hatfull, Asai and Sivanathan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: David I. Hanauer, hanauer@pitt.edu; Viknesh Sivanathan, sivanathanv@hhmi.org
†Present address:
Michelle Kanther, Department of Bio-Medical Sciences, Philadelphia College of Osteopathic Medicine, Philadelphia, PA, United States