Skip to main content

ORIGINAL RESEARCH article

Front. Res. Metr. Anal., 11 May 2023
Sec. Research Methods

Addressing the educational challenges of urban poverty: a case for solution-based research

\r\nMacey CartwrightMacey Cartwright1Erin O&#x;CallaghanErin O'Callaghan2Sara StacySara Stacy3Casey HordCasey Hord4Heidi Kloos
Heidi Kloos1*
  • 1Department of Psychology, College of Arts and Sciences, University of Cincinnati, Cincinnati, OH, United States
  • 2Criminology, Law and Justice, University of Illinois at Chicago, Chicago, IL, United States
  • 3Department of Psychology, Michigan State University, East Lansing, MI, United States
  • 4Department of Special Education, College of Education, University of Cincinnati, Cincinnati, OH, United States

Introduction: Math achievement for economically disadvantaged students remains low, despite positive developments in research, pedagogy, and funding. In the current paper, we focused on the research-to-practice divide as possible culprit. Our argument is that urban-poverty schools lack the stability that is necessary to deploy the trusted methodology of hypothesis-testing. Thus, a type of efficacy methodology is needed that could accommodate instability.

Method: We explore the details of such a methodology, building on already existing emancipatory methodologies. Central to the proposed solution-based research (SBR) is a commitment to the learning of participating students. This commitment is supplemented with a strength-and-weaknesses analysis to curtail researcher bias. And it is supplemented with an analysis of idiosyncratic factors to determine generalizability. As proof of concept, we tried out SBR to test the efficacy of an afterschool math program.

Results: We found the SBR produced insights about learning opportunities and barrier that would not be known otherwise. At the same time, we found that hypothesis-testing remains superior in establishing generalizability.

Discussion: Our findings call for further work on how to establish generalizability in inherently unstable settings.

Highlights

- We argue that persistent instability is a pivotal characteristic of urban poverty.

- As initial approximation, a type of efficacy research is proposed, referred to as solution-based research (SBR), to accommodate persistent instability.

- We use SBR to test the efficacy of student-guided math practice after school.

- SBR results show that student-guided math practice is engaging and can lead to learning when the right support is in place.

- Our findings indicate that SBR is necessary in addressing educational disparity.

Introduction

It's the travel, not the road, that gets you there.

Matt Hires

Despite numerous positive developments (Kroeger et al., 2012; Schoenfeld, 2016; Neal et al., 2018; Outhwaite et al., 2018), math achievement remains low for students from marginalized communities. For example, the level of math proficiency reported by the National Assessment of Educational Progress [U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics, and National Assessment of Educational Progress (NAEP), 2019] has remained largely unchanged for children from economically disadvantaged urban communities (e.g., from 18% proficient in 2009 to 17% proficient in 2019). In the current paper, we focus specifically on the research-to-practice divide as a possible culprit for this schism.

A biased research-to-practice divide

To explain how a research-to-practice divide could be responsible for educational disparity, note that educational success depends on deploying pedagogy that is deemed effective. Hypothesis-testing is the gold standard to determine such efficacy, trusted by journal editors, policy makers, and practitioners. At the same time, it is well-known that hypothesis-testing falls short in reaching marginalized communities (e.g., Brown, 1992; Berliner, 2002; Camilli et al., 2006; Dawson et al., 2006; Balazs and Morello-Frosch, 2013; Sandoval, 2013; Tseng and Nutley, 2014). Building on this work, we highlight a mismatch between the assumptions of hypothesis-testing and the characteristics of urban poverty.

To explain the mismatch, note that the pressures of poverty are serious enough that they cannot be addressed by individuals chipping away at them on their own. No individual can address hunger or crime single-handedly. Instead, a concerted effort of all stakeholders is required. Unfortunately, the constant crisis mode of poverty cuts into the resources needed to grow a cohesive network of individuals working together (cf., Camazine et al., 2003; Evans et al., 2010). Individuals are instead forced to respond to pressures on their own. This yields sporadic efforts that are bound to disappoint. In turn, such ratchet of unsuccessful efforts yields persistent instability (see Figure 1 for a schematic of this logic).

FIGURE 1
www.frontiersin.org

Figure 1. Schematic of the argument that poverty creates persistent instability.

The argument of the poverty-pressure-instability link can also be made at the level of school life. Central here is that urban impoverished schools often fail to reach the educational targets, which earns them the public label of “low-performing,” “failing,” “in need of improvement,” or “on probation,” (Fleischman and Heppen, 2009). Such label comes with the stinging pressure to improve educational outcomes. Yet, it does not come with enough resources to build the cohesiveness among individuals working together to alleviate the pressure. Students, school staff, and parents are therefore forced to operate on their own: A teacher might try a new technology, a parent might organize homework help after school, etc. These efforts are short-lived under the pressure of the massive educational shortfall, leading to persistent instability.

The argument of the poverty-pressure-instability link can even be made at the micro-level of a math classroom. Central here is that the distribution of students' math scores is likely to be strikingly wide, spanning several grade levels (see Appendix A for data in support). Thus, a middle-school math teacher might encounter students who are one, two, and even three grade levels behind, as well as students who are proficient. Such large spread of proficiency creates yet another type of pressure, namely to provide individualized instruction. This pressure is again likely to require a concerted effort of stakeholders (i.e., to organize small-group support). And the resources required to grow such coordination are again missing in poverty. The outcome is a cycle of sporadic efforts and inevitable failures, resulting again in instability.

Incidentally, the hypothesis-testing methodology depends heavily on stability. This is because its logic assumes the presence of a theoretical distribution of scores that represent the status quo (i.e., the “Null” distribution). This distribution can only be estimated when there is enough stability (i.e., the unchanging “instruction as usual”). Yet there is no stable status quo in the ever-changing crisis of poverty, as individuals attempt to address the various pressures of student learning. Put differently, there is no meaningful chance probability in urban poverty, the same way there is no chance probability for the perfect storm (Bernal et al., 1995). For this reason, the central assumption of hypothesis-testing is violated in poverty.

There have indeed been efforts to bring efficacy research to unstable settings (Fenwick et al., 2015). Prominent is the so-called action research, a type of efficacy research carried out by practitioners (Stringer, 2008; see also continuous improvement; Kaufman and Zahn, 1993; Park et al., 2013). The idea is that practitioners familiar with idiosyncratic constraints could adjust the research protocol accordingly. Design-based research (DBR) is yet another efficacy methodology designed to function in potentially unstable settings (Wang and Hannafin, 2005; Kelly et al., 2008). It builds flexibility into the research protocol in order to adjust to idiosyncratic constraints of a setting (Barab and Squire, 2004; Anderson and Shattuck, 2012; Coburn et al., 2013).

While these efforts have expanded the reach of efficacy research into marginalized communities (e.g., Design-Based Research Collective, 2003), both action research and DBR nevertheless rely on a hypothesis-testing protocol. For example, action research counts on practitioners to find a pocket of stability to roll out a hypothesis-testing protocol. And DBR assumes that the idiosyncratic constraints, once known, remain stable for the duration of the research. Thus, these methodologies merely postpone the need of stability, rather than escaping it. Our objective is to develop a type of efficacy research that can produce generalizable insights about an inherently unstable setting.

A proposal: efficacy research for unstable settings

As initial approximation, we consider the case-study methodology as guide (Lipka et al., 2005; Baxter and Jack, 2008; Yin, 2009; Karsenty, 2010). This methodology is intended to study complex phenomena—phenomena that are marked by nonlinear fluctuations (cf., Holland, 2014). The basis for case studies is a collaboration between researchers and participants (Lambert, 2013). A similar collaboration is featured in community-based participatory research (CBPR; Israel et al., 1998; Minkler and Wallerstein, 2003; Levine et al., 2005; Ozanne and Anderson, 2010; Wallerstein and Duran, 2010). In CBPR, community members drive the research by having a shared goal between the researchers and community members (see also activism research, advocacy research, transformative research).

Building on these approaches, we propose a methodology that is framed by an educational goal shared between researchers and practitioners. Specifically, the so-called Solution-Based Research (SBR) is organized by a collaborative effort to improve the learning of students who participate in the research. Following the logic of case-study design and CBPR, such commitment to an educational mission can provide the necessary stability to uncover insights in an inherently unstable system. Note, however, that SBR—unlike case-study design and CBPR—seeks to produce efficacy results about a pedagogical intervention.

One could argue that a close involvement on the ground could elicit researcher bias (cf., Malterud, 2001; Mann, 2003). In response, we consulted the methodology of evaluation research as guide. This type of methodology is concerned with the effectiveness of a program without the use of hypothesis-testing. Insights about a program are instead obtained via a thoughtful analysis of the program's strengths and weaknesses (Patton, 1990; Quinn, 2002; Narayanasamy, 2009; Ghazinoory et al., 2011). The idea is that, while researcher insights are subjective, the process of highlighting both strengths and weaknesses of an intervention forces a mindset of discovery. In turn, such mindset creates a balanced approach that could keep biases in check. We propose that SBR employs a strengths-and-weaknesses analysis too.

In addition to curtailing researcher bias, efficacy research also needs a strategy by which to determine whether findings can be generalized. In hypothesis-testing, generalizability is accomplished by estimating the chance probability of descriptive results. Given the absence of such chance distribution, we propose to accomplish generalizability via a systematic analysis of idiosyncratic factors vis-à-vis existing literature. This same idea is used in both case-study research and evaluation research to obtain transferable insights. Thus, there is precedence for the claim that an open-mined consideration can address the question of generalizability.

Taken together, SBR has three central features (see Table 1 for an overview). First, SBR adds to the stability of educational settings by supporting the learning of students. This requires a partnership with the educational setting, namely to identify the needs and resources relevant to the intervention. Second, SBR involves a critical analysis of strengths and weaknesses of the intervention, namely to curtail researcher bias. And third, SBR involves a generalizability analysis to determine the degree to which observed trends might hold up in the larger population. In what follows, we illustrate SBR with a concrete example.

TABLE 1
www.frontiersin.org

Table 1. Overview of the proposed SBR methodology.

SRB in action: an efficacy study on an afterschool math program

Introduction

As example case, we sought to investigate the benefits of math practice, the idea being that students gain competence when they solve math problems at an appropriate difficulty level (Frye et al., 2013; Jansen et al., 2013; Haelermans and Ghysels, 2017). We were specifically interested in a type of math practice that gives students agency about what to practice. Student autonomy is known to act as a powerful motivator, likely to increase engagement (Pink, 2009; León et al., 2015). In order to make such individualized practice feasible, we used technology in combination with adult facilitators (Karsenty, 2010; Bayer et al., 2015; Stacy et al., 2017). Our study sought to explore: To what extent is the proposed student-guided math practice effective?

Method

Table 2 provides an overview of the method used for this study. Importantly, students who participate in SBR are not research participants per se. This is because all of the activities students complete are designed for educational purposes. Thus, SRB method does not have a Participants section.

TABLE 2
www.frontiersin.org

Table 2. Overview of the example SBR method.

Setting

The partnering school was a large urban high school that covered Grades 7–12 (~150 students per grade level). At least 90% of students were classified as economically disadvantaged at the time the study took place. The relevant school metrics were predictably troubling: 84% of students scored below proficient on state-wide math assessments, chronic absenteeism was 26%, and high-school dropout was 27%. For 7th-graders specifically, only 20% of students passed the state-wide math test in the year before the research was carried out.

At the same time, the school had extensive collaborations with local organizations, which has led to numerous initiatives (e.g., STEM exposure, career explorations, support in health and wellbeing). It also attracted funding through the 21st Century Community Learning Centers (21st CCLC) funding mechanism to provide academic enrichment opportunities during out-of-school time (James-Burdumy et al., 2005; Leos-Urbel, 2015; Ward et al., 2015). The partnership with researchers was created to design an afterschool math enrichment program. The decision was made to carry out an afterschool math program for 7th and 8th grade students.

Intervention

The intervention was a student led afterschool math practice program, blending technology and the human element. Regarding technology, we opted for the math practice app IXL (IXL Learning, 2016). This app provides access to a comprehensive library of practice sets from all K-12 Common Core topics, organized by grade level and math topic. Regarding the human element, we opted for college-student volunteers to act as facilitators. There was no requirement for facilitators to be math proficient beyond general college readiness. Their task was to help students choose practice sets that were neither too easy nor too difficult for them. For example, should a student visibly struggle with a practice set, facilitators had to guide the student toward an easier practice set.

Following the requirement of the 21st CCLC funding, the math-practice program was offered after school. Two sessions were offered per week to accommodate the schedule of as many students as possible (one hour per session). An extensive reward system was in place to encourage student attendance (e.g., raffle tickets, tokens to be redeemed, opportunities for students to make up a demerit). There were also information booths at school events to let parents know about this opportunity, as well as reminder calls to parents the night before a session. A meal was available to students prior to each session.

Students accessed IXL on a provided tablet or laptop that connected to the school's Wi-Fi. To encourage a prompt start, a session began with a “warm-up” practice set that earned students a treat. After the warm-up, students were free to decide what practice set to work on. When they solved a problem correctly, an encouraging statement appeared on the app (e.g., “good job”), after which the next math problem was loaded. When students entered an incorrect answer, an explanation appeared and students had to press a button to move onto the next math problem. A “Smart Score” visible on the screen tracked the student's performance within each practice set. This value increased with every correct answer and decreased with every incorrect answer, culminating in a score of 100.

Measuring tools

Field notes. We developed an observation protocol to capture the behavior of students and facilitators during a session. For students, positive indicators were to actively work on IXL or seek help from a facilitator. Negative indicators were to be distracted or frustrated, or to quit the math practice altogether. For facilitators, positive indicators were to be attentive to students' work or provide encouragement to a frustrated or inactive student. Negative indicators were to be distracted (e.g., chatting with other adults) or to give lengthy explanations to students. Field notes also served as an attendance record.

Student math proficiency. We gauged students' math proficiency at the onset of the program. Two subscales of the Woodcock-Johnson test battery (Version IV) were used for this purpose: math fluency and calculation competence. The math fluency subscale is a 3-min test of simple one-digit operations (addition, subtraction, multiplication). The calculation-competence subscale is an untimed test of arithmetic, fractions, algebra, etc. Both subscales return the student's grade equivalence (GE) score. We also gauged students' learning via the analytics obtained by IXL (e.g., duration of practice, type of practice, error reports).

Student attitudes toward math. Two measures were developed to capture students' attitudes toward math, one administered at the onset of the program, and one administered after each session. Appendix B shows the intake survey: It focused on whether students like math (e.g., “Is there something that you like about math?”), their beliefs about math competence (e.g., “How good do you think you are at math?”), and their coping skills when encountering a difficult math problem (e.g., “Do you let someone help you?”). After each session, we also asked: “What did your face look like what it was time for math today?” There were five answer options ranging from “not nervous” to “very, very nervous.”

Feedback from facilitators and school staff. A program-satisfaction survey was administered to facilitators at the end of each semester. Anonymously, facilitators were asked to describe positive and negative aspects of the program. Facilitators were also interviewed after each session to capture possible opportunities and barriers to learning. Teachers and school administrators were interviewed throughout the year about the program and to explore potential solutions to issues we encountered. They were also interviewed at the end of the year to capture perceived strengths and weaknesses of the program. All members of the school staff involved with the program were available for at least three interviews over the course of the year.

Results

With permission from the school's administrators, de-identified data were released for research purposes. These data were then mined for two reasons: to determine the strengths and weaknesses of the student-guided math practice, and to determine the extent to which the observed findings could yield generalizable conclusions. For the current study, we were specifically interested in whether the intervention was engaging and led to learning.

Analysis of strengths and weaknesses

Our approach was to generate an initial list of strengths and weaknesses regarding student engagement and learning. This list was then checked against available evidence and modified iteratively until the set of strengths and weaknesses was balanced in quantity and quality. Below, we describe the finalized list of items (see Table 3).

TABLE 3
www.frontiersin.org

Table 3. Outcome of the strengths-and-weaknesses analysis.

Was self-guided math practice engaging? Our observations showed that students were eager to start their practice. They also were willing to practice math for the duration of the full hour. This applied to a large variety of students, independently of their initial proficiency (see Figure 2 for a distribution of proficiency scores), and independently of their attitudes toward math (27% reported to be “not so good” in math, 68% reported that they dislike math). Even students who attended involuntarily were compelled to practice math. In fact, some students asked to continue practicing after the hour was up. Teachers and visitors to the classroom commented on the positive energy and student focus. To quote a facilitator (survey): “with enough [adult] support, it worked like a well-oiled machine.”

FIGURE 2
www.frontiersin.org

Figure 2. Proficiency scores expressed as proportion of students scoring at a certain grade level (N = 45 7th-graders; N = 48 8th-graders). There was a large variability in proficiency: While some students scored at the 1st- and 2nd-grade proficiency level, other students scored above their actual grade level. Average proficiency in math fluency was close to grade level for 7th grade students (M = 6.97, SD = 2.32), but below grade level for 8th grade students (M = 6.50, SD = 2.79). Average proficiency in calculation competence was substantially below grade level for both 7th (M = 4.55, SD = 1.36) and 8th grade students (M = 5.25, SD = 2.14). There was a significant correlation between fluency and calculation competence, r(24) = 0.70, p < 0.05.

Comments from students and facilitators indicated that each of the program's aspects was well-received (student autonomy, technology, facilitators): Students were appreciative of being given a choice about what to practice, and there was no obvious misuse of this privilege. Students also found the IXL app exceedingly easy to navigate, needing virtually no assistance with using the app correctly. And students connected well with the facilitators, especially when there was consistency in student-facilitator pairing. Students clearly enjoyed the interactions with the facilitators they were familiar with, and they were visibly upset when they could not work with “their” facilitator.

Yet, student engagement did not yield reliable attendance. Considerable efforts were undertaken by school staff to make the afterschool space work. In some cases, coaches even mandated attendance of their players. These efforts led to sporadic increases in attendance (see Figure 3). However, they had little effect on students' voluntary attendance (e.g., only 19% of N = 93 students attended regularly). In fact, they had the unintended consequence of leading to sharp fluctuations in attendance (from 1 to 24 students per session). In turn, the student-facilitator ratio fluctuated as well, ranging from 1:1 to 4:1. This created uncertainty for both students and facilitators (e.g., routines could not solidify; excessive time had to be spent explaining the program to new-comers). Facilitators also reported on being overwhelmed at times, as it was difficult to manage large groups of students.

FIGURE 3
www.frontiersin.org

Figure 3. Attendance of students and facilitators across the year, plotted for each meeting (M).

Did student-guided math practice lead to learning? We found that students clearly benefited from math practice. IXL analytics showed that students who participated in five or more sessions (N = 18) improved in division (80% of students), operations with integers (75% of students), multiplication (50% of students) and equations (27% of students). The comments of students and teachers further substantiated these findings. For example, an 8th-grader explained to a teacher visiting the program, “This is why I was so good at slopes in class today—I've been practicing.” Neither initial math proficiency nor math attitude appeared related to degree of learning.

At the same time, learning was interrupted when students made errors. Observations showed that students were frustrated over errors—even attempting to abandon math practice all together (e.g., switching off the tablet, laying their head on the desk). According to facilitator input, students also failed to learn from their errors: Students showed little interest in the explanations of errors offered by IXL, and they continued to make the same mistakes when facilitators took it upon themselves to explain an error. Switching to an easier practice set was far from straightforward: Observations and student responses indicated that students had little insight about their own gaps in math proficiency, making it difficult to choose appropriately challenging practice sets.

Generalizability analysis

To what extent could our descriptive findings apply generally? Our approach was to first narrow down the list of findings of interest. For each chosen finding, we then mined the literature to determine whether published results align with it. Next, we considered idiosyncratic factors that could account for the finding. Based on this information, we then speculated on whether the finding could be generalized (see Table 4 for an overview of the outcome of this analysis).

TABLE 4
www.frontiersin.org

Table 4. Outcome of the generalizability analysis.

Is student engagement likely to generalize? Descriptively, we found that all three elements of the student-guided math practice were motivating to students: Students appreciated being given a choice, they enjoyed using the IXL app, and they bonded with the facilitators. These findings are strongly aligned with existing literature (Baker et al., 1996; Slavin et al., 2009; George, 2012; Leh and Jitendra, 2013). Regarding idiosyncratic factors, we found that student engagement was high, independently of their attitudes, math proficiency, or the reason for attending the program. Therefore, it is likely that the combination of autonomy, technology, and facilitators is motivating to students more generally.

Is the fluctuation in attendance likely to generalize? Descriptively, we could not establish reliable attendance after school, despite extensive efforts to do so. This finding was surprising, given that the afterschool space is a popular choice for learning (Lauer et al., 2006; Apsler, 2009; Afterschool Alliance, 2014). It is possible, therefore, that our finding stems from idiosyncratic factors, whether of the program, the school culture, or student life (e.g., the stigma of remedial math). At the same time, the fluctuations in student attendance mirror the instability characteristic of urban poverty. Like education itself, afterschool attendance too might require the concerted effort of an underdeveloped network that results in a cacophony of well-meaning but unorganized efforts.

Is student learning likely to generalize? Descriptively, we found that students improved in many of the skills they practiced. This finding is unsurprising given the general understanding that practice leads to learning (Jonides, 2004; Woodward et al., 2012). At the same time, learning was not uniform. For example, practicing division problems led to far more improvement than practicing equation problems. Exploring the published literature, we had difficulty finding research that could shed light on such nuanced effect of math practice. In fact, we were surprised to find relatively little research on math practice, given that math practice is common in students' life (e.g., homework). Thus, we conclude that it remains unclear whether our practice results can generalize.

Is the effect of errors likely to generalize? Descriptively, we found that errors interrupted learning, whether because students got frustrated, because they had difficulty learning from their mistakes, or because they lacked strategic knowledge about what to practice next. There is indeed research on how pressure leads to strong emotional responses, which, in turn, curtails decision-making (Pearman, 2017). At the same time, there are numerous idiosyncratic factors that could account for our findings. For example, facilitators sometimes offered a reward for finishing a practice set, inadvertently making errors relevant. And IXL is not designed to make difficulty levels obvious, thus failing to guide students' choices. Note also that errors were far less disruptive when the student-facilitator ratio was small. Therefore, we conclude that these findings generalize when settings lack the necessary support.

Summary on student-guided math practice

Our findings suggest that student-guided math practice is an effective pedagogical tool to promote math practice among students from urban impoverished communities. We also found evidence for learning, at least for some math topics. At the same time, afterschool attendance fluctuated sharply and student errors often led to frustrations and inefficiencies that interrupted learning. Further research is needed to investigate whether afterschool attendance is a broader concern in urban-poverty schools. Research is also needed to explore ways to support students' math practice in ways that minimizes frustrations and inefficiencies.

General discussion

Our starting point was the argument that education in urban impoverished neighborhoods suffers from a research-to-practice divide because it lacks the stability that is needed for hypothesis-testing. Solution-based research was proposed in response, namely to produce efficacy results in settings marked by persistent instability. SBR involves (1) a commitment to the learning of participating students, (2) an analysis of both strengths and weaknesses of the intervention, and (3) an analysis of idiosyncratic factors vis-à-vis existing literature to determine generalizability. In light of our study, we discuss each of these aspects.

Consider first the commitment to the learning of participating students. We found that this aspect of SBR worked remarkably well to get research off the ground. Teachers and school staff were eager to help, and they made available needed resources, space, and information. The explicit goal of supporting students also generated a lot of good-will to make the program a success, even among individuals who were not directly related to the project (e.g., sports coaches). The flexibility in the research protocol might have empowered individuals to get involved, creating a collaborative atmosphere among teachers, school staff, and parents. Thus, the educational goal of SBR—put in place to accommodate instability—also forged an educational community. Such community building is likely to be at the heart of addressing the pressures of poverty.

Next, consider SBR's critical analysis of the intervention's strengths and weaknesses. Here too, we had encouraging findings. Notably, the exploratory search for positives and negatives of the intervention brought out diverse points of views of researchers and practitioners. These interactions led to a learning experience for individuals who approached math education from very different vantage points. Thus, preconceived notions had to be revised, in effect neutralizing potential biases. A nuanced picture about opportunities and barriers to learning emerged instead. For example, discussions with practitioners helped highlight the plight of students averse to math, as well as the enormous challenge of math teachers working in large-class settings.

At the same time, we encountered some shortcomings of the strengths-and-weaknesses analysis, especially when it comes to deciding on what findings to consider. In the current study, we focused on the intervention's effects on student engagement and learning. However, this was not the only focus we tried out. In fact, we went through several iteration of strengths and weaknesses, examining separate aspects of the program and its outcomes (e.g., effect on student-facilitator relations). Arguably, such unconstrained process slows down the analysis.

Finally, consider SBR's generalizability analysis. Here we had to select pertinent findings, mine the literature for confirmation, and then consider idiosyncratic factors that could be at play. Needless to say, it was unfeasible to carry out systematic literature reviews on issues that arose. It was also impossible to consider all possible idiosyncratic factors without full information about students' lives. Instead, we found ourselves relying on circumstantial information. Thus, we had to enlist our intuition when deriving claims about generalizability. By comparison, neither circumstantial information nor intuitions are needed for hypothesis-testing. One merely has to enter the data and then read off the result of whether the criterion of generalizability is met.

Seeing that a hypothesis-testing protocol remains superior on the question of generalizability, one could argue that impoverished communities could forego their own efficacy research and rely instead on findings obtained from affluent schools. This might be the logic behind the requirement that low-performing schools should employ evidence-based interventions without concern about where the evidence was obtained (e.g., No Child Left Behind, 2002; Every Student Success Act, 2015). Indeed, the efficacy results compiled for practitioners rarely list the demographics of participants [e.g., see U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance, What Works Clearinghouse (2019), edreports.org, achievethecore.org]. Our findings advise against this logic: The educational landscape of urban poverty seems sufficiently unique to warrant its own efficacy research (cf., Anna-Karenina principle of “All unhappy families are different”; Diamond, 1997). For this reason, SBR might be indispensable.

Conclusions

The quote at the top of our paper anticipates the difference between efficacy research (the travel) and the well-trodden path of hypothesis-testing (the road). Under ideal circumstances, efficacy research takes place on the established road of hypothesis-testing. Urban poverty is not such an ideal circumstance, however. Its persistent instability runs counter to the assumptions of hypothesis-testing. In response, we explored a type of efficacy research that can accommodate persistent instability, building on already existing methodologies. Though the proposed solution-based research does not have an established rule book, our findings highlight several reasons why it might be worth continuing the travel. The hope is that such travel might eventually make a road.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by Institutional Review Board of the University of Cincinnati (Protocol # 2014-4138). The patients/participants provided their written informed consent to participate in this study.

Author contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.

Acknowledgments

We thank the school administration and teachers for partnering with us on bringing afterschool math enrichment to their students. We also thank the Center of Community Engagement from the University of Cincinnati for their help with recruiting facilitators. A special thanks to the facilitators who volunteered their time and energy to support students' learning. Members of the University of Cincinnati Children's Cognitive Research Lab aided with program preparation and data entry. Financial assistance for the research was provided in part by university funding.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Afterschool Alliance (2014). Taking a Deeper Dive into Afterschool: Positive Outcomes and Promising Practices. Washington, DC: Afterschool Alliance.

Google Scholar

Anderson, T., and Shattuck, J. (2012). Design-based research: a decade of progress in education research? Educ. Res. 41, 16–25. doi: 10.3102/0013189X11428813

PubMed Abstract | CrossRef Full Text | Google Scholar

Apsler, R. (2009). After-school programs for adolescents: a review of evaluation research. Adolescence 44, 1–19.

PubMed Abstract | Google Scholar

Baker, J. D., Rieg, S. A., and Clendaniel, T. (1996). An investigation of an after school math tutoring program: university tutors + elementary students = a successful partnership. Education 127, 287–293.

Google Scholar

Balazs, C. L., and Morello-Frosch, R. (2013). The three Rs: how community-based participatory research strengthens the rigor, relevance, and reach of science. Environ. Just. 6, 9–16. doi: 10.1089/env.2012.0017

PubMed Abstract | CrossRef Full Text | Google Scholar

Barab, S., and Squire, K. (2004). Design-based research: putting a stake in the ground. J. Learn. Sci. 13, 1–14. doi: 10.1207/s15327809jls1301_1

CrossRef Full Text | Google Scholar

Baxter, P., and Jack, S. (2008). Qualitative case study methodology: study design and implementation for novice researchers. Qual. Rep. 13, 544–559. doi: 10.46743/2160-3715/2008.1573

CrossRef Full Text | Google Scholar

Bayer, A., Grossman, J. B., and DuBois, D. L. (2015). Using volunteer mentors to improve the academic outcomes of underserved students: the role of relationships. J. Community Psychol. 43, 408–429. doi: 10.1002/jcop.21693

CrossRef Full Text | Google Scholar

Berliner, D. (2002). Comment: educational research: the hardest science of all. Educ. Res. 31, 18–20. doi: 10.3102/0013189X031008018

CrossRef Full Text | Google Scholar

Bernal, G., Bonilla, J., and Bellido, C. (1995). Ecological validity and cultural sensitivity for outcome research: issues for the cultural adaptation and development of psychosocial treatments with Hispanics. J. Abnorm. Child Psychol. 23, 67–82. doi: 10.1007/BF01447045

PubMed Abstract | CrossRef Full Text | Google Scholar

Brown, A. L. (1992). Design experiments: theoretical and methodological challenges in creating complex interventions in classroom settings. J Learn. Sci. 2, 141–178. doi: 10.1207/s15327809jls0202_2

CrossRef Full Text | Google Scholar

Camazine, S., Deneubourg, J. L., Franks, N. R., Sneyd, J., Bonabeau, E., and Theraula, G. (2003). Self-Organization in Biological Systems. Princeton, NJ: Princeton University Press.

Camilli, G., Elmore, P. B., and Green, J. L. (Eds) (2006). Handbook of Complementary Methods for Research in Education. Mahwah, NJ: Lawrence Erlbaum Associates.

Coburn, C. E., Penuel, W., and Geil, K. E. (2013). Research-practice Partnerships: A Strategy for Leveraging Research for Educational Improvement in School Districts. New York, NY: William T. Grant Foundation.

Google Scholar

Dawson, T. L., Fischer, K. W., and Stein, Z. (2006). Reconsidering qualitative and quantitative research approaches: a cognitive developmental perspective. New Ideas Psychol. 24, 229–239. doi: 10.1016/j.newideapsych.2006.10.001

CrossRef Full Text | Google Scholar

Design-Based Research Collective (2003). Design-based research: an emerging paradigm for educational inquiry. Educ. Res. 32, 5–8. doi: 10.3102/0013189X032001005

CrossRef Full Text | Google Scholar

Diamond, J. M. (1997). Guns, Germs, and Steel. New York, NY: W.W. Norton.

Evans, G. W., Eckenrode, J., and Marcynyszyn, L. A. (2010). “Chaos and the macrosetting: the role of poverty and socioeconomic status,” in Chaos and its Influence on Children's Development: An Ecological Perspective, eds G. W. Evans, and T. D. Wachs (Arlington, VA: American Psychological Association), 225–238. doi: 10.1037/12057-014

CrossRef Full Text | Google Scholar

Every Student Success Act (2015). Every Student Succeeds (ESSA) Act of 2015. Pub. L. No. 114-95 U.S.C. Available online at: congress.gov/114/plaws/publ95/PLAW-114publ95.pdf

Fenwick, T., Edwards, R., and Sawchuk, P. (2015). Emerging Approaches to Educational Research: Tracing the Socio-material. New York, NY: Routledge. doi: 10.4324/9780203817582

CrossRef Full Text | Google Scholar

Fleischman, S., and Heppen, J. (2009). Improving low-performing high schools: searching for evidence of promise. Future Child. 19, 105–133. doi: 10.1353/foc.0.0021

PubMed Abstract | CrossRef Full Text | Google Scholar

Frye, D., Baroody, A. J., Burchinal, M., Carver, S. M., Jordan, N. C., and McDowell, J. (2013). Teaching Math to Young Children. Educator's Practice Guide. What Works Clearinghouse. NCEE 2014-4005. What Works Clearinghouse. Available online at: https://ies.ed.gov/ncee/wwc/Docs/practiceguide/early_math_pg_111313.pdf

George, M. (2012). Autonomy and motivation in remedial mathematics. Primus 22, 255–264. doi: 10.1080/10511970.2010.497958

PubMed Abstract | CrossRef Full Text | Google Scholar

Ghazinoory, S., Abdi, M., and Azadegan-Mehr, M. (2011). SWOT methodology: a state-of-the-art review for the past, a framework for the future. J. Bus. Econ. Manag. 12, 24–48. doi: 10.3846/16111699.2011.555358

CrossRef Full Text | Google Scholar

Haelermans, C., and Ghysels, J. (2017). The effect of individualized digital practice at home on math skills-Evidence from a two-stage experiment on whether and why it works. Comput. Educ. 113, 119–134. doi: 10.1016/j.compedu.2017.05.010

CrossRef Full Text | Google Scholar

Holland, J. H. (2014). Complexity: A Very Short Introduction. Oxford: OUP Oxford. doi: 10.1093/actrade/9780199662548.001.0001

CrossRef Full Text | Google Scholar

Israel, B. A., Schulz, A. J., Parker, E. A., and Becker, A. B. (1998). Review of community-based participatory research: assessing partnership approaches to improve public health. Annu. Rev. Public Health 19, 173–202. doi: 10.1146/annurev.publhealth.19.1.173

PubMed Abstract | CrossRef Full Text | Google Scholar

IXL Learning (2016). Available online at: www.ixl.com

James-Burdumy, S., Dynarski, M., Moore, M., Deke, J., Mansfield, W., Pistorino, C., et al. (2005). When Schools Stay Open Late: The National Evaluation of the 21st Century Community Learning Centers Program. Final Report. Washington, DC: US Department of Education.

Jansen, B. R., De Lange, E., and Van der Molen, M. J. (2013). Math practice and its influence on math skills and executive functions in adolescents with mild to borderline intellectual disability. Res. Dev. Disabil. 34, 1815–1824. doi: 10.1016/j.ridd.2013.02.022

PubMed Abstract | CrossRef Full Text | Google Scholar

Jonides, J. (2004). How does practice makes perfect? Nat. Neurosci. 7, 10–11. doi: 10.1038/nn0104-10

PubMed Abstract | CrossRef Full Text | Google Scholar

Karsenty, R. (2010). Nonprofessional mathematics tutoring for low-achieving students in secondary schools: a case study. Educ. Stud. Math. 74, 1–21. doi: 10.1007/s10649-009-9223-z

CrossRef Full Text | Google Scholar

Kaufman, R., and Zahn, D. (1993). Quality Management Plus: The Continuous Improvement of Education. Newbury Park, CA: Corwin Press, Inc.

Kelly, A. E., Baek, J. Y., Lesh, R. A., and Bannan-Ritland, B. (2008). “Enabling innovations in education and systematizing their impact,” in Handbook of Design Research Methods in Education: Innovations in Science, Technology, Engineering, and Mathematics Learning and Teaching, eds A. E. Kelly, R. A. Lesh, and J. Y. Baek (New York, NY: Routledge), 3–18.

Google Scholar

Kroeger, L. A., Brown, R. D., and O'Brien, B. A. (2012). Connecting neuroscience, cognitive, and educational theories and research to practice: a review of mathematics intervention programs. Special issue: neuroscience perspectives on early development and education. Early Educ. Dev. 23, 37–58 doi: 10.1080/10409289.2012.617289

CrossRef Full Text | Google Scholar

Lambert, R. (2013). Constructing and resisting disability in mathematics classrooms: a case study exploring the impact of different pedagogies. Educ. Stud. Math. 89, 1–18. doi: 10.1007/s10649-014-9587-6

CrossRef Full Text | Google Scholar

Lauer, P. A., Akiba, M., Wilkerson, S. B., Apthorp, H. S., Snow, D., and Martin-Glenn, M. L. (2006). Out-of-school-time programs: a meta-analysis of effects for at-risk students. Rev. Educ. Res. 76, 275–313. doi: 10.3102/00346543076002275

CrossRef Full Text | Google Scholar

Leh, J. M., and Jitendra, A. K. (2013). Effects of computer-mediated versus teacher-mediated instruction on the mathematical word problem-solving performance of third-grade students with mathematical difficulties. Learn. Disabil. Q. 36, 68–79. doi: 10.1177/0731948712461447

CrossRef Full Text | Google Scholar

León, J., Núñez, J. L., and Liew, J. (2015). Self-determination and STEM education: effects of autonomy, motivation, and self-regulated learning on high school math achievement. Learn. Individ. Differ. 43, 156–163. doi: 10.1016/j.lindif.2015.08.017

CrossRef Full Text | Google Scholar

Leos-Urbel, J. (2015). What works after school? The relationship between after-school program quality, program attendance, and academic outcomes. Youth Soc. 47, 684–706. doi: 10.1177/0044118X13513478

PubMed Abstract | CrossRef Full Text | Google Scholar

Levine, M., Perkins, D. D., and Perkins, D. V. (2005). Principles of Community Psychology: Perspectives and Applications. New York, NY: Oxford University Press.

Google Scholar

Lipka, J., Hogan, M. P., Webster, J. P., Yanez, E., Adams, B., Clark, S., et al. (2005). Math in a cultural context: two case studies of a successful culturally based math project. Anthropol. Educ. Q. 6, 367–385. doi: 10.1525/aeq.2005.36.4.367

CrossRef Full Text | Google Scholar

Malterud, K. (2001). Qualitative research: standards, challenges, and guidelines. Lancet 358, 483–488. doi: 10.1016/S0140-6736(01)05627-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Mann, C. J. (2003). Observational research methods. Research approach II: cohort, cross sectional, and case-control studies. Emerg. Med. J. 20, 54–60. doi: 10.1136/emj.20.1.54

PubMed Abstract | CrossRef Full Text | Google Scholar

Minkler, M., and Wallerstein, N. (2003). “Part one: introduction to community-based participatory research,” Community-based Participatory Research for Health, eds M. Minkler, and N. Wallerstein (San Francisco, CA: Jossey-Bass), 5–24.

Narayanasamy, N. (2009). Participatory Rural Appraisal: Principles, Methods and Application. New Delhi: SAGE Publications India. doi: 10.4135/9788132108382

PubMed Abstract | CrossRef Full Text | Google Scholar

Neal, J. W., Neal, Z. P., Lawlor, J. A., Mills, K. J., and McAlindon, K. (2018). What makes research useful for public school educators? Adm. Policy Ment. Health Ment. Health Serv. Res. 45, 432–446. doi: 10.1007/s10488-017-0834-x

PubMed Abstract | CrossRef Full Text | Google Scholar

No Child Left Behind (2002). No Child Left Behind (NCLB) Act of 2001, Pub. L. No. 107–110, § 101, Stat. 1425. Available online at: https://www.congress.gov/107/plaws/publ110/PLAW-107publ110.pdf

Outhwaite, L. A., Faulder, M., Gulliford, A., and Pitchford, N. J. (2018). Raising early achievement in math with interactive apps: a randomized control trial. J. Educ. Psychol. 111, 284. doi: 10.1037/edu0000286

PubMed Abstract | CrossRef Full Text | Google Scholar

Ozanne, J. L., and Anderson, L. (2010). Community action research. J. Public Policy Mark. 29, 123–137. doi: 10.1509/jppm.29.1.123

CrossRef Full Text | Google Scholar

Park, S., Hironaka, S., Carver, P., and Nordstrum, L. (2013). Continuous Improvement in Education. Advancing Teaching–improving Learning. White Paper. Princeton, NJ: Carnegie Foundation for the Advancement of Teaching.

Google Scholar

Patton, M. Q. (1990). Qualitative Evaluation and Research Methods, 2nd ed. Newbury Park, CA: Sage Publications, Inc.

Google Scholar

Pearman, F. A. (2017). The effect of neighborhood poverty on math achievement: evidence from a value-added design. Educ. Urban Soc. 51, 289–307. doi: 10.1177/0013124517715066

CrossRef Full Text | Google Scholar

Pink, D. (2009). Drive: The Surprising Truth about What Motivates Us. New York, NY: Riverhead Books.

Google Scholar

Quinn, P. M. (2002). Qualitative Research and Evaluation Methods. Newbury Park, CA: Sage Publications Inc.

Google Scholar

Sandoval, C. (2013). Methodology of the Oppressed (Vol. 18). Minneapolis, MN: University of Minnesota Press.

Google Scholar

Schoenfeld, A. H. (2016). Research in mathematics education. Rev. Res. Educ. 40, 497–528. doi: 10.3102/0091732X16658650

CrossRef Full Text | Google Scholar

Slavin, R. E., Lake, C., and Groff, C. (2009). Effective programs in middle and high school mathematics: a best-evidence synthesis. Rev. Educ. Res., 79, 839–911. doi: 10.3102/0034654308330968

CrossRef Full Text | Google Scholar

Stacy, S. T., Cartwright, M., Arwood, Z., Canfield, J. P., and Kloos, H. (2017). Addressing the math-practice gap in elementary school: are tablets a feasible tool for informal math practice? Front. Psychol. 8, 179. doi: 10.3389/fpsyg.2017.00179

PubMed Abstract | CrossRef Full Text | Google Scholar

Stringer, E. T. (2008). Action Research in Education. Upper Saddle River, NJ: Pearson Prentice Hall.

Google Scholar

Tseng, V., and Nutley, S. (2014). “Building the infrastructure to improve the use and usefulness of research in education,” in Using Research Evidence in Education: From the Schoolhouse Door to Capitol Hill, eds A. J. Daly, and K. S. Finnigan (New York, NY: Springer), 163–175. doi: 10.1007/978-3-319-04690-7_11

CrossRef Full Text | Google Scholar

U.S. Department of Education Institute of Education Sciences National Center for Education Evaluation Regional Assistance. (2019). What Works Clearinghouse. Available online at: ies.ed.gov/ncee/wwc/

Google Scholar

U.S. Department of Education Institute of Education Sciences, National Center for Education Statistics, and National Assessment of Educational Progress (NAEP). (2019). Mathematics Assessments.

Google Scholar

Wallerstein, N., and Duran, B. (2010). Community-based participatory research contributions to intervention research: the intersection of science and practice to improve health equity. Am. J. Public Health 100(S1), S40–S46. doi: 10.2105/AJPH.2009.184036

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, F., and Hannafin, M. J. (2005). Design-based research and technology-enhanced learning environments. Educ. Technol. Res. Dev. 53, 5–23. doi: 10.1007/BF02504682

CrossRef Full Text | Google Scholar

Ward, C., Gibbs, B. G., Buttars, R., Gaither, P. G., and Burraston, B. (2015). Assessing the 21st century after-school program and the educational gains of LEP participants: a contextual approach. J. Educ. Stud. Placed Risk. 20, 312–335. doi: 10.1080/10824669.2015.1050278

CrossRef Full Text | Google Scholar

Woodward, J., Beckmann, S., Driscoll, M., Franke, M., Herzig, P., Jitendra, A., et al. (2012). Improving Mathematical Problem Solving in Grades 4 through 8. IES Practice Guide. NCEE 2012-4055. What Works Clearinghouse.

Google Scholar

Yin, R. K. (2009). Case Study Research, Vol. 4. Thousand Oaks, CA: Sage.

Google Scholar

Appendix A

TABLE A1
www.frontiersin.org

Table A1. Math performance of 4th- and 8th-grade students in 2017, reported in NAEP, 2019.

Appendix B

Math intake survey for students

Attitudes toward math:

1. Is there something that you like about math? [If yes] What do you like about math?

2. Is there something that you don't like about math? [If yes] What is it you don't like about math?

yes

3. What does your face look like when it is time for math?

4. What does your face look like when math is over?

5. What would your face look like if you were to find out you never had to do math again?

6. Do you think you would like to have a job that has a lot of math? [If yes] what would you like about a job that involves a lot of math? It's OK if you are not sure. Just say “I'm not sure.” [If no] what would you not like about a job that involves a lot of math?

Perceived competence:

1. How many people do you think are good at math? Are all people good at math, many people, some people, or only a few people?

2. How good do you think you are at math? Are you very good, good, okay, or not so good?

3. How good do you think boys are at math? Are they very good, good, okay, or not so good?

4. How good do you think girls are at math? Are they very good, good, okay, or not so good?

Coping strategies:

Imagine you have a difficult math problem on homework. Did that ever happen to you? What do you remember about it? Can you tell me about it? What do you do when you have a problem like what you just told me about?

a. Do you say to yourself: “it isn't so serious”?

b. Do you do something to take your mind off it?

c. Do you wonder what to do?

d. Do you say to yourself: “I know I can solve the problem”?

e. Do you ask somebody what to do?

f. Do you try to get out of having to do it?

g. Do you keep worrying and thinking about the problem?

h. Do you want to give up?

i. Do you get grumpy?

j. Do you make a plan to fix the problem?

k. Do you let someone help you?

l. Do you try to figure it out and solve the problem?

m. Do you get mad at the teacher?

n. Do you pretend you don't care?

o. Do you say to yourself: “I know I can fix this”?

Keywords: out-of-school time, student autonomy, tutors, community-based participatory research, action research, design-based research

Citation: Cartwright M, O'Callaghan E, Stacy S, Hord C and Kloos H (2023) Addressing the educational challenges of urban poverty: a case for solution-based research. Front. Res. Metr. Anal. 8:981837. doi: 10.3389/frma.2023.981837

Received: 03 September 2022; Accepted: 27 March 2023;
Published: 11 May 2023.

Edited by:

Surapaneni Krishna Mohan, Panimalar Medical College Hospital and Research Institute, India

Reviewed by:

María-Antonia Ovalle-Perandones, Complutense University of Madrid, Spain
Christos Tryfonopoulos, University of the Peloponnese, Greece

Copyright © 2023 Cartwright, O'Callaghan, Stacy, Hord and Kloos. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Heidi Kloos, heidi.kloos@uc.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.