Skip to main content

ORIGINAL RESEARCH article

Front. Public Health, 11 December 2023
Sec. Infectious Diseases: Epidemiology and Prevention
This article is part of the Research Topic Artificial Intelligence Solutions for Global Health and Disaster Response: Challenges and Opportunities View all 26 articles

Forecasting daily COVID-19 cases with gradient boosted regression trees and other methods: evidence from U.S. cities

\r\nAnindya Sen
Anindya Sen1*Nathaniel T. StevensNathaniel T. Stevens2N. Ken TranN. Ken Tran3Rishav R. AgarwalRishav R. Agarwal4Qihuang ZhangQihuang Zhang5Joel A. Dubin,Joel A. Dubin2,3
  • 1Department of Economics, University of Waterloo, Waterloo, ON, Canada
  • 2Department of Statistics and Actuarial Science, University of Waterloo, Waterloo, ON, Canada
  • 3School of Public Health Sciences, University of Waterloo, Waterloo, ON, Canada
  • 4Cheriton School of Computer Science, University of Waterloo, Waterloo, ON, Canada
  • 5Department of Epidemiology, Biostatistics and Occupational Health, McGill College, Montreal, QC, Canada

Introduction: There is a vast literature on the performance of different short-term forecasting models for country specific COVID-19 cases, but much less research with respect to city level cases. This paper employs daily case counts for 25 Metropolitan Statistical Areas (MSAs) in the U.S. to evaluate the efficacy of a variety of statistical forecasting models with respect to 7 and 28-day ahead predictions.

Methods: This study employed Gradient Boosted Regression Trees (GBRT), Linear Mixed Effects (LME), Susceptible, Infectious, or Recovered (SIR), and Seasonal Autoregressive Integrated Moving Average (SARIMA) models to generate daily forecasts of COVID-19 cases from November 2020 to March 2021.

Results: Consistent with other research that have employed Machine Learning (ML) based methods, we find that Median Absolute Percentage Error (MAPE) values for both 7-day ahead and 28-day ahead predictions from GBRTs are lower than corresponding values from SIR, Linear Mixed Effects (LME), and Seasonal Autoregressive Integrated Moving Average (SARIMA) specifications for the majority of MSAs during November-December 2020 and January 2021. GBRT and SARIMA models do not offer high-quality predictions for February 2021. However, SARIMA generated MAPE values for 28-day ahead predictions are slightly lower than corresponding GBRT estimates for March 2021.

Discussion: The results of this research demonstrate that basic ML models can lead to relatively accurate forecasts at the local level, which is important for resource allocation decisions and epidemiological surveillance by policymakers.

1 Introduction

On May 5, 2023, the World Health Organization (WHO) officially ended the global COVID-19 emergency, referring to increased population immunity, fewer deaths, and reduced pressure on hospitals. The COVID-19 pandemic, which was first declared an international crisis by WHO on January 30, 2020, resulted in severe lockdowns, closure of international borders, devastating economic costs upheaval and the deaths of at least seven million people across the world.1 Hundreds of published and working research papers have attempted to evaluate the efficacy of traditional statistical models and Machine Learning (ML)/Artificial Intelligence (AI) methods in forecasting COVID-19 cases. Accurate short-term forecasts can inform government decision-making in terms of resource allocation to health practitioners and hospitals, as well as in deciding on the magnitude and severity of lockdowns and timing of re-openings.

The focus on ML and AI methods might be explained by the poor performance of Susceptible-Infected-Removed (SIR) models—traditionally used by epidemiologists to predict the spread of infectious diseases—in forecasting daily COVID-19 counts (13). However, many sophisticated statistical methods developed for COVID-19 modeling and forecasting, such as the models from the Institute of Health Metrics and Evaluations, the University of Texas at Austin, and the Los Alamos National Laboratory, have also yielded unsatisfactory results (4). Hence, there is value in identifying alternative models that are relatively easy to implement and interpret, and that are capable of producing accurate predictions. This study evaluates the efficacy of Gradient Boosted Regression Tree (GBRT), Susceptible, Infectious, or Recovered (SIR), Seasonal Autoregressive Integrated Moving Average (SARIMA), and Linear Mixed Effects (LME) models in forecasting daily trends in COVID-19 cases across 25 cities in the U.S. (Albuquerque, Atlanta, Baltimore, Boston, Charlotte, Chicago, Cleveland, Dallas, Denver, Detroit, Houston, Indianapolis, Los Angeles, Louisville, Memphis, Miami, New York, Oklahoma, Phoenix, Pittsburgh, Portland, Sacramento, San Francisco, Seattle, Tampa). We forecast daily COVID-19 case rates one-week and four-weeks ahead over different testing periods. The models chosen are reasonably basic, but this choice is intentional and motivated by the desire to explore the efficacy of simpler models that are relatively easily interpretable and computationally efficient, while also crossing disciplinary boundaries to benchmark the performance of traditional methods employed by researchers across different fields.

The choice of these forecasting models and periods is consistent with other studies. For example, Chumachenko et al. (5) used Random Forest, K-Nearest Neighbors, and Gradient Boosting methods to forecast COVID-19 cases for Germany, Japan, South Korea, and Ukraine with respect to 3, 7, 10, 14, 21, and 30 days. The objective of the study is similar to ours, in terms of assisting public health agencies to identify models that could generate predictions to address various pandemic containment challenges. Krivtsov et al. (6) is a relatively recent example of research based on more complex Neural Network (NN) methods using data for the same countries. Mohammadi and Chumachenko (7) employ ARIMA methods to forecast cases for Ukraine. Other examples of ARIMA based studies at the country level include Dansana et al. (8), Singh et al. (9), and Sahai et al. (10). Devaraj et al. (11) use ARIMA and other deep learning methods. Fang et al. (12) compare the performances for XGBoost and ARIMA specifications using U.S. time-series data and Liu et al. (13) employ ARIMA models to study US national data as well.

Our focus on short-term predictions is guided by the importance of such forecasts for the allocation of local health resources, such as the supply of personal protective equipment (PPE), adequate testing infrastructure, and the availability of hospital care teams. Further, our use of data across several U.S. Metropolitan Statistical Areas (MSAs) or cities is a contribution to the literature as most studies rely on either cross-country, national, state/province, or county level data. While employing aggregated data has benefits, identifying models that are capable of relatively accurate forecasts at the local level can result in more targeted decisions by policymakers. In this respect, we are unaware of any study that has attempted to forecast COVID-19 case counts using a panel of MSAs in the United States. Our study is also a contribution to the literature given that a large fraction of studies, which have attempted to forecast the spread of COVID-19 in the U.S., have not provided any benchmarking of their forecasts against the truth, or stated their limitations (14). Constructing forecasting models at a local level is challenging, given the need to account for unobserved jurisdiction level heterogeneity, and corresponding volatility in daily COVID-19 cases which we observe for several U.S. cities. Such volatility often disappears when data are aggregated across jurisdictions, resulting in a smoother time-series of observations in training datasets, and hence, more accurate predictions. However, such predictions may not be very useful for policymakers interested in epidemic trends within a specific jurisdiction.

A primary objective of our study is to evaluate the performance of traditional SIR models employed by epidemiologists, given findings on the inaccuracy of COVID-19 case forecasts, relative to other methods. A possible explanation behind the poor performance of SIR models is because of the need for properly accounting for relevant geographical characteristics such as the number and distribution of outbreaks, and population size and density (15). The results from SIR modeling are compared to GBRT models intended to evaluate the efficacy of a common ML model. Besides being used by other research, our choice of GBRTs is also motivated by the ease in which models can be implemented by policymakers with limited knowledge of Machine Learning or Artificial Intelligence methods. SARIMA modeling is commonly used for time-series forecasting by economists and is especially useful when the modeled data has pronounced trend and seasonality. LME models are popular with researchers working at the intersection of statistics and health. They enable the researcher to flexibly control for the potentially confounding effects of unit-specific (in this paper, city-specific) heterogeneity through the accommodation of random effects, which themselves accommodate the correlation of within-city repeated measures over time. LME specifications also allow information to be borrowed across geographies, which might result in more accurate predictions, especially for cities that experience considerable within-city variation during the training phase and whose case rates may be spatially correlated with other cities. We do not explore the performance of Neural Network or Long Short-Term Memory (LSTM) models as we restrict our analysis of methods that are relatively straightforward to implement. Finally, given results from other studies [for example, (1620)], which suggest benefits from the use of ML/statistical models in tandem with social mobility/internet data, we downloaded Apple mobility data for each MSA to evaluate their potential in generating more accurate forecasts.

Citing all papers that have employed such methods in forecasting daily COVID-19 case trends is beyond the scope of this study. Focusing on studies that have employed either county or state/province level data, Altieri et al. (21) and Liu et al. (22) construct one- and two-week-ahead forecasts (of case or death counts) using either different types of linear and exponential predictors or Bayesian methods. Chu and Qureshi (1), Chen et al. (2), Stevens et al. (23), and Sen et al. (24) find different autoregressive time series and ML models to be capable of comparable or superior short-term out-of-sample forecasts of daily cases relative to SIR models. In terms of cities, Wathore et al. (25) rely on deep learning models such as LSTM to forecast cases for 8 cities in India, U.S., and Sweden. Their study contains a summary of other LSTM based papers. Devaraj et al. (11) also contains a detailed discussion of deep learning-based papers. Zhang et al. (26) develop a hybrid predictive model of COVID-19 cases based on autoregressive and LSTM models, which they test for 8 counties in California and some countries.

The limited amount of research exploiting variation across cities over time is probably because of the lack of publicly available data. For U.S. cities, we surmount this difficulty by employing county level data collected by the New York Times and constructing corresponding MSA level daily case counts. As a result, we are able to clearly see patterns and differences in COVID-19 cases across some major U.S. metropolitan areas during the first and second waves of the pandemic. Another contribution of our study is that we produce forecasts during time periods that coincide with the peak of the second wave of infections, specifically during November 15 - December 12, 2020, along with subsequent time periods, which saw significant declines in cases counts for many MSAs. This is in contrast to many early published studies that are focused on forecasts of the first wave and Summer of 2020.

Our choice of this time period is also motivated by the considerable volatility in daily case counts observed across cities during November–December 2020, making the forecasting exercise more challenging. We also conduct forecasting exercises for January–March 2021 as well, to evaluate model efficacy during periods of significant declines in daily case counts. The need to ensure that training data for forecasting models capture dynamic changes in the spread of the virus has been noted by other studies (11, 25, 27). Employing data from these time periods is further justified given the rise in population vaccination rates from March 2021 onwards, and the widespread use of home testing kits from late 2021 onwards, which impacts the reliability of official statistics, given the possibility of under-reporting of positive tests to health authorities.

For most MSAs, GBRT and SARIMA models produce forecasts for November 15–December 12, 2020, with lower Median Absolute Percentage Errors (MAPEs) than corresponding one-week ahead predictions produced by LME and SIR models and are consistent with other studies that find SIR models to produce inaccurate forecasts of the incidence and spread of COVID-19. Apple mobility data do not make a significant difference for the forecast accuracy of SARIMA models. With respect to 7-day ahead forecasts, GBRTs produce MAPEs lower than SARIMA models for most MSAs for the November-December 2020, January, and February 2021 testing periods. On the other hand, SARIMA MAPEs are lower for the March 2021 testing period. Likewise, the 28-day ahead forecasts produced by SARIMA models generate lower MAPE values in March 2021. However, for the other months considered, 28-day ahead GBRT forecasts tend to be associated with lower MAPE values.

2 Methods and materials

2.1 Data

Daily COVID-19 case data at the county level were downloaded from the Github repository maintained by the New York Times2 We note that the COVID-19 data maintained by the New York Times and John Hopkins University3 have been widely employed by researchers. Using Federal Information Processing System (FIPS) codes, county-level data were mapped to Metropolitan Statistical Areas (MSAs), which include city cores and adjoining suburbs, to provide daily case counts at the MSA level. We herein use the terms city and MSA interchangeably. Our choice of MSAs was based on investigating COVID-19 trends in the largest cities across the U.S, while ensuring representation across different regions.

Figure 1 shows significant variation in COVID-19 daily cases across MSAs and over time. Some MSAs such as Miami, Phoenix, Oklahoma City, Atlanta, Dallas, Charlotte, Tampa, Houston, San Francisco, and Sacramento had much higher daily case counts per 100,000 of population during July and August of 2020, relative to the first wave in March and April of the same year. In contrast, New York, Boston, and Chicago had much higher case counts per 100,000 of population or per capita daily cases during the first wave. Most cities experienced a decline in COVID-19 cases during September, which was succeeded by a rapid increase during November and December that coincided with intensive campaigning associated with the U.S. Presidential Election. Increases in daily case counts during this time period were succeeded by declines that began sometime during December 2020 or January 2021 for almost all MSAs, and that continued through March 2021.

FIGURE 1
www.frontiersin.org

Figure 1. Standardized daily case counts per 100,000 people in 25 U.S. cities (March 22, 2020–March 31, 2021).

Five other points should be noted. First, some cities such as Atlanta, Boston, Cleveland, and Houston, have peaks in daily case counts that are clearly anomalous with previous and successive trends. This can be traced to errors in data collection and reporting. The most common explanation we were able to find is that spikes are attributable to the reporting of a significant backlog of cases that, for some reason, were not reported correctly earlier.4 Second, most cities seem to experience considerable volatility in daily COVID-19 case counts during November and December 2020. Third, days with zero cases reported are likely due to batch reporting in previous days or non-working days and were changed to 0.5 to prevent our performance metrics from having nonsensical results when dividing by zero, as well as to allow us to model the natural logarithm of the counts (specifically, the log of standardized case counts per 100,000) for the LME and SARIMA models. Fourth, unlike other MSAs, Detroit experienced a significant rise in daily case counts during March 2021. Daily COVID-19 cases also rose modestly during March 2021 in Pittsburgh and Boston. Fifth, across all MSAs, case rates are clearly non-stationary with pronounced trend and seasonality. Effective modeling and forecasting should account for this correlation structure. Table 1 contains means and standard deviations of standardized daily case counts per 100,000 for each MSA for our training period of March 22, 2020–November 14, 2020.

TABLE 1
www.frontiersin.org

Table 1. Sample statistics.

2.2 Methods

We begin by introducing notation that will be used for all the different methods introduced in this section. Let t = 0, 1, 2, …, T represent time, which here we take to be measured in days. Thus, for a given analysis of an MSA's daily case counts, we will observe a maximum of T + 1 time points. Each city is represented by the index i = 1, 2, …, N, where N = 25 in this paper's analyses. The response variable, denoted by Yi,t, is documented COVID-19 case counts per 100,000 people, standardized by the city population size. In particular,

Yi,t=(#new cases in city on day t/Population size of city i)        × 100,000

represents the per capita daily case count (per 100,000 people) in city i at time point t. Note that this standardization is necessary to ensure that case rates are comparable across cities of different sizes. For models where lagged case values are included as predictor variables, the earliest first day used for the response will be constrained by the number of lagged terms; for example, if we include predictors Yt−1 and Yt−2, then the first response in the model will be at t = 2. In our models specified in the following subsections, the first date to be used for Yt is March 22, 2020, since this is the date that we start to see (at least some) non-zero case counts in all 25 cities.

As discussed in Section 1, some studies have used social media or cellular data to model population mobility and forecast the incidence and spread of COVID-19. Therefore, we use three separate social mobility indices released by Apple, that measure walking, driving, and transit use, respectively, for certain geographies. While most studies have relied on Google mobility data, Google does not offer publicly available data at the MSA level. The Apple data are collected from Apple Maps app usage through individual iPhones. The indices show changes in relative volume of directions requests per country/region, sub-region, or city compared to a baseline volume on January 13th, 2020. While we do not claim these indices necessarily capture population-wide mobility trends, using these data at least allows us to evaluate the effects of social mobility on daily COVID-19 cases counts, and assess the usefulness of this information in constructing daily predictions of COVID-19 cases in MSAs. Moreover, the walking mobility indicator is highly correlated with temperature and therefore serves as a useful proxy for weather, which is known to be an important factor in transmission (28).

In terms of quantifying the forecasting performance of different models, we consider the Median Absolute Error (MAE) and Median Absolute Percentage Error (MAPE) of predicted daily case counts for each Metropolitan Statistical Area. The MAE is the median of the absolute value of the difference between actual and predicted daily case counts over the chosen testing period, while the MAPE is the median of the absolute value of the corresponding percentage difference. While MAEs are reported, our comparisons are primarily based on the MAPE as it is a standardized metric interpreted equivalently across all cities. Hence, for a city i, if the actual observed daily case count is Yt and the predicted value from a chosen model is Ŷt, the MAE is the median absolute value of their difference or |Yt − Ŷt|, calculated from daily values over the testing period. The MAPE is therefore the median of |Yt-ŶtYt|×100 over the testing sample. For further sensitivity analysis, we also report the proportion of daily forecasts with an absolute percentage forecast error <20%.

2.2.1 Linear Mixed Effects (LME) models

To assess the statistical importance of Apple mobility indicators, we ran LME regressions using Yi,t (i.e., standardized case count per 100,000 for city i at time j) as the dependent variable and based on data pooled across cities and over days for the training period. As a sensitivity analysis, we also ran the LME regressions for the time period March 22 - August 31, 2020, before the beginning of the steady rise in case counts for most cities. The LME models considered contain the seven-day lag of the walking index, one and two day lags of the dependent variable along with a weekend indicator variable.5 Estimation results are reported in Table 2. First, while these specific results are not reported, we note that lagged dependent variables of a higher order than 3 days were statistically insignificant in these models. Second, one and two-day lags in the dependent variable are statistically significant at the 1% level. Third, the seven-day lagged Apple mobility Walking Index is also statistically significant at the 1% level. These results offer some support to the inclusion of a mobility index in the models used to forecast daily case counts. On the other hand, we note that the inclusion of the Apple mobility index does not significantly improve the fit of the LME models as measured through the AIC and Log-Likelihood, and so its inclusion may not drastically improve a model's forecasting capability.

TABLE 2
www.frontiersin.org

Table 2. Baseline linear mixed effect model summaries (standard errors in parentheses).

In terms of the specific LME model:

log(Yi,t)=β0+b0,i+β1log(Yi,t-1)+β2log(Yi,t-2)+β3WEt                     +εi,t,    (1)

where:

• β0 is the population-level intercept;

b0,i is a city-specific random intercept, with b0,i ~ N(02, σb0);

• β1 is the fixed-effect parameter connected with one-day lagged (log) count for city i;

• β2 is the fixed-effect parameter connected with two-day lagged (log) count for city i;

• β3 is the fixed-effect parameter connected with the weekend indicator, common to all cities. Specifically, if day t falls on a weekend, then WEt = 1, otherwise WEt = 0;

• εi,t is the model error term, with εi = (εi,1, . . ., εi,T)t ~ NT (0, σ2IT), and εi independent of b0, i.

Note we have initially assumed that the model error terms are conditionally independent, with the random intercept being the only term to induce correlation between the repeated measures of daily counts within the same city. We also see that all cities have a common T repeated measures observed. We found that a modified version of (1) provides a better fit to the standardized city-level daily COVID-19 case count data. Specifically, allowing for heterogeneity in both time-varying predictors, i.e., attaching city-specific random effects to each, and properly handling existing heteroskedasticity among the model error terms was also required, leading to:

log(Yi,t)=β0+b0,i+(β1+b1,i)log(Yi,t-1)                    +(β2+b2,i)log(Yi,t-2)+β3WEt+εi,t,    (2)

where the changes relative to (1) are:

b1,i is a city-specific random effect connected to the first-order lagged (log) counts, with b1,i ~ N(02, σb1);

b2,i is a city-specific random effect connected to the second-order lagged (log) counts, with b2,i ~ N(02, σb2); we allow for all three random effects to be correlated, though each are independent of εi;

εi ~ NT (0, Σ), with Σ diagonal, but weighted in such a way to remove the original heteroskedasticity in the model error terms.

From (2), we can forecast one time point ahead (e.g., one day ahead), then use that forecast as the first-order lagged term to forecast one additional day ahead (where the prior first-order lagged term is now the second-order lagged term), and so on, until we have forecasted forward the desired number of days. Note that linear mixed effect models have units borrow strength from other units for a given unit's trajectory prediction. Though not all cities are aligned in time in terms of model dynamics, not nearly as much borrowing of strength from other city's predictions will affect a given city's predictions, due to the extensive collection of repeated measures per city paired with the relatively reasonable within-city variation. Based on the results in Table 2, the Apple Walking Index is employed as a predictor for some specifications. Estimation of LME models and forecasts were conducted using the lme4 package from the R programming language.

2.2.2 Susceptible-Infected-Removed (SIR) model

SIR models are the dominant methodology to model the spread of epidemics; see, for example, Tolles and Luong (29) for further details. The SIR model uses differential equations to describe the dynamic status of an individual switching between three compartments in a population at time t: susceptible (St), infected (It), and removed (Rt) (including recovered and deceased individuals) and is a standard approach employed by epidemiologists to forecast disease spread. Nt is the total population at time t and is identified by Nt = St +It +Rt, where β is the average number of contacts per infectious person per time unit, γ is the transition rate from It to Rt, and Rt includes recovered and deceased individuals. The SIR model is then given by the following ordinary differential equations:

Stt=-βItStN    (3)
Itt=βItStN-γIt    (4)
Rtt=γIt    (5)

where β is the average number of contacts per infectious person per time unit and γ is the transition rate from It to Rt. While the SIR model is based on the modeling of St, It, and Rt, our focus here is the daily infection numbers and aims to conduct predictions such that the prediction errors are minimized. With the daily number of confirmed cases on day t considered as the difference of St and St−1, we calculate the predicted number of daily confirmed cases on day t, denoted Ŷt(β, γ), as follows:

Ŷt(β,γ)=St-1-St=It-It-1+Rt-Rt-1    (6)

for t = 1, …, T, where T represents the end of the study period. We implicitly assume that N, β, and γ are constant over time.

Then, the parameters β and γ can be obtained by minimizing the squared prediction error

PE(β,γ)=t=1T{YtY^t(β,γ)}2    (7)

for β and γ. Yt is the total number of daily cases in a province at time t and is not in per capita terms. We convert the forecasts into per capita terms. Our forecasting procedures model predictions of total daily cases as these are the numbers that are reported and are of interest to policymakers in tracking the incidence and spread of COVID-19. This is consistent with the approach employed by Chen et al. (2). Estimation of the SIR models and construction of forecasts were done using the R programming language. Specifically, we utilized the EpiDynamics and bbmle packages to estimate the parameters of the SIR model.

2.2.3 Box-Jenkins time series modeling

In this paper we also employ the Box-Jenkins class of time series models referred to herein as SARIMA (Seasonal Autoregressive Integrated Moving Average) models and we use the notation log(Yi,t) ~ SARIMA(p, d, q)(P, D, Q)[m], where, again, Yi,t is the standardized case count per 100,000 in city i on day t. Thus, for each city i = 1, 2, . . ., N, we separately fit a model of the form

ϕ(B)Φ(Bm)D(1-B)dlog(Yi,t)=θ(B)Θ(Bm)Dεi,t,    (8)

where B is the backshift operator6 and m is the period of the seasonal component, which here we take to be 7 given the daily data. We also assume that εi,t  i i d~N(0, σ2) and we define ϕ(z), θ(z), Φ(z), and Θ(z) respectively to be the following pth order, qth order, P th order, and

Qth order polynomials:

ϕ(z)=1-ϕ1z-ϕ2z2-...- ϕpzpθ(z)=1+θ1z+θ2z2+...+ θpzqΦ(z)=1-Φ1z-Φ2z2-...- ΦpzPΘ(z)=1+Θ1z+Θ2z2+...+ ΘpzQ

Simple algebra shows that the response for city i on day t, log(Yi,t), is therefore a weighted sum of historical log case counts. Note that different values of the non-seasonal (p, d, q) and seasonal (P, D, Q) orders give rise to different configurations of the model, accounting for different forms of correlation structure in the observed time series. We choose the values of p, d, q, P, D, Q that minimize the corrected Akaike information criterion to ensure the model fits the observed data well. The models themselves are fit using maximum likelihood estimation. As with the LME model, the 7-day lag of the Apple Walking Index is also employed as an exogenous variable in some specifications. These time-series models were estimated employing the forecast package in the R programming language.

2.2.4 Gradient Boosted Regression Trees (GBRTs)

Gradient Boosted Regression Trees are a commonly used ML algorithm based on decision trees to produce forecasts of particular outcomes. The algorithm sequentially tests the predictive power of different trees to reduce forecast errors, until no further improvements can be made. These predictions are then combined through a weighted average of regressions to produce a final prediction. Although there are other more sophisticated ML methods that can be employed to generate predictions, our choice of GBRTs was motivated by their relative simplicity and interpretability, as well as their ease of implementation through software such as R and Python. This makes them an attractive choice for policymakers with limited resources and fairly limited experience with machine learning.

We assume again our sample contains T + 1 observations for each MSA i. In particular, we assume that we observe the response variable Yi,t (standardized case count per 100,000 for city i at time j) and a vector of predictors given by Xi,t for i = 1,2,…, N and t = 0, 1, 2, …., T. The model that forecasts Yi,t based on Xi,t is a weighted additive model of the form

Yi,t=k=1Kαkfk(Xi,t)+εi,t,    (9)

where fk(·) for k = 1,2, …, K are regression trees, αk are weights, and εi,t is an error term. The algorithm estimates both the weights αk for k = 1,2, …, K and fk(·) by sequentially minimizing a penalized differentiable convex loss function of Ytk=1Kαkfk(Xk) with respect to both αk and fk(·) over K boosting iterations for k = 1,2, …, K. These estimates can then be used to produce forecasts of Yi,t. For the Xi,t variables, we employ: the weekend dummy variable; city-specific fixed effects; one- to seven-day lags of the dependent variable; and seven- to ten-day lags of all the Apple mobility variables. GBRTs enable the use of all Apple mobility variables and more lagged dependent variables without concern of collinearity issues, as weak predictor variables are given less weight in constructing forecasts. The XGBoost package in Python was used to estimate the GBRT models and construct predictions. Code for all of the above models and algorithms are available upon request.

3 Results

We begin with predictions that are 1 week or 7-day ahead forecasts for each of the 25 MSAs for November 15–December 12, 2020 based on all the models. Given the poor performance of SIR and LME models, we then evaluate the sensitivity of our findings by constructing forecasts from SARIMA and GBRT models for different time periods in January, February, and March of 2021. Figure 2 visualizes MAPE values and the proportion of daily predictions with a forecast error lower than 20% for daily standardized COVID-19 cases (per 100,000 of population) based on LME, SIR, SARIMA, SARIMA models with the Apple Walking Index (referred to as SARIMAA), and GBRTs for each MSA. Note the values displayed in these plots are also tabulated in Table 3. In Figure 2, Panels A and C report the proportion of daily predictions with a forecast error lower than 20% with respect to 7-day and 28-day forecast performances, respectively. Panels B and D contain corresponding visualizations for MAPE values based on 7-day and 28-day forecast errors. For Panels B and D we seek to identify the methodology that has the lowest average MAPE values across cities. On the other hand, for Panels A and C we are interested in the methodology which has the highest average of the proportion of daily predictions with a forecast error lower than 20%, across cities.

FIGURE 2
www.frontiersin.org

Figure 2. Forecast performance as a function of MSA for 7-day (A, B) and 28-day (C, D) forecasting strategies for November 15-December 12, 2020. (B, D) Quantify forecast performance by MAPE (smaller values are better); (A, C) do so using the proportion of daily predictions with a forecast error lower than 20% (larger values are better).

TABLE 3
www.frontiersin.org

Table 3. 7-day forecasting performance from LME, SIR, and SARIMA models November 15-December 12, 2020.

3.1 LME and SIR 7-day ahead

With respect to 7-day ahead predictions, LME models produce lower MAPEs for 9 MSAs, relative to SIR models. However, the LME MAPEs are also quite high, with Baltimore having the lowest one at 28.63%. SIR models also do not offer consistently accurate forecasts across cities. With the exception of Denver, Louisville, and Portland, SIR MAPE values are always above 25% for such models. Further, SIR models perform poorly on the basis of the proportion of daily forecasts falling within 20% of the actual value, as there are 7 cities for which the value of this measure is zero. It is also important to note the extremely high MAPE values of approximately 90% and 100% for Boston and New York, respectively, as these two cities experienced extremely high per capita daily case counts during the first wave of infections. Further, the MAPE results indicate that even the use of a long training period is not sufficient to enable the SIR models to acknowledge the subsequent downward trend in daily infections and readjust to generate more accurate forecasts for the testing period.

3.2 SARIMA 7-day ahead

Relative to SIR and LME forecasts, SARIMA models produce forecasts with MAPEs that are lower for the majority of cities. SARIMA forecasts yield MAPEs smaller than 20% for four cities and smaller than 25% for six more cities. Another observation is that SARIMA modeling generates predictions that are within tighter bounds for almost each city. Specifically, there is only one MSA (Cleveland) with a MAPE >40% with SARIMA modeling. On the other hand, the number of MSAs with MAPEs >40% based on LME and SIR models are 17 and 15, respectively. Another measure that supports the notion that SARIMA models produce superior forecasts to SIR and LME methods, is the fact that no MSA has a proportion of absolute forecast errors lower than 20% equal to zero, based on SARIMA predictions. Further, for 11 MSAs the proportion of such observations, generated from SARIMA models, is >0.4. Based on these findings, we conclude that SARIMA is a superior forecasting strategy for the vast majority of MSAs. Employing the Apple Walking Index does not make much of a difference in SARIMA forecasts. As we can see in Figure 2, with the exception of a few MSAs, SARIMA MAPEs with and without the Apple Walking Index are comparable. For seventeen cities the difference between MAPEs is <3 percentage points. The proportion of daily forecasts falling within 20% of the actual values are also quite comparable between SARIMA forecasts with and without Apple mobility data.

3.3 GBRT 7-day ahead

For most MSAs, GBRTs produce more accurate forecasts of the methods considered. Specifically, GBRTs generate predictions with lower MAPEs for 17 MSAs, relative to corresponding forecasts from SARIMA models with and without the Apple Mobility Index. Eleven MSAs have GBRT forecasts with MAPEs lower than 20%, with Baltimore, Boston, Charlotte, Chicago, Denver, Indianapolis, Miami, and New York having the lowest values. These results suggest that GBRTs are capable of producing relatively accurate 7-day ahead forecasts even during time periods of steep increases and some volatility in daily case counts. The GBRT results are driven by the lagged dependent variables, and not the Apple mobility variables. Specifically, the feature importance scores of the one- to five-day lagged dependent variables range from 0.01 to 0.29. Among the Apple mobility indicators, only the 8-day lagged Driving Index has a feature score above zero, but which is still low at 0.01. These results are consistent with the satisfactory performance of SARIMA specifications without the Apple Walking Index for some MSAs, as the predictors in these models are lagged dependent variables.

3.4 SARIMA, SARIMAA, and GBRT 28-day ahead

Based on the noticeably poorer performance of the LME and SIR models, the 28-day ahead forecasts (Figures 2C, D) are based on SARIMA, SARIMAA, and GBRT models only. The predictions are constructed from a single sequence of 28-day ahead forecasts (without model updating) for each of the 25 MSAs for November 15–December 12, 2020. Note the values displayed in these plots are also tabulated in Table 4. Here, we note that SARIMA models produce lower MAPEs for a majority of MSAs, as there are 14 MSAs for which SARIMA MAPEs are lower than corresponding GBRT values. However, when the SARIMA MAPEs are larger than the GBRT MAPEs, they are often much larger, and when SARIMA MAPEs are smaller than GBRT MAPEs, they tend only to be marginally smaller. Consequently, across all MSAs the average SARIMA MAPE is higher than the corresponding GBRT value, despite individual SARIMA MAPEs being smaller for a larger number of cities. As such, GBRTs appear to be the best choice for forecasting both 7-day ahead and 28-day ahead daily standardized COVID-19 case counts for November 15–December 12, 2020.

TABLE 4
www.frontiersin.org

Table 4. 28-day forecasting performance from GBRT and SARIMA models with and without Apple Mobility November 15- December 12, 2020.

3.5 SARIMA and GBRT 7-day and 28-day ahead for January, February, and March 2021

A relevant question is whether these findings are robust to the use of other time periods. Figure 3 visualizes the MAPEs associated with 7-day and 28-day ahead SARIMA and GBRT forecasts for the following testing periods: January 1–28 2021; February 1–28 2021; and March 1–28 2021. For the sake of brevity, we only report MAPEs. The proportion of daily forecasts falling within 20% of the actual values for these time periods are available upon request. As discussed, daily case counts dropped significantly for most MSAs during these months, making the evaluation of the predictive abilities an interesting contrast to our previous exercise of constructing forecasts during a period of rising daily COVID-19 cases during November-December 2020. We do not construct forecasts for Oklahoma City and Seattle, given the presence of a significant number of zeros in daily case values that seem anomalous to case counts in other days.

FIGURE 3
www.frontiersin.org

Figure 3. Forecast performance as a function of MSA for January, February, and March 2021. Forecast performance is quantified by MAPE for two different model types and two different forecasting strategies.

The results visualized in Figure 3 (and tabulated in Tables 5, 6) demonstrate that for almost all MSAs, with respect to 7-day ahead forecasts, GBRTs offer superior predictions relative to SARIMA models for the month of January. GBRT MAPEs are lower than 20% for 9 MSAs, with another 8 MSAs having MAPEs ranging from 20 to 25%. Therefore, for a number of MSAs, GBRTs were able to recognize the change in the trend in daily case counts, from increases to a steady decline. Like the November–December 2020 results, GBRTs are also able to produce more accurate 28-day predictions relative to SARIMA models. GBRT MAPEs are lower than 20% for 6 MSAs, with another 7 MSAs having MAPEs between 20% and 25%. In terms of specific MSAs, Albuquerque, Atlanta, Baltimore, Boston, Cleveland, Louisville, Miami, New York, San Francisco, and Tampa all have GBRT-generated 7-day ahead MAPEs lower than 20% for January 2021, with most of the MSAs (with the addition of Chicago) also possessing low GBRT MAPEs for 28-day ahead forecasts. Including Denver and Indianapolis, these are the same MSAs with low GBRT MAPEs during November - December 2020.

TABLE 5
www.frontiersin.org

Table 5. MAPE values—7-day- and 28-day forecasts from SARIMA models January—March 2021.

TABLE 6
www.frontiersin.org

Table 6. MAPE values—7-day- and 28-day forecasts from GBRT models January—March 2021.

Some of the same trends are visible for February forecasts with GBRT 7-day ahead MAPEs lower than corresponding SARIMA values for 15 MSAs. Further, 9 MSAs have GBRT MAPEs lower than 20% and 4 MSAs possess GBRT MAPEs for from 20 to 25%. On the other hand, both GBRTs and SARIMA models produce low-quality forecasts with high 28-day ahead MAPEs for the same February time period; 17 or more cities possessing MAPE values >50%, and both methods result in MAPEs of over 100% for 8 MSAs. These results suggest that both GBRT and SARIMA models were unable to accurately predict the downward trend in daily case rates that occurred in February for most MSAs. While we cannot confirm this, a possible reason for the decline in daily case rates might be the increased availability and uptake in vaccinations during and preceding February 2021. To the best of our knowledge, data on city-level vaccination rates are unavailable and, therefore, cannot be included as predictors to test this hypothesis.

In contrast to the February 2021 results, both GBRT and SARIMA March 2021 MAPEs become much lower for 7-day ahead forecasts and are comparable to previous months. GBRT and SARIMA performance are comparable in the sense that the number of MSAs are roughly split between the two methods in terms of which approach has the lowest MAPE. With both methods, 10 MSAs have 7-day ahead MAPEs lower than 20%, while another 8 MSAs possess MAPEs between 20 and 25%. On the other hand, SARIMA seems to be a superior method for generating 28-day ahead predictions in March 2021. Specifically, SARIMA forecasts yield MAPEs that are either comparable or much lower than corresponding MAPE values from GBRT models for most MSAs.

To summarize, the GBRT and SARIMA models are the most reliable in terms of producing 7-day ahead daily predictions with the lowest MAPE values for November-December 2020. With respect to 28-day ahead forecasts for the same time-period, while SARIMA and GBRT MAPES are comparable for several MSAs, there are MSAs where SARIMA MAPEs are much higher. Using Apple Mobility data does not considerably improve forecasts from SARIMA models. GBRT models similarly outperform SARIMA models for 7-day and 28-day ahead predictions for January, but neither model produces high quality predictions for February 2021, with MAPE values from GBRT models, on average, being slightly lower. For March 2021, 7-day ahead forecasts from GBRT and SARIMA models are comparable, with 28-day ahead SARIMA MAPE values being somewhat lower than corresponding GBRT estimates.

4 Discussion

Findings from previous studies suggest that the standard SIR model used by epidemiologists for disease predictions resulted in inaccurate forecasts for multiple jurisdictions. Further, we are unaware of any other study that has attempted to construct daily forecasts of a panel of U.S. cities. While there is an abundance of research based on country-level data, research attempting to predict daily COVID-19 cases across cities is scarce (11). This is unfortunate, as cities or MSAs have often been the center of significant occurrences and spread of COVID-19 infections. Besides difficulties associated with obtaining city level data, forecasting daily cases is extremely challenging given the considerable heterogeneity in daily case trends across cities. Perhaps even more important are factors such as differences in testing rates, lack of uniformity in data collection protocols, and inaccurate processing of data, that make comparisons of cases across jurisdictions difficult (30).

This study explores the efficacy of SIR, LME, SARIMA, and GBRT models in producing one and four-week ahead predictions. The choice of these models is premised on different advantages. GBRTs are an example of supervised ML methods that have proved to be effective in different cases (12, 24) and can be developed with less knowledge than more sophisticated machine learning methods, such as deep learning. LME models accommodate city-level heterogeneity, enabling the researcher to borrow information across geographies, and hence do not rely exclusively on time-series variation within geographies. SIR models are the conventional workhorses in epidemiology, but do not have the flexibility to accommodate time-specific changes in external factors, such as population mobility, that can plausibly impact spread of infections. SARIMA models are relatively simple to implement and as noted earlier, are useful when the modeled data have pronounced trend and seasonality. These models have been employed by other studies (11, 12).

We evaluate model performance primarily through MAPEs and find that the prediction accuracy associated with LME and SIR forecasts tends to be inferior relative to that of SARIMA and GBRT forecasts. Compared to SARIMA, GBRTs generate 7-day ahead and 28-day forecasts with lower MAPEs for a vast majority of MSAs for most months. While SARIMA produces lower MAPEs for 28-day ahead predictions for March 2021, the good performance of GBRTs more broadly make it a suitable choice for modeling and forecasting daily COVID-19 case counts during time periods of both increase and decrease in infections. From a policy perspective, these results are important as they imply the availability of superior forecasting methods relative to conventional epidemiological methods. The performance of GBRTs should be noted, given that they are relatively straightforward to use through available statistical packages. The performance of these models is comparable to findings from other studies that use more disaggregated data than at the country level. For example, using provincial level data from Canada, Sen et al. (24) find MAPEs from GBRTs ranging from 8 to 30% for 2 week ahead daily forecasts. Zhang et al. (26) obtain Mean Absolute Percentage Errors ranging from 4 to 8% for 8 counties in California, but not for time-periods of 4 weeks or over.

We note the following limitations. It is important to acknowledge that several of the chosen models make strong and potentially restrictive assumptions including, but not limited to, errors that are independently and identically distributed normal random variables. Such assumptions have been evaluated using a battery of residual diagnostic tools. Although many models and many specifications of these models were considered, we present results only for those whose assumptions were reasonably satisfied. To demonstrate this, we include Figures 46 which depict residuals, autocorrelation function and normal quantile-quantile plots of the residuals of the SARIMA model fit to the March 2020-December 2020. As is evident, we do not observe serious violations of stationarity or normality.

FIGURE 4
www.frontiersin.org

Figure 4. Model residuals associated with the SARIMA model fit to March 2020-December 2020 data.

FIGURE 5
www.frontiersin.org

Figure 5. Autocorrelation function (ACF) plots of model residuals associated with the SARIMA model fit to March 2020-December 2020 data.

FIGURE 6
www.frontiersin.org

Figure 6. Normal QQ-plots of model residuals associated with the SARIMA model fit to March 2020-December 2020 data.

5 Conclusion

Along with other papers that have emerged over the past 2 years, the results of this study suggest that GBRTs can also be used for predicting the spread of highly infectious diseases on a daily basis. These findings suggest that the relatively basic ML modeling can lead to vital insights for government resource allocation and decision-making, and result in superior disease surveillance relative to conventional epidemiological methods. Future work will investigate the benefits of employing more complex deep learning and neural network-based methods, which have the trade-off of being more complex and possibly more accurate but also more difficult to interpret, especially for a wider audience of policy practitioners.

Data availability statement

The original contributions presented in the study are publicly available. This data can be found here: https://github.com/kentranz/socialMobilityCOVID/blob/master/data/all.csv.

Author contributions

AS: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Writing—original draft, Writing—review & editing. NS: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Writing—original draft, Writing—review & editing. NT: Data curation, Formal analysis, Investigation, Methodology, Software, Writing—original draft. RA: Data curation, Formal analysis, Investigation, Methodology, Software, Writing—original draft. QZ: Data curation, Formal analysis, Investigation, Methodology, Software, Writing—original draft. JD: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Supervision, Writing—original draft, Writing—review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. Funding for this study was obtained to: NT gratefully acknowledges financial assistance from NSERC Discovery Grant RGPIN-2019-04212. JD acknowledges financial support from NSERC Discovery Grant RGPIN-2020-04382.

Acknowledgments

The authors acknowledge helpful comments from the two referees.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

1. ^Please see https://www.cbc.ca/news/health/canada-who-pandemic-no-longer-emergency (last accessed June 1st 2023) for further details.

2. ^https://github.com/nytimes/covid-19-data

3. ^https://coronavirus.jhu.edu/map.html

4. ^See, for example, https://www.click2houston.com/news/local/2020/09/23/explained-why-nearly-15000-new-coronavirus-cases-were-reported-in-the-houston-area-in-one-day/ and https://www.wsbtv.com/news/local/georgia-records-highest-daily-increase-coronavirus-cases-since-bef7IAY2EPRURABPICVYA3PO7AEJ4/.

5. ^Only the Walking Index was employed in these specifications because of a high correlation with the Driving Index (r = 0.85) and the fact that the Transit Index was consistently insignificant in preliminary regressions that contained all three mobility indicators. More complicated forms of autocorrelation are handled differently in the various models, but generally through the inclusion of lagged case rates as predictor variables.

6. ^B is a notational operator that shifts subscripts t back by 1: BYt = Yt−1. And when raised to the kth power, it shifts subscripts t back by k: BkYt = Ytk.

References

1. Chu B, Qureshi S. Predicting the covid-19 pandemic in Canada and the U.S. Econ Bull. (2020) 40:2565–85.

Google Scholar

2. Chen LP, Zhang Q, Yi GY, He W. Model-based forecasting for Canadian covid-19 data. PLoS ONE. (2021) 16:e0244536. doi: 10.1371/journal.pone.0244536

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Moein S, Nickaeen N, Roointan A, Borhani N, Heidary Z, Javanmard SH, et al. Inefficiency of sir models in forecasting covid-19 epidemic: a case study of isfahan. Sci Rep. (2021) 11:1–9. doi: 10.1038/s41598-021-84055-6

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Ioannidis JP, Cripps S, Tanner MA. Forecasting for covid-19 has failed. Int J Forecast. (2022) 38:423–38. doi: 10.1016/j.ijforecast.2020.08.004

CrossRef Full Text | Google Scholar

5. Chumachenko D, Meniailov I, Bazilevych K, Chumachenko T, Yakovlev S. Investigation of statistical machine learning models for covid-19 epidemic process simulation: random forest, k-nearest neighbors, gradient boosting. Computation. (2022) 10:86. doi: 10.3390/computation10060086

CrossRef Full Text | Google Scholar

6. Krivtsov S, Meniailov I, Bazilevych K, Chumachenko D. Predictive model of covid-19 epidemic process based on neural network. Radioelect Comp Syst. (2022) 4. doi: 10.32620/reks.2022.4.01

CrossRef Full Text | Google Scholar

7. Mohammadi A, Chumachenko D. COVID-19 epidemic process simulation using ARIMA model. In: International Scientific and Technical Conference on Integrated Computer Technologies in Mechanical Engineering-Synergetic Engineering, ICTM 2021 (2022). p. 353–63.

Google Scholar

8. Dansana D, Kumar R, Das Adhikari J, Mohapatra M, Sharma R, Priyadarshini I, et al. Global forecasting confirmed and fatal cases of covid-19 outbreak using autoregressive integrated moving average model. Front Public Health. (2020) 8:580327. doi: 10.3389/fpubh.2020.580327

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Singh S, Sundram BM, Rajendran K, Law KB, Aris T, Ibrahim H, et al. Forecasting daily confirmed COVID-19 cases in Malaysia using ARIMA models. J Infect Dev Ctries. (2020) 14:971–6. doi: 10.3855/jidc.13116

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Sahai AK, Rath N, Sood V, Singh MP. ARIMA modelling and forecasting of covid-19 in top five affected countries. Diabetes Metab Syndr Clin Res Rev. (2020) 14:1419–27. doi: 10.1016/j.dsx.2020.07.042

CrossRef Full Text | Google Scholar

11. Devaraj J, Madurai Elavarasan R, Pugazhendhi R, Shafiullah GM, Ganesan S, Jeysree AK, et al. Forecasting of COVID-19 cases using deep learning models: is it reliable and practically significant? Results Phys. (2021) 21:103817. doi: 10.1016/j.rinp.2021.103817

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Fang ZG, Yang SQ, Lv CX, An SY, Wu W. Application of a data-driven XGBoost model for the prediction of COVID-19 in the USA: a time-series study. BMJ Open. (2022) 12:e056685. doi: 10.1136/bmjopen-2021-056685

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Liu Q, Fung DLX, Lac L, Hu P. A novel matrix profile-guided attention LSTM model for forecasting covid-19 cases in USA. Front Public Health. (2021) 9:741030. doi: 10.3389/fpubh.2021.741030

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Nixon K, Jindal S, Parker F, Reich NG, Ghobadi K, Lee EC, et al. An evaluation of prospective covid-19 modelling studies in the USA: from data to science translation. Lancet Digital Health. (2022) 4:e738–47. doi: 10.1016/S2589-7500(22)00148-0

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Bisin A, Moro A. JUE insight: learning epidemiology by doing: the empirical implications of a spatial-sir model with behavioral responses. J Urban Econ. (2022) 127:103368. doi: 10.1016/j.jue.2021.103368

CrossRef Full Text | Google Scholar

16. Wang H, Yamamoto N. Using a partial differential equation with google mobility data to predict covid-19 in Arizona. Math Biosci Eng. (2020) 17:4891–904. doi: 10.3934/mbe.2020266

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Garcia-Cremades S, Morales-Garcia J, Hernandez-Sanjaime R, Martinez-Espan~a R, Bueno-Crespo A, Hernandez-Orallo E, et al. Improving prediction of covid-19 evolution by fusing epidemiological and mobility data. Sci Rep. (2021) 11:1–16. doi: 10.1038/s41598-021-94696-2

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Guan G, Dery Y, Yechezkel M, Ben-Gal I, Yamin D, Brandeau ML. Early detection of covid-19 outbreaks using human mobility data. PLoS ONE. (2021) 16:1–17. doi: 10.1371/journal.pone.0253865

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Ilin C, Annan-Phan S, Tai XH, Mehra S, Hsiang S, Blumenstock JE. Public mobility data enables covid-19 forecasting and management at local and global scales. Sci Rep. (2021) 11:1–11. doi: 10.1038/s41598-021-92892-8

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Amusa LB, Twinomurinzi H, Okonkwo CW. Modeling covid-19 incidence with Google Trends. Front Res Metrics Anal. (2022) 7:1003972. doi: 10.3389/frma.2022.1003972

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Altieri N, Barter RL, Duncan J, Dwivedi R, Kumbier K, Li X, et al. Curating a covid-19 data repository and forecasting county-level death counts in the United States. Harvard Data Sci Rev. (2020). doi: 10.1162/99608f92.1d4e0dae

CrossRef Full Text | Google Scholar

22. Liu L, Moon HR, Schorfheide F. Panel forecasts of country-level covid-19 infections. J Econ. (2021) 220:2–22. doi: 10.1016/j.jeconom.2020.08.010

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Stevens NT, Sen A, Kiwon F, Morita PP, Steiner SH, Zhang Q. Estimating the effects of non-pharmaceutical interventions (npis) and population mobility on daily covid-19 cases: evidence from Ontario. Can Public Policy. (2022) 48:144–61. doi: 10.3138/cpp.2021-022

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Sen A, Baker JD, Zhang Q, Agarwal RR, Lam JP. Do more stringent policies reduce daily COVID-19 case counts? Evidence from Canadian provinces. Econ Anal Pol. (2023) 78:225–42. doi: 10.1016/j.eap.2023.03.006

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Wathore R, Rawlekar S, Anjum S, Gupta A, Bherwani H, Labhasetwar N, et al. Improving performance of deep learning predictive models for covid-19 by incorporating environmental parameters. Gondwana Res. (2023) 114:69–77. doi: 10.1016/j.gr.2022.03.014

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Zhang Y, Tang S, Yu G. An interpretable hybrid predictive model of covid-19 cases using autoregressive model and LSTM. Sci Rep. (2023) 13:6708. doi: 10.1038/s41598-023-33685-z

CrossRef Full Text | Google Scholar

27. Chimmula VKR, Zhang L. Time series forecasting of COVID-19 transmission in Canada using LSTM networks. Chaos Solitons Fractals. (2020) 135:109864. doi: 10.1016/j.chaos.2020.109864

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Wang J, Tang K, Feng K, Lin X, Lv W, Chen K, et al. Impact of temperature and relative humidity on the transmission of covid-19: a modelling study in China and the United States. BMJ Open. (2021) 11:e043863. doi: 10.1136/bmjopen-2020-043863

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Tolles J, Luong T. Modeling epidemics with compartmental models. JAMA. (2020) 323:2515–6. doi: 10.1001/jama.2020.8420

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Alvarez E, Bielska IA, Hopkins S, Belal AA, Goldstein DM, Slick J, et al. Limitations of COVID-19 testing and case data for evidence-informed health policy and practice. Health Res Pol Syst. (2020) 21. doi: 10.1186/s12961-023-00963-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: daily COVID-19 cases, epidemiological surveillance, Metropolitan Statistical Areas, Gradient Boosted Regression Trees, Seasonal Autoregressive Integrated Moving Average (SARIMA), Susceptible, Infectious, or Recovered (SIR), Linear Mixed Effects

Citation: Sen A, Stevens NT, Tran NK, Agarwal RR, Zhang Q and Dubin JA (2023) Forecasting daily COVID-19 cases with gradient boosted regression trees and other methods: evidence from U.S. cities. Front. Public Health 11:1259410. doi: 10.3389/fpubh.2023.1259410

Received: 20 July 2023; Accepted: 13 November 2023;
Published: 11 December 2023.

Edited by:

Tetyana Chumachenko, Kharkiv National Medical University, Ukraine

Reviewed by:

Héctor Armando Rincón León, Mexican Social Security Institute (IMSS), Mexico
Amit Kumar Banerjee, Indian Institute of Chemical Technology (CSIR), India

Copyright © 2023 Sen, Stevens, Tran, Agarwal, Zhang and Dubin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Anindya Sen, asen@uwaterloo.ca

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.