Comparison of statistical methods used to meta-analyse results from interrupted time series studies: an empirical study
| dc.contributor.author | Korevaar, Elizabeth | |
| dc.contributor.author | Turner, Simon L. | |
| dc.contributor.author | Forbes, Andrew B. | |
| dc.contributor.author | Karahalios, Amalia | |
| dc.contributor.author | Taljaard, Monica | |
| dc.contributor.author | McKenzie, Joanne E. | |
| dc.date.accessioned | 2024-02-13T04:21:19Z | |
| dc.date.available | 2024-02-13T04:21:19Z | |
| dc.date.issued | 2024-02-10 | |
| dc.date.updated | 2024-02-13T04:21:19Z | |
| dc.description.abstract | Abstract Background The Interrupted Time Series (ITS) is a robust design for evaluating public health and policy interventions or exposures when randomisation may be infeasible. Several statistical methods are available for the analysis and meta-analysis of ITS studies. We sought to empirically compare available methods when applied to real-world ITS data. Methods We sourced ITS data from published meta-analyses to create an online data repository. Each dataset was re-analysed using two ITS estimation methods. The level- and slope-change effect estimates (and standard errors) were calculated and combined using fixed-effect and four random-effects meta-analysis methods. We examined differences in meta-analytic level- and slope-change estimates, their 95% confidence intervals, p-values, and estimates of heterogeneity across the statistical methods. Results Of 40 eligible meta-analyses, data from 17 meta-analyses including 282 ITS studies were obtained (predominantly investigating the effects of public health interruptions (88%)) and analysed. We found that on average, the meta-analytic effect estimates, their standard errors and between-study variances were not sensitive to meta-analysis method choice, irrespective of the ITS analysis method. However, across ITS analysis methods, for any given meta-analysis, there could be small to moderate differences in meta-analytic effect estimates, and important differences in the meta-analytic standard errors. Furthermore, the confidence interval widths and p-values for the meta-analytic effect estimates varied depending on the choice of confidence interval method and ITS analysis method. Conclusions Our empirical study showed that meta-analysis effect estimates, their standard errors, confidence interval widths and p-values can be affected by statistical method choice. These differences may importantly impact interpretations and conclusions of a meta-analysis and suggest that the statistical methods are not interchangeable in practice. | |
| dc.identifier.citation | BMC Medical Research Methodology. 2024 Feb 10;24(1):31 | |
| dc.identifier.uri | https://doi.org/10.1186/s12874-024-02147-z | |
| dc.identifier.uri | https://doi.org/10.20381/ruor-30149 | |
| dc.identifier.uri | http://hdl.handle.net/10393/45945 | |
| dc.language.rfc3066 | en | |
| dc.rights.holder | The Author(s) | |
| dc.title | Comparison of statistical methods used to meta-analyse results from interrupted time series studies: an empirical study | |
| dc.type | Journal Article |
