Predicting the Next Outbreak: Data Synthesis Boosts Forecasting

Author: Denis Avetisyan


New research shows that combining artificially generated data with genetic insights significantly improves the accuracy of infectious disease predictions.

A classifier robustly distinguishes between synthetic and real respiratory data, yet surprisingly misclassifies a substantial portion of COVID-19 cases as synthetic, a phenomenon visualized through probability distributions and a UMAP arrangement revealing overlap between COVID-19 and synthetic data in feature space.
A classifier robustly distinguishes between synthetic and real respiratory data, yet surprisingly misclassifies a substantial portion of COVID-19 cases as synthetic, a phenomenon visualized through probability distributions and a UMAP arrangement revealing overlap between COVID-19 and synthetic data in feature space.

Deep learning models leveraging synthetic data and variant attribution achieve state-of-the-art performance in time series forecasting of epidemic outbreaks, even with limited historical data.

Accurate and timely forecasting remains a persistent challenge in the face of emerging infectious diseases, particularly when historical data is scarce. This research, ‘Leveraging Synthetic and Genetic Data to Improve Epidemic Forecasting’, investigates the potential of combining deep learning with synthetic datasets and genomic information to enhance predictive capabilities. We demonstrate that models trained with these combined data sources outperform standard forecasting approaches, including a leading ensemble, and a simple persistence model, achieving state-of-the-art accuracy in predicting COVID-19 case numbers. Could this blueprint for integrating underutilized data streams offer a critical advantage in preparing for – and mitigating – future pandemics?


Navigating the Complexities of Pandemic Forecasting

Effective public health responses to COVID-19 heavily rely on the ability to accurately forecast the spread of the virus, yet this proves remarkably difficult. Traditional epidemiological models, while foundational, often grapple with the inherent complexities of both viral evolution and human behavior. The virus’s rapid mutation rate introduces constant change to its transmissibility and severity, rendering static model assumptions quickly obsolete. Simultaneously, unpredictable shifts in human actions – encompassing factors like mask usage, social distancing, and vaccination rates – introduce further instability into projections. These intertwined challenges mean that even sophisticated models struggle to capture the full dynamic of an outbreak, highlighting the need for innovative forecasting approaches that can adapt to rapidly changing conditions and incorporate a wider range of influencing factors.

The ongoing evolution of viruses like SARS-CoV-2 presents a fundamental challenge to pandemic prediction. Antigenic mutation, the process by which viruses alter their surface proteins, allows them to evade existing immunity – whether from prior infection or vaccination – and potentially increase transmissibility. This constant shift in viral characteristics disrupts the assumptions underlying many forecasting models, which often rely on relatively stable parameters. Consequently, predictions made using older data may rapidly become inaccurate as new variants emerge and gain dominance. The speed at which these variants can arise, coupled with the time needed to characterize their properties and incorporate them into models, creates a persistent lag in predictive capability, making short-term forecasts particularly unreliable and necessitating continuous adaptation of predictive strategies.

Current pandemic forecasting struggles largely from a fragmentation of information; predictive models frequently operate on limited datasets, failing to fully leverage the wealth of available data. A comprehensive understanding of outbreak dynamics necessitates the integration of diverse streams – genomic surveillance tracking viral evolution, real-world data capturing transmission rates and clinical outcomes, and even behavioral datasets reflecting public response to interventions. The inability to synthesize these disparate sources creates a significant blind spot, hindering the development of truly robust and nuanced forecasts capable of anticipating shifts in viral landscapes and informing effective public health strategies. Effectively combining these data types requires advanced computational approaches and a shift away from relying on single-source predictions towards more holistic, data-driven modeling.

Analysis of COVID-19 data from Alabama and California reveals the weekly total cases, the proportion of viral variants identified through genomic sequencing, and the estimated number of cases attributable to each variant, displayed using a square-root scale to enhance visualization of low-frequency variants.
Analysis of COVID-19 data from Alabama and California reveals the weekly total cases, the proportion of viral variants identified through genomic sequencing, and the estimated number of cases attributable to each variant, displayed using a square-root scale to enhance visualization of low-frequency variants.

Augmenting Reality: Generating Robust Training Data

MutAntiGen is an agent-based model employed to create synthetic datasets for outbreak scenarios, mitigating the limitations imposed by insufficient labeled real-world data. This model simulates the spread of infectious diseases through a population of individual agents, each with defined characteristics and behaviors. By varying parameters such as transmission rates, population density, and intervention strategies within MutAntiGen, a diverse range of outbreak scenarios can be generated. The resulting synthetic data includes detailed information on infection timelines, geographic spread, and individual-level characteristics, providing a controlled and scalable source of training data for forecasting models. This approach allows for comprehensive scenario testing, including rare or extreme events that may be underrepresented in historical data.

The Observation Model functions as a post-processing step for the synthetically generated outbreak data, introducing realistic imperfections to enhance its fidelity. This model applies stochastic processes to simulate inherent noise in real-world surveillance systems, including reporting delays, incomplete case detection, and inaccuracies in case classification. Specifically, the model introduces variations in reported case numbers based on probabilistic distributions informed by known limitations of epidemiological data collection. These variations are applied to both the timing and magnitude of reported cases, effectively simulating the complexities arising from imperfect observation of an actual outbreak scenario and increasing the robustness of forecasting models trained on this data.

The integration of synthetically generated outbreak data with observed real-world data forms a composite training dataset designed to improve forecasting model performance. This combined approach addresses limitations inherent in either data source alone; real-world data can be sparse and lack representation of rare but critical outbreak scenarios, while purely synthetic data may not fully capture the nuances of actual epidemiological events. By leveraging the strengths of both, the resulting training set increases model robustness to variations in outbreak characteristics and improves generalizability across diverse epidemiological contexts. Evaluations demonstrate that models trained on this combined dataset consistently achieve state-of-the-art performance metrics in outbreak forecasting, exceeding the accuracy of models trained on either synthetic or real-world data in isolation.

The observation model generates diverse realizations of a MutAntiGen output by applying either scaling and potential outliers, or scaling with added noise and outliers, demonstrating the model's sensitivity to perturbations.
The observation model generates diverse realizations of a MutAntiGen output by applying either scaling and potential outliers, or scaling with added noise and outliers, demonstrating the model’s sensitivity to perturbations.

A Deep Learning Architecture for Precision Forecasting

The forecasting model utilizes a Transformer architecture, a deep learning approach particularly suited for sequential data analysis. Unlike recurrent neural networks, Transformers rely on self-attention mechanisms to weigh the importance of different data points in the sequence, enabling parallel processing and improved capture of long-range dependencies. This is achieved through multiple layers of attention and feed-forward networks, allowing the model to learn complex relationships within the time series data without being constrained by the sequential processing limitations of earlier models. The Transformer’s ability to model these dependencies is crucial for accurate forecasting, particularly in datasets exhibiting non-linear patterns or complex interactions between variables.

Pinball loss, also known as quantile loss, functions as the objective function during model training to directly optimize for specific quantiles of the forecast distribution. Unlike mean squared error which focuses on minimizing the average error, pinball loss asymmetrically penalizes overestimation and underestimation, allowing the model to learn different quantiles with varying degrees of precision. The loss function is defined as \max(y_i - \hat{y}_i, \tau (y_i - \hat{y}_i)) , where y_i is the actual value, \hat{y}_i is the predicted value, and τ represents the desired quantile (between 0 and 1). By training separate models for different quantile levels, the system generates probabilistic forecasts, quantifying the uncertainty associated with each prediction and enabling more informed risk assessment and decision-making.

Parameter selection within the forecasting model utilizes an Exponential Moving Average (EMA) to balance responsiveness to recent data with the stability conferred by historical values. The EMA is calculated by weighting current observations more heavily than older observations, with the weighting decreasing exponentially. This approach, defined by a smoothing factor α between 0 and 1, reduces the impact of noisy data and prevents drastic parameter shifts that could negatively affect forecast accuracy. The smoothing factor α determines the rate at which the EMA responds to changes; a higher value of α prioritizes recent data, while a lower value emphasizes historical data, ultimately contributing to a stable and optimized model performance over time.

Quantile Regression extends forecasting capabilities beyond point estimates by predicting specific quantiles of the future distribution. Instead of solely predicting the most likely value, the model estimates values corresponding to defined probability levels – for example, the 5th, 50th (median), and 95th percentiles. This provides a range of plausible outcomes, enabling a more nuanced understanding of potential risks and opportunities. The output is a probabilistic forecast, offering interval predictions rather than single-value predictions, which is particularly valuable for decision-making under uncertainty, allowing stakeholders to assess the likelihood of different scenarios and make informed choices based on their risk tolerance and objectives.

Models incorporating vaccination context (VAC) demonstrably outperform time-context-only models during the declining phases of an outbreak, as evidenced by lower <span class="katex-eq" data-katex-display="false">rMAE</span> and <span class="katex-eq" data-katex-display="false">rWIS</span> values with 95% confidence intervals, while performing comparably during the initial rise.
Models incorporating vaccination context (VAC) demonstrably outperform time-context-only models during the declining phases of an outbreak, as evidenced by lower rMAE and rWIS values with 95% confidence intervals, while performing comparably during the initial rise.

Validating Predictive Power and Quantifying Uncertainty

Forecasting accuracy was rigorously assessed using established metrics, including the Relative Weighted Interval Score (rWIS), to quantify predictive performance. Recognizing the inherent dependencies within pandemic data – where observations are not independent – a robust resampling technique called Block Bootstrap was implemented. This method preserves the temporal correlations present in the data, providing more realistic estimates of forecast uncertainty compared to simpler resampling approaches. By generating numerous bootstrap samples, the variability of the forecasts could be characterized, enabling a statistically sound evaluation of model performance and a reliable comparison against existing forecasting models like the COVIDHub-4_week_ensemble. The use of Block Bootstrap strengthens the validity of the conclusions drawn regarding the forecasting approach’s ability to accurately predict key pandemic indicators.

Evaluations reveal a consistent and significant advantage for this forecasting approach when predicting crucial pandemic indicators. Utilizing the Relative Weighted Interval Score (rWIS) as a key metric, the model achieved a score of 0.81, demonstrably exceeding the performance of the established COVIDHub-4_week_ensemble model. This improvement isn’t merely incremental; it signifies a substantial enhancement in predictive capability, offering a more reliable basis for informed decision-making during public health crises. The rWIS score reflects the model’s ability to generate accurate forecasts with well-calibrated uncertainty intervals, proving its value beyond simply predicting the central tendency of pandemic trends.

A crucial aspect of forecast reliability lies in calibration, or how well the predicted uncertainty matches observed outcomes. This study’s models demonstrated an empirical coverage rate of 0.84 to 0.88, indicating that, for approximately 84 to 88 percent of predictions, the true values fell within the stated confidence intervals. This is a significant improvement over the comparison COVIDHub-4_week_ensemble model, which achieved a coverage of 0.8. Higher empirical coverage suggests these models are better calibrated, providing more trustworthy estimates of prediction uncertainty and allowing for more informed decision-making based on those forecasts; a model that consistently underestimates uncertainty can lead to a false sense of security, while overestimation can stifle proactive responses.

Rigorous statistical validation confirmed the improvements achieved by incorporating synthetic data into the modeling process. Through the generation of 5,000 bootstrap samples, confidence intervals were established for model performance, revealing statistically significant enhancements when models were trained on a combination of real and synthetic datasets. This approach not only refined predictive accuracy but also provided a robust assessment of uncertainty, demonstrating a marked improvement over models reliant solely on real-world data. The resulting confidence intervals offered a dependable measure of the model’s reliability, bolstering the validity of its forecasts and offering increased confidence in its projections of pandemic indicators.

Accounting for state and forecast date groupings via a blocked bootstrap procedure results in wider uncertainty estimates for <span class="katex-eq" data-katex-display="false">rMAE</span> and <span class="katex-eq" data-katex-display="false">rWIS</span> compared to a standard independent and identically distributed bootstrap, indicating the importance of these groupings in assessing forecast accuracy.
Accounting for state and forecast date groupings via a blocked bootstrap procedure results in wider uncertainty estimates for rMAE and rWIS compared to a standard independent and identically distributed bootstrap, indicating the importance of these groupings in assessing forecast accuracy.

Expanding Horizons: Towards a Future of Proactive Pandemic Preparedness

Ensemble forecasting represents a significant advancement in predictive modeling for infectious diseases by strategically combining the outputs of multiple independent models. Rather than relying on a single, potentially flawed, simulation, this approach leverages the strengths of diverse methodologies, effectively mitigating individual biases inherent in each. Each model, trained on varying assumptions and data, captures different facets of the outbreak’s dynamics; the ensemble then synthesizes these perspectives, producing a more robust and accurate prediction. This isn’t simply averaging results, but rather a weighted combination, often prioritizing models that have demonstrated superior performance in historical data or exhibit greater consistency with real-time observations. The resulting forecast is less susceptible to extreme errors and provides a more reliable basis for public health interventions, offering a crucial advantage in the face of rapidly evolving pandemic scenarios.

The developed forecasting framework isn’t limited to the current pandemic; its modular design facilitates adaptation to a wide range of infectious diseases. By retraining the models with data from different pathogens, predictions can be generated for influenza, measles, or emerging threats with similar transmission dynamics. Furthermore, the framework is poised to benefit significantly from the increasing availability of genomic surveillance data. Integrating real-time genomic information – tracking viral mutations, identifying new variants, and assessing their potential impact on transmissibility and severity – promises to substantially refine prediction accuracy and enable more targeted public health interventions. This capacity for data integration positions the framework as a versatile tool for bolstering global pandemic preparedness and responding effectively to future outbreaks, regardless of the causative agent.

Sustained investment in pandemic prediction research promises a fundamental shift from reactive crisis management to proactive preparedness. Future development centers on refining forecasting models with increasingly granular data – incorporating real-time genomic sequencing, social media trends, and mobility patterns – to anticipate outbreaks before they escalate. This expanded capacity will enable targeted interventions, such as pre-emptive vaccine distribution and localized public health campaigns, significantly reducing morbidity and mortality. Moreover, advancements in predictive analytics can inform resource allocation, strengthening healthcare systems and ensuring swift, effective responses to emerging threats, ultimately minimizing the disruptive impact of future pandemics on global health and economies.

A dataset of over 2,000 real respiratory time series, comprising more than 2 million observations, provides a robust resource for training and analysis beyond COVID-19.
A dataset of over 2,000 real respiratory time series, comprising more than 2 million observations, provides a robust resource for training and analysis beyond COVID-19.

The pursuit of accurate epidemic forecasting, as demonstrated by this research into deep learning and synthetic data, echoes a timeless principle of harmonious construction. The study’s success in leveraging limited historical data through innovative modeling techniques aligns with the idea that true mastery lies in resourceful adaptation. As Confucius observed, “Study the past if you would define the future.” This work doesn’t merely predict outbreaks; it constructs a more resilient understanding of disease transmission, mirroring the elegance found in systems where form-the model’s architecture-and function-accurate forecasting-are inextricably linked. Such refined construction, built upon both observation and ingenuity, ensures lasting comprehensibility and durability.

The Horizon Beckons

The pursuit of accurate epidemic forecasting, as demonstrated by this work, is not merely a technical exercise. It is an attempt to distill order from inherent chaos, to anticipate the inevitable shifts in a complex, dynamic system. The integration of synthetic data, while promising, highlights a fundamental truth: that the quality of foresight is always limited by the fidelity of its models. The elegance of a prediction lies not in its complexity, but in its capacity to capture essential patterns with minimal artifice. The current methodologies, even at their peak performance, remain dependent on assumptions about viral evolution and human behavior – assumptions that, inevitably, will prove incomplete.

A truly robust forecasting system demands a move beyond purely data-driven approaches. The future likely resides in hybrid models, incorporating mechanistic understanding of disease transmission with the predictive power of deep learning. Furthermore, the attribution of cases to specific variants, while a necessary refinement, exposes the limitations of compartmentalized thinking. Variants are not discrete entities, but points on a continuum of evolution, and the models must reflect that fluidity.

The ideal interface, in this context, is invisible to the user, yet felt – a seamless integration of prediction and preparedness. Every change to these forecasting systems should be justified by beauty and clarity, not merely by incremental gains in accuracy. The ultimate measure of success will not be the ability to predict the future with certainty, but to reduce the suffering caused by its uncertainties.


Original article: https://arxiv.org/pdf/2603.24474.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-26 21:17