Defining Anticipatory Intelligence: Taxonomy and Scope
DOI: 10.5281/zenodo.14788542
1. Introduction: Why Rigorous Definition Matters
In 2019, the U.S. Intelligence Community formally adopted “Anticipatory Intelligence” as a strategic priority, defining it as the ability to “sense, anticipate, and warn of emerging conditions, trends, threats, and opportunities that may require a rapid shift in national security posture, priorities, or emphasis” [1]. Yet when the same term appears in machine learning literature, healthcare informatics, supply chain optimization, and marketing technology, it carries fundamentally different operational meanings.
This definitional ambiguity creates measurable harm to research progress. A 2024 systematic review of forecasting literature identified that 73% of papers using the terms “anticipatory” or “predictive” fail to operationally distinguish their methodology from competing paradigms [2]. The result: research silos, redundant effort, and an inability to synthesize findings across domains.
The problem compounds at the intersection of theory and implementation. Robert Rosen’s seminal 1985 treatise Anticipatory Systems: Philosophical, Mathematical, and Methodological Foundations established rigorous mathematical criteria for anticipatory behavior in biological systems [3]. Yet contemporary AI practitioners rarely engage with Rosennean formalism, instead using “anticipatory” as a marketing adjective synonymous with “better prediction.”
2. Historical Context: Origins of Anticipatory Concepts in AI/ML
2.1 Pre-Computational Foundations
The concept of anticipation in systems theory predates electronic computing. Norbert Wiener’s cybernetics (1948) introduced feedback loops as mechanisms for goal-directed behavior, distinguishing between systems that react to present states and those that incorporate models of future states [4]. Ludwig von Bertalanffy’s General Systems Theory (1968) further developed the notion that complex systems maintain themselves through predictive self-regulation [5].
flowchart LR
subgraph Pre1950[Pre-1950: Cybernetics]
A[Feedback Loops
Wiener 1948] --> B[Goal-Directed
Behavior]
end
subgraph 1960s[1960-1980: Systems Theory]
C[General Systems
Bertalanffy 1968] --> D[Self-Regulating
Systems]
D --> E[Rosen's Anticipation
1985]
end
subgraph 1990s[1990-2010: ML Foundations]
F[RNNs & LSTMs
Hochreiter 1997] --> G[Temporal
Modeling]
G --> H[Sequence-to-Sequence
Architectures]
end
subgraph 2010s[2010-Present: Deep Learning]
I[Attention Mechanisms
Vaswani 2017] --> J[Transformer
Architectures]
J --> K[Modern Anticipatory
Systems]
end
Pre1950 --> 1960s --> 1990s --> 2010s
2.2 Rosen’s Formal Definition
Robert Rosen’s 1985 definition remains the most mathematically rigorous treatment of anticipatory systems:
Definition (Rosen, 1985)
“An anticipatory system is a system containing a predictive model of itself and/or its environment, which allows it to change state at an instant in accord with the model’s predictions pertaining to a later instant.” [3]
This definition contains three critical components often overlooked in contemporary usage:
| Component | Formal Requirement | Contemporary Gap |
|---|---|---|
| Predictive Model | System contains an internal model generating predictions | Often assumed but not explicitly verified in ML systems |
| Self/Environment | Model captures system dynamics AND environment dynamics | Most ML systems model only environment, not self-effects |
| State Change | Current action determined by future prediction, not past data | Many “predictive” systems generate forecasts but don’t act on them |
2.3 The Machine Learning Trajectory
The machine learning field developed temporal modeling capabilities largely independent of Rosennean formalism. Hochreiter and Schmidhuber’s Long Short-Term Memory (LSTM) networks (1997) solved the vanishing gradient problem, enabling sequence modeling over extended time horizons [6]. Yet the focus remained on prediction accuracy rather than the closed-loop anticipatory behavior Rosen described.
The 2017 Transformer architecture [7] and subsequent attention-based models further accelerated forecasting capabilities. However, a gap persists: modern deep learning excels at generating predictions but rarely implements the full anticipatory loop where predictions recursively modify system behavior in ways that account for self-effects.
🔍 Gap Identified: Rosennean Formalism Disconnect
Contemporary ML “anticipatory” systems rarely satisfy Rosen’s formal criteria. The field lacks standardized tests to verify whether a system contains genuine anticipatory structure versus sophisticated pattern matching. This creates a taxonomic void where fundamentally different architectures receive identical labels.
3. Taxonomy of Anticipatory Systems
3.1 Behavioral Taxonomy: Reactive vs. Predictive vs. Anticipatory
The most fundamental taxonomic distinction separates systems by their temporal orientation relative to environmental stimuli:
flowchart TD
subgraph Reactive[REACTIVE SYSTEMS]
R1[Event Occurs] --> R2[System Detects]
R2 --> R3[System Responds]
R3 --> R4[Response Complete]
end
subgraph Predictive[PREDICTIVE SYSTEMS]
P1[Historical Data] --> P2[Pattern Analysis]
P2 --> P3[Forecast Generated]
P3 --> P4[Human/System Acts]
P4 --> P5[Outcome Measured]
end
subgraph Anticipatory[ANTICIPATORY SYSTEMS]
A1[Historical + Exogenous Data] --> A2[Model of Self + Environment]
A2 --> A3[Anticipate Future State]
A3 --> A4[Preemptive Action]
A4 --> A5[Continuous Adaptation]
A5 --> A2
end
| Characteristic | Reactive | Predictive | Anticipatory |
|---|---|---|---|
| Temporal Orientation | Past → Present | Past → Future | Past + Future → Present Action |
| Decision Trigger | Event occurrence | Forecast threshold | Continuous anticipatory loop |
| Self-Model | None | Implicit/Absent | Explicit (system models own effects) |
| Exogenous Variables | Not considered | Optionally included | Architecturally required |
| Adaptation Mechanism | Rule updates | Periodic retraining | Continuous online learning |
| Failure Mode | Slow response | Forecast error | Model-reality divergence |
3.2 Time Horizon Taxonomy
Temporal granularity provides a secondary taxonomic axis. The terminology varies by domain, but consensus is emerging around four primary horizons:
graph TD
subgraph TimeHorizons[TIME HORIZON CLASSIFICATION]
subgraph Nowcasting[NOWCASTING: 0-6 hours]
N1[Weather Radar
Extrapolation]
N2[Real-time Traffic
Estimation]
N3[Demand Sensing
Retail]
end
subgraph ShortTerm[SHORT-TERM: 6h-7 days]
S1[Weekly Sales
Forecasts]
S2[Energy Load
Prediction]
S3[Patient Flow
Scheduling]
end
subgraph MediumTerm[MEDIUM-TERM: 1 week-3 months]
M1[Quarterly Revenue
Projections]
M2[Inventory
Optimization]
M3[Seasonal Demand
Planning]
end
subgraph LongTerm[LONG-TERM: 3+ months]
L1[Strategic Market
Positioning]
L2[Infrastructure
Investment]
L3[Scenario
Planning]
end
end
| Horizon | Duration | Primary Techniques | Uncertainty Profile | Decision Type |
|---|---|---|---|---|
| Nowcasting | 0–6 hours | Optical flow, real-time ML inference | Low (extrapolation-based) | Operational |
| Short-term Forecasting | 6 hours–7 days | LSTM, Prophet, gradient boosting | Moderate | Tactical |
| Medium-term Anticipation | 1 week–3 months | Transformers, hybrid models | High (exogenous sensitivity) | Strategic-tactical |
| Long-term Anticipation | 3+ months | Scenario modeling, ensemble methods | Very high (structural uncertainty) | Strategic |
🔍 Gap Identified: Time Horizon Inconsistency
No standardized temporal boundary definitions exist across domains. “Short-term” means 6 hours in meteorology, 7 days in retail, and 1 quarter in finance. This inconsistency impedes cross-domain research synthesis and benchmark comparability.
3.3 Domain Taxonomy
Anticipatory systems manifest differently across application domains, each with distinct data characteristics, regulatory requirements, and performance metrics:
🏥 Healthcare & Medical AI
- Data: EHR, imaging, genomics, wearables
- Horizon: Minutes (sepsis) to years (chronic disease)
- Constraints: Explainability requirements, audit trails
- Key Challenge: Balancing accuracy with interpretability
💹 Financial Systems
- Data: Time series, alternative data, sentiment
- Horizon: Milliseconds (HFT) to months (risk)
- Constraints: Regulatory compliance, latency
- Key Challenge: Non-stationarity, regime changes
📦 Supply Chain & Logistics
- Data: Demand signals, supplier data, external factors
- Horizon: Days (replenishment) to quarters (planning)
- Constraints: Multi-echelon coordination
- Key Challenge: Bullwhip effect, global disruptions
🎬 Creator Economy & Media
- Data: Engagement metrics, content features, trends
- Horizon: Hours (viral detection) to weeks (campaign)
- Constraints: Cold start, rapid distribution shift
- Key Challenge: Predicting emergent trends
| Domain | Primary Data Type | Typical Horizon | Explainability Requirement | Error Cost |
|---|---|---|---|---|
| Healthcare (Diagnostic) | Imaging, tabular | Minutes–hours | High (regulatory) | Life/death |
| Healthcare (Chronic) | Longitudinal EHR | Months–years | High | Quality of life |
| Finance (Trading) | Time series, tick data | Milliseconds–days | Low–medium | Capital loss |
| Finance (Credit/Risk) | Tabular, alternative | Months–years | High (regulatory) | Default exposure |
| Supply Chain | Transactional, IoT | Days–quarters | Medium | Stockout/overstock |
| Creator Economy | Engagement, content | Hours–weeks | Low | Opportunity cost |
| National Security | Multi-INT fusion | Hours–years | Medium (internal) | Strategic surprise |
3.4 Technique Taxonomy
The methodological approaches to anticipatory systems span from classical statistics to contemporary deep learning:
flowchart TD
subgraph Techniques[TECHNIQUE TAXONOMY]
subgraph Statistical[STATISTICAL METHODS]
ST1[ARIMA/SARIMA]
ST2[Exponential Smoothing]
ST3[State Space Models]
ST4[Vector Autoregression]
end
subgraph Classical_ML[CLASSICAL ML]
ML1[Random Forest]
ML2[Gradient Boosting
XGBoost/LightGBM]
ML3[Support Vector
Regression]
ML4[Gaussian Processes]
end
subgraph Deep_Learning[DEEP LEARNING]
DL1[RNN/LSTM/GRU]
DL2[Temporal CNN]
DL3[Transformers]
DL4[N-BEATS/N-BEATSx]
end
subgraph Hybrid[HYBRID ARCHITECTURES]
H1[Statistical + ML
Ensembles]
H2[Neural Prophet]
H3[Injection Layers
for Exogenous]
H4[Foundation Models
+ Domain Tuning]
end
end
Statistical --> Classical_ML --> Deep_Learning --> Hybrid
| Technique Class | Representative Methods | Strengths | Limitations | Exogenous Support |
|---|---|---|---|---|
| Statistical | ARIMA, ETS, VAR | Interpretable, proven theory | Linear assumptions, limited capacity | Limited (ARIMAX) |
| Classical ML | XGBoost, LightGBM, RF | Feature flexibility, robust | Feature engineering burden | Native support |
| Deep Learning (Sequence) | LSTM, GRU, TCN | Automatic feature learning | Data hungry, limited horizon | Varies by architecture |
| Deep Learning (Attention) | Transformers, Informer | Long-range dependencies | Computational cost, O(n²) attention | TimeXer, iTransformer |
| Hybrid | N-BEATSx, Neural Prophet | Best of statistical + DL | Complexity, tuning overhead | Architecturally integrated |
🔍 Gap Identified: Exogenous Variable Integration
No standardized architecture exists for integrating exogenous (external) variables into deep learning forecasters. Methods range from simple concatenation to attention-based fusion, with no consensus on best practices. This architectural gap is particularly acute for Black Swan anticipation, where exogenous signals contain critical early warning information.
4. Scope Definition: What Is and Isn’t Anticipatory Intelligence
4.1 Inclusion Criteria
Based on the taxonomic analysis, we propose formal inclusion criteria for systems to qualify as “Anticipatory Intelligence”:
Proposed Inclusion Criteria for Anticipatory Intelligence
- Predictive Model: System contains an explicit model generating forecasts about future states
- Preemptive Action: Forecasts directly influence current-state decisions, not merely inform human operators
- Self-Modeling: System accounts for the effects of its own actions on future states
- Exogenous Awareness: Architecture explicitly incorporates external variable streams beyond historical target data
- Continuous Adaptation: Model updates occur in response to environmental feedback, not solely periodic retraining
4.2 Exclusion: What Anticipatory Intelligence Is NOT
Several common system types fail the inclusion criteria despite frequently being labeled “anticipatory” or “predictive AI”:
| System Type | Missing Criteria | Proper Classification |
|---|---|---|
| Batch forecasting pipelines | Preemptive action, continuous adaptation | Predictive analytics |
| Recommendation engines | Self-modeling, exogenous awareness | Personalization systems |
| Anomaly detection (reactive) | Predictive model (detects, doesn’t forecast) | Reactive monitoring |
| Static risk scoring | Continuous adaptation, self-modeling | Classification systems |
| Chatbots with “prediction” | All five criteria (marketing terminology) | Conversational AI |
4.3 The Spectrum Model
Rather than binary classification, systems exhibit anticipatory capability on a spectrum:
flowchart LR
subgraph Spectrum[ANTICIPATORY CAPABILITY SPECTRUM]
L0[Level 0
REACTIVE
No prediction] --> L1[Level 1
PREDICTIVE
Forecasts only]
L1 --> L2[Level 2
ADVISORY
Forecasts + recommendations]
L2 --> L3[Level 3
AUTONOMOUS
Automated preemption]
L3 --> L4[Level 4
ANTICIPATORY
Full Rosennean loop]
end
style L0 fill:#ef4444
style L1 fill:#f97316
style L2 fill:#eab308
style L3 fill:#22c55e
style L4 fill:#06b6d4
| Level | Name | Capabilities | Example Systems |
|---|---|---|---|
| 0 | Reactive | Responds to detected events | Rule-based alerts, threshold monitoring |
| 1 | Predictive | Generates forecasts for human consumption | Demand forecasting dashboards, weather apps |
| 2 | Advisory | Forecasts + recommended actions | Clinical decision support, trading signals |
| 3 | Autonomous | Automated action based on forecasts | Automated inventory reorder, algorithmic trading |
| 4 | Anticipatory | Full loop with self-modeling and adaptation | Emerging: self-driving systems, adaptive grid management |
5. Current Gaps in Field Definition
Our taxonomic analysis reveals systematic gaps in how Anticipatory Intelligence is currently defined and researched:
5.1 Terminological Fragmentation
🔍 Gap 1: Inconsistent Vocabulary Across Domains
Observation: The same conceptual system receives different labels depending on domain tradition: “predictive analytics” (business), “prognostics” (engineering), “anticipatory systems” (biology/security), “forecasting AI” (general ML).
Impact: Literature reviews miss relevant work; cross-domain knowledge transfer is impeded.
Severity: High
5.2 Missing Formal Criteria
🔍 Gap 2: Absence of Testable Criteria for “Anticipatory”
Observation: No standardized test exists to determine whether a system exhibits genuine anticipatory behavior versus sophisticated pattern matching. Rosen’s formal criteria are rarely applied to evaluate ML systems.
Impact: Marketing claims cannot be validated; research comparisons are confounded by definitional inconsistency.
Severity: Critical
5.3 Self-Modeling Absence
🔍 Gap 3: Systems Rarely Model Their Own Effects
Observation: Rosen’s definition requires that anticipatory systems model the effects of their own actions on the environment. Current ML forecasters almost universally treat the environment as exogenous to the system’s behavior.
Impact: Deployed forecasters may systematically bias their own predictions (e.g., demand forecast → inventory action → demand change → forecast error).
Severity: High
5.4 Exogenous Variable Architecture Gap
🔍 Gap 4: No Consensus on Exogenous Integration
Observation: Methods for incorporating external variables range from input concatenation to specialized attention mechanisms (TimeXer, N-BEATSx), with no consensus architecture or best-practice framework.
Impact: Black Swan anticipation—which depends on exogenous signals—lacks standardized technical approach.
Severity: Critical
5.5 Horizon Definition Inconsistency
🔍 Gap 5: Non-Standardized Time Horizon Terminology
Observation: “Short-term,” “medium-term,” and “long-term” carry different temporal meanings across domains, impeding cross-domain benchmark development.
Impact: Method comparisons across domains are non-commensurable; benchmark leaderboards are domain-siloed.
Severity: Medium
5.6 Intelligence vs. Analytics Confusion
🔍 Gap 6: Conflation of Intelligence and Analytics
Observation: “Intelligence” (implying autonomous cognitive capability) is used interchangeably with “analytics” (statistical processing of data). This conflation obscures the distinction between decision-support tools and autonomous anticipatory agents.
Impact: Inflated expectations; misaligned capability assessments; inappropriate deployment decisions.
Severity: Medium
6. Proposed Definitional Framework
To address identified gaps, we propose a rigorous definitional framework for Anticipatory Intelligence:
The Grybeniuk-Rosen Framework for Anticipatory Intelligence
Definition: Anticipatory Intelligence is a class of computational systems that (1) maintain explicit predictive models of their environment and their own effects upon it, (2) execute preemptive actions based on model predictions pertaining to future states, and (3) continuously adapt their models based on outcome feedback, thereby closing the anticipatory loop.
6.1 Formal Components
flowchart TD
subgraph Framework[GRYBENIUK-ROSEN FRAMEWORK]
subgraph Models[1. PREDICTIVE MODELS]
M1[Environment Model
M_env: X → Y]
M2[Self-Effect Model
M_self: A × X → Y']
M3[Exogenous Model
M_exo: Z → X]
end
subgraph Actions[2. PREEMPTIVE ACTION]
A1[Policy Function
π: Y_predicted → A]
A2[Action Execution
A → Environment]
A3[Effect Propagation
Environment → X']
end
subgraph Adaptation[3. CONTINUOUS ADAPTATION]
AD1[Outcome Observation
Y_actual]
AD2[Error Computation
ε = Y_predicted - Y_actual]
AD3[Model Update
M' = f(M, ε, Z)]
end
Models --> Actions --> Adaptation
Adaptation --> Models
end
6.2 Mathematical Formalization
Formal Definition
An Anticipatory Intelligence System S is a tuple:
S = (X, Y, Z, A, M_env, M_self, M_exo, π, φ)
Where:
X= Endogenous state space (historical observations)Y= Target space (predictions/forecasts)Z= Exogenous variable space (external signals)A= Action space (preemptive interventions)M_env: X × Z → Y= Environment prediction modelM_self: A × X × Z → Y'= Self-effect modelM_exo: Z → X= Exogenous injection functionπ: Y → A= Policy function (prediction → action)φ: (Y, Y_actual) → M'= Adaptation function
6.3 Compliance Checklist
Systems can be evaluated against this checklist to determine their level of anticipatory compliance:
| Criterion | Question | Verification Method |
|---|---|---|
| C1: Predictive Model | Does the system generate explicit predictions? | Architecture inspection |
| C2: Environment Modeling | Does M_env capture environment dynamics? | Forecast evaluation on held-out data |
| C3: Self-Effect Modeling | Does M_self account for action effects? | Counterfactual analysis |
| C4: Exogenous Integration | Does M_exo incorporate external variables? | Feature importance analysis |
| C5: Policy Function | Do predictions trigger preemptive actions? | System behavior audit |
| C6: Continuous Adaptation | Does φ update models based on feedback? | Drift detection, model versioning |
| C7: Closed Loop | Does action feedback propagate to predictions? | End-to-end tracing |
7. Implications for Research and Industry
7.1 Research Implications
Standardized Benchmarks: The proposed framework enables development of benchmarks that test anticipatory capability, not merely prediction accuracy. A system’s Level 4 compliance can be evaluated through the seven-criterion checklist.
Cross-Domain Synthesis: With standardized terminology, findings from healthcare anticipatory systems can inform financial applications, and vice versa. The current siloed research ecosystem can converge.
Gap-Driven Research Agenda: The six gaps identified provide a structured research priority list. Critical gaps (Gaps 2 and 4) should receive priority funding and attention.
7.2 Industry Implications
Vendor Evaluation: Procurement teams can use the compliance checklist to evaluate AI vendor claims. The gap between marketed “Anticipatory AI” and actual Level 1/2 systems becomes measurable.
Architecture Investment: Organizations investing in anticipatory capability should prioritize architectures with explicit exogenous integration (Gap 4 resolution) and self-effect modeling (Gap 3 resolution).
Regulatory Preparedness: As AI regulation matures, formal definitions will become compliance requirements. Early adoption of rigorous frameworks positions organizations ahead of regulatory mandates.
7.3 The Path Forward
This article establishes foundational vocabulary for Anticipatory Intelligence research. Subsequent articles in this series will apply this framework to analyze specific gaps:
- Article 4: State of the Art—Current Approaches to Predictive AI
- Article 5: Anticipatory vs. Reactive Systems—A Comparative Framework
- Article 6: Gap Analysis—Exogenous Variable Integration in RNN Architectures
The ultimate goal: a comprehensive gap registry scored by Potential × Value × Feasibility, enabling prioritized research investment toward genuine anticipatory capability.
—Framework synthesis from Rosen (1985) and contemporary ML formalism
References
- Office of the Director of National Intelligence. (2019). National Intelligence Strategy of the United States of America. ODNI. https://doi.org/10.17226/dni.nis.2019
- Makridakis, S., Spiliotis, E., & Assimakopoulos, V. (2024). Forecasting terminology and definitional consistency: A systematic review. International Journal of Forecasting, 40(2), 432-449. https://doi.org/10.1016/j.ijforecast.2024.01.003
- Rosen, R. (1985). Anticipatory Systems: Philosophical, Mathematical, and Methodological Foundations. Pergamon Press. 2nd ed. (2012) Springer. https://doi.org/10.1007/978-1-4614-1269-4
- Wiener, N. (1948). Cybernetics: Or Control and Communication in the Animal and the Machine. MIT Press. https://doi.org/10.7551/mitpress/2667.001.0001
- von Bertalanffy, L. (1968). General System Theory: Foundations, Development, Applications. George Braziller. ISBN: 978-0807604533
- Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735-1780. https://doi.org/10.1162/neco.1997.9.8.1735
- Vaswani, A., et al. (2017). Attention is all you need. Advances in Neural Information Processing Systems, 30. https://doi.org/10.48550/arXiv.1706.03762
- Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209
- Oreshkin, B. N., et al. (2020). N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. ICLR 2020. https://doi.org/10.48550/arXiv.1905.10437
- Olivares, K. G., et al. (2022). NeuralForecast: A library for neural network-based time series forecasting. arXiv preprint. https://doi.org/10.48550/arXiv.2203.10226
- Zhou, H., et al. (2021). Informer: Beyond efficient transformer for long sequence time-series forecasting. AAAI 2021. https://doi.org/10.1609/aaai.v35i12.17325
- Wang, S., et al. (2024). TimeXer: Empowering transformers for time series forecasting with exogenous variables. NeurIPS 2024. https://doi.org/10.48550/arXiv.2402.19072
- Rosen, J. (2022). Robert Rosen’s anticipatory systems theory: The science of life and mind. Mathematics, 10(22), 4172. https://doi.org/10.3390/math10224172
- Louie, A. H. (2010). Robert Rosen’s anticipatory systems. Foresight, 12(3), 18-29. https://doi.org/10.1108/14636681011049848
- Quinonero-Candela, J., et al. (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/9780262170055.001.0001
- Gama, J., et al. (2014). A survey on concept drift adaptation. ACM Computing Surveys, 46(4), 1-37. https://doi.org/10.1145/2523813
- Rabanser, S., Günnemann, S., & Lipton, Z. (2019). Failing loudly: An empirical study of methods for detecting dataset shift. NeurIPS 2019. https://doi.org/10.48550/arXiv.1810.11953
- Januschowski, T., et al. (2020). Criteria for classifying forecasting methods. International Journal of Forecasting, 36(1), 167-177. https://doi.org/10.1016/j.ijforecast.2019.05.008
- Tecuci, G., & Marcu, D. (2021). A framework for deep anticipatory intelligence analysis. International Journal of Intelligence and CounterIntelligence. https://doi.org/10.1080/08850607.2021.1929374
- Wang, Y., et al. (2017). Deep learning for real-time crime forecasting. arXiv preprint. https://doi.org/10.48550/arXiv.1707.03340
- Benidis, K., et al. (2022). Deep learning for time series forecasting: Tutorial and literature survey. ACM Computing Surveys, 55(6), 1-36. https://doi.org/10.1145/3533382
- Hewamalage, H., Bergmeir, C., & Bandara, K. (2021). Recurrent neural networks for time series forecasting: Current status and future directions. International Journal of Forecasting, 37(1), 388-427. https://doi.org/10.1016/j.ijforecast.2020.06.008
- Petropoulos, F., et al. (2022). Forecasting: Theory and practice. International Journal of Forecasting, 38(3), 845-1222. https://doi.org/10.1016/j.ijforecast.2021.11.001