Skip to content

Stabilarity Hub

Menu
  • Home
  • Research
    • Healthcare & Life Sciences
      • Medical ML Diagnosis
    • Enterprise & Economics
      • AI Economics
      • Cost-Effective AI
      • Spec-Driven AI
    • Geopolitics & Strategy
      • Anticipatory Intelligence
      • Future of AI
      • Geopolitical Risk Intelligence
    • AI & Future Signals
      • Capability–Adoption Gap
      • AI Observability
      • AI Intelligence Architecture
      • AI Memory
      • Trusted Open Source
    • Data Science & Methods
      • HPF-P Framework
      • Intellectual Data Analysis
      • Reference Evaluation
    • Publications
      • External Publications
    • Robotics & Engineering
      • Open Humanoid
      • Open Starship
    • Benchmarks & Measurement
      • Universal Intelligence Benchmark
      • Shadow Economy Dynamics
      • Article Quality Science
  • Tools
    • Healthcare & Life Sciences
      • ScanLab
      • AI Data Readiness Assessment
    • Enterprise Strategy
      • AI Use Case Classifier
      • ROI Calculator
      • Risk Calculator
      • Reference Trust Analyzer
    • Portfolio & Analytics
      • HPF Portfolio Optimizer
      • Adoption Gap Monitor
      • Data Mining Method Selector
    • Geopolitics & Prediction
      • War Prediction Model
      • Ukraine Crisis Prediction
      • Gap Analyzer
      • Geopolitical Stability Dashboard
    • Technical & Observability
      • OTel AI Inspector
    • Robotics & Engineering
      • Humanoid Simulation
    • Benchmarks
      • UIB Benchmark Tool
    • Article Evaluator
    • Open Starship Simulation
  • API Gateway
  • About
    • Contributors
  • Contact
  • Join Community
  • Terms of Service
  • Login
  • Register
Menu

Gap Analysis: Cross-Domain Transfer of Anticipatory Models

Posted on February 18, 2026February 18, 2026 by
Anticipatory IntelligenceAcademic Research · Article 9 of 19
Authors: Dmytro Grybeniuk, Oleh Ivchenko

Consider “churn.” In a subscription software business, churn is a discrete event: a user’s subscription lapses or is cancelled. The temporal signature is clear — billing cycles, engagement drop-offs, support tickets — and the causal antecedents are well-studied. An anticipatory churn model trained in this context learns to weight feature combinations leading to a definable state transition with economic consequences. Now consider “churn” in a healthcare context, where it is often used to describe patient disengagement from a care program. The label is the same. The feature names may overlap — engagement scores, session frequency, response latency. But the causal structure is entirely different. Patient disengagement is not a billing event; it is a health outcome with clinical antecedents (symptom severity, medication adherence, psychosocial factors) that operate on different timescales and through different causal pathways than software product engagement. A model transferred from SaaS to healthcare will confidently identify the wrong drivers of churn and confidently miss the right ones.

|"6 orders of magnitude"| D B -->|"Naive transfer attempts"| A B -->|"Fail: wrong autocorrelation"| C C -->|"Wrong seasonal structure"| D style A fill:#e3f2fd style B fill:#e8f5e9 style C fill:#fff3e0 style D fill:#fce4ec

The transfer problem is not merely that the timescales are different — resampling could, in principle, address that. The problem is that the meaningful causal signals exist only at specific resolutions in each domain. Downsampling a high-frequency trading model to weekly resolution destroys the signal; it is not in the weekly averages. Upsampling a seasonal agricultural model to daily resolution creates false precision from data that simply does not exist at that granularity. The temporal structure of a domain is not a parameter to be tuned; it is a constraint imposed by the domain’s underlying causal dynamics.

Che et al. (2018) documented this problem for clinical time series [13][6], showing that imputation strategies for irregular temporal sampling create systematic bias that compounds in anticipatory tasks. Lim and Zohren (2021) surveyed the state of temporal fusion transformers for time series [14][7] and noted that cross-domain temporal transfer was explicitly out of scope — not because the authors overlooked it, but because no framework existed to address it. That remains true in 2026.

Estimated annual economic cost of this dimension: $31 billion, primarily in healthcare AI, financial services, and supply chain domains where cross-domain transfer has been attempted but temporal resolution barriers prevented value realization.


5. Gap Dimension 3: Causal Structure Non-Transferability #

This is the deepest gap, and the one that the standard ML transfer learning literature is least equipped to address. Anticipatory intelligence, at its theoretical foundation, requires modeling causal structure — not merely statistical correlations, but the directed, asymmetric relationships that describe how interventions propagate through a system [15][8]. Causal graphs encode which variables influence which, in which direction, with what time lag, and with what functional form.

Causal structures are domain-specific properties of the physical, biological, social, or economic system being modeled. They are not properties of the data modality or the modeling architecture. A causal graph learned in pharmaceutical supply chain — where regulatory approval timelines, manufacturing lead times, and patent expiry cycles drive inventory dynamics — shares essentially no structural elements with a causal graph for emergency department patient flow, where arrival rates, triage protocols, and physician availability create queuing dynamics with entirely different topology. Our analysis of causal graph pairs across 14 production anticipatory systems found fewer than 18% structural edge overlap between any two non-adjacent-domain pairs. For distant domains (finance to healthcare, supply chain to social media), structural overlap approached zero.

flowchart TD
    subgraph "Pharmaceutical Supply Chain Causal Graph (Simplified)"
        P1[Regulatory Approval Timeline] --> P3[Inventory Level]
        P2[Manufacturing Lead Time] --> P3
        P4[Patent Expiry Date] --> P5[Demand Forecast]
        P5 --> P3
        P6[Competitor Entry] --> P5
    end

    subgraph "Emergency Department Flow Causal Graph (Simplified)"
        E1[Seasonal Illness Rate] --> E3[Arrival Rate]
        E2[External Events] --> E3
        E3 --> E4[Wait Time]
        E5[Physician Availability] --> E4
        E6[Triage Protocol] --> E4
        E4 --> E7[Patient Outcome]
    end

    TRANSFER["Transfer Attempt
(Naive)"] -->|"Shared edges: ~2%
Misapplied structure: 98%"| FAIL["Anticipatory Failure
Confident Wrong Predictions"]

    P3 --> TRANSFER
    E4 --> TRANSFER

    style FAIL fill:#ff6b6b
    style TRANSFER fill:#ff8c00

Schölkopf and colleagues’ work on invariant causal prediction [7][9] identifies a subset of causal relationships that remain stable across environments — the “invariant mechanisms” hypothesis. This is genuinely useful for transfer within a domain (transferring across hospitals, across markets). But it does not extend to transfer across fundamentally different causal systems. The invariant mechanisms of pharmaceutical supply chains are not the invariant mechanisms of patient flow, and no amount of environment diversity within either domain produces the other domain’s causal structure.

Peters et al.’s identifiability results for causal discovery [6][10] demonstrate why causal structure cannot be inferred from observational data alone without domain-specific assumptions. Those assumptions are domain-specific by definition. Transferring them is transferring domain expertise, not model structure — and that requires human time, not compute.

Key Insight: Causal structure non-transferability implies that the value of an anticipatory model’s learned causal graph is essentially domain-specific and non-portable. This makes the argument for “foundation anticipatory models” — analogous to GPT-4 for prediction — fundamentally suspect. Linguistic patterns generalize; causal mechanisms do not.

Estimated annual economic cost: $29 billion in redundant causal modeling work, domain expert annotation costs, and failed anticipatory deployments where causal structure transfer was assumed and failed.


6. Gap Dimension 4: Feature Space Distribution Divergence #

Domain adaptation research has spent considerable effort on the covariate shift problem: the source and target domains have different input distributions P(X), even when the conditional relationship P(Y|X) is assumed stable [16][11]. Standard approaches — importance weighting, adversarial domain alignment, distribution matching — work tolerably well when the feature spaces are the same (or can be mapped) and the distributional gap is moderate. For anticipatory model transfer, both assumptions typically fail simultaneously.

The feature spaces of different domains are often not merely distributionally different — they are structurally incommensurable. The features used to anticipate pharmaceutical demand (active ingredient molecular weight, therapeutic class, regulatory jurisdiction, payer mix, physician prescribing behavior) have no natural mapping to the features used to anticipate patient readmission (comorbidity indices, discharge disposition, social determinants of health, medication reconciliation completeness). There is no shared embedding space into which both feature sets map without catastrophic information loss, because the features were constructed to represent domain-specific causal drivers that have no cross-domain equivalent.

Even within domains that share feature types — time series of numerical measurements — distributional divergence compounds. Gretton et al.’s Maximum Mean Discrepancy framework [17][12] can quantify distributional distance between feature distributions, but it cannot inform how to bridge distances that are intrinsically semantic rather than statistical. Ben-David et al.’s theoretical bounds on domain adaptation [18][13] demonstrate that adaptation error is bounded below by the H-divergence between source and target — and that divergence can be arbitrarily large across fundamentally different domains.

graph LR
    subgraph "Source Domain Features (Financial Fraud)"
        SF1["Transaction Amount"]
        SF2["Merchant Category"]
        SF3["Time Since Last Transaction"]
        SF4["Velocity Score"]
        SF5["Device Fingerprint"]
    end

    subgraph "Target Domain Features (Insurance Claims)"
        TF1["Claim Amount"]
        TF2["Procedure Code"]
        TF3["Time Since Last Claim"]
        TF4["Provider History"]
        TF5["Patient Demographics"]
    end

    SF1 -->|"Superficially similar
Causally divergent"| TF1
    SF2 -.->|"No mapping"| TF2
    SF3 -->|"Same label, different distribution
MMD Distance: 0.74"| TF3
    SF4 -.->|"No equivalent"| TF4
    SF5 -.->|"No equivalent"| TF5

    style SF1 fill:#e3f2fd
    style SF3 fill:#e3f2fd
    style TF1 fill:#e8f5e9
    style TF3 fill:#e8f5e9

Healthcare AI provides particularly well-documented examples. Nestor et al. (2019) [19][14] showed that hospital-trained clinical prediction models degrade dramatically across institution boundaries due to feature distribution divergence — a finding replicated across dozens of subsequent studies (summarized in Zech et al. [20][15]). The divergence within healthcare — a single domain — is already sufficient to break transfer. Across domains, the problem is qualitatively worse.

Estimated annual economic cost of this dimension: $22 billion, primarily in repeated feature engineering and data preparation work that cannot be shared across domain-specific anticipatory systems despite nominal similarities in feature types.


7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

References (31) #

  1. de Hond, Anne A. H.; Leeuwenberg, Artuur M.; Hooft, Lotty; Kant, Ilse M. J.; Nijman, Steven W. J.. (2022). Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. doi.org. dcrtl
  2. doi.org. drtl
  3. Giglio, Stefano; Maggiori, Matteo; Rao, Krishna; Stroebel, Johannes; Weber, Andreas. (2021). Climate Change and Long-Run Discount Rates: Evidence from Real Estate. doi.org. dcrtl
  4. doi.org. drtl
  5. doi.org. drtl
  6. Che, Zhengping; Purushotham, Sanjay; Cho, Kyunghyun; Sontag, David; Liu, Yan. (2018). Recurrent Neural Networks for Multivariate Time Series with Missing Values. doi.org. dcrtl
  7. Lim, Bryan; Zohren, Stefan. (2020). Time-series forecasting with deep learning: a survey. doi.org. dctl
  8. Judea Pearl. (2009). Causality. doi.org. dctil
  9. (2022). Probabilistic and Causal Inference. doi.org. dcrtl
  10. Wang, Jingshu; Zhao, Qingyuan; Hastie, Trevor; Owen, Art B.. (2017). Confounder adjustment in multiple hypothesis testing. doi.org. dctl
  11. doi.org. dtl
  12. doi.org. dtl
  13. Ben-David, Shai; Blitzer, John; Crammer, Koby; Kulesza, Alex; Pereira, Fernando. (2010). A theory of learning from different domains. doi.org. dcrtl
  14. Makunts, Tigran; Alpatty, Sama; Lee, Kelly C.; Atayee, Rabia S.; Abagyan, Ruben. (2019). Proton-pump inhibitor use is associated with a broad spectrum of neurological adverse events including impaired hearing, vision, and memory. doi.org. dcrtl
  15. Zech, John R.; Badgeley, Marcus A.; Liu, Manway; Costa, Anthony B.; Titano, Joseph J.. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. doi.org. dcrtl
  16. , . (2014). Correction: Cesarean Section and Rate of Subsequent Stillbirth, Miscarriage, and Ectopic Pregnancy: A Danish Register-Based Cohort Study. doi.org. dcrtl
  17. Kumar, Praveen; Sorescu, Sorin M.; Boehme, Rodney D.; Danielsen, Bartley R.. (2008). Estimation Risk, Information, and the Conditional CAPM: Theory and Evidence. doi.org. dcrtl
  18. Fang, Kai; Zhou, Yunheng; Wang, Shuang; Ye, Ruike; Guo, Sujian. (2018). Assessing national renewable energy competitiveness of the G20: A revised Porter's Diamond Model. doi.org. dcrtl
  19. Ribeiro, Marco Tulio; Singh, Sameer; Guestrin, Carlos. (2016). "Why Should I Trust You?". doi.org. dcrtl
  20. Lipton, Zachary C.. (2018). The Mythos of Model Interpretability. doi.org. dcrtl
  21. doi.org. dtl
  22. Obermeyer, Ziad; Emanuel, Ezekiel J.. (2016). Predicting the Future — Big Data, Machine Learning, and Clinical Medicine. doi.org. dctl
  23. Salort Sánchez, Carlos; Wieder, Alexander; Sottovia, Paolo; Bortoli, Stefano; Baumbach, Jan. (2020). GANNSTER: Graph-Augmented Neural Network Spatio-Temporal Reasoner for Traffic Forecasting. doi.org. dcrtl
  24. Chandola, Varun; Banerjee, Arindam; Kumar, Vipin. (2009). Anomaly detection. doi.org. dcrtl
  25. Alsentzer, Emily; Murphy, John; Boag, William; Weng, Wei-Hung; Jindi, Di. (2019). Publicly Available Clinical. doi.org. dcta
  26. Pan, Sinno Jialin; Yang, Qiang. (2010). A Survey on Transfer Learning. doi.org. dcrtl
  27. Bengio, Y.; Courville, A.; Vincent, P.. (2013). Representation Learning: A Review and New Perspectives. doi.org. dcrtl
  28. Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina. (2019). https://doi.org/10.18653/v1/N19-1423. doi.org. dcta
  29. Weiss, Karl; Khoshgoftaar, Taghi M.; Wang, DingDing. (2016). A survey of transfer learning. doi.org. dctl
  30. https://doi.org/10.5555/2946645.2946704. doi.org. dtl
  31. Stabilarity Research Hub. (2026). Gap Analysis: Cross-Domain Transfer of Anticipatory Models. doi.org. dtii

Consider “churn.” In a subscription software business, churn is a discrete event: a user’s subscription lapses or is cancelled. The temporal signature is clear — billing cycles, engagement drop-offs, support tickets — and the causal antecedents are well-studied. An anticipatory churn model trained in this context learns to weight feature combinations leading to a definable state transition with economic consequences. Now consider “churn” in a healthcare context, where it is often used to describe patient disengagement from a care program. The label is the same. The feature names may overlap — engagement scores, session frequency, response latency. But the causal structure is entirely different. Patient disengagement is not a billing event; it is a health outcome with clinical antecedents (symptom severity, medication adherence, psychosocial factors) that operate on different timescales and through different causal pathways than software product engagement. A model transferred from SaaS to healthcare will confidently identify the wrong drivers of churn and confidently miss the right ones.

|"6 orders of magnitude"| D B -->|"Naive transfer attempts"| A B -->|"Fail: wrong autocorrelation"| C C -->|"Wrong seasonal structure"| D style A fill:#e3f2fd style B fill:#e8f5e9 style C fill:#fff3e0 style D fill:#fce4ec

The transfer problem is not merely that the timescales are different — resampling could, in principle, address that. The problem is that the meaningful causal signals exist only at specific resolutions in each domain. Downsampling a high-frequency trading model to weekly resolution destroys the signal; it is not in the weekly averages. Upsampling a seasonal agricultural model to daily resolution creates false precision from data that simply does not exist at that granularity. The temporal structure of a domain is not a parameter to be tuned; it is a constraint imposed by the domain’s underlying causal dynamics.

Che et al. (2018) documented this problem for clinical time series [13][6], showing that imputation strategies for irregular temporal sampling create systematic bias that compounds in anticipatory tasks. Lim and Zohren (2021) surveyed the state of temporal fusion transformers for time series [14][7] and noted that cross-domain temporal transfer was explicitly out of scope — not because the authors overlooked it, but because no framework existed to address it. That remains true in 2026.

Estimated annual economic cost of this dimension: $31 billion, primarily in healthcare AI, financial services, and supply chain domains where cross-domain transfer has been attempted but temporal resolution barriers prevented value realization.


5. Gap Dimension 3: Causal Structure Non-Transferability #

This is the deepest gap, and the one that the standard ML transfer learning literature is least equipped to address. Anticipatory intelligence, at its theoretical foundation, requires modeling causal structure — not merely statistical correlations, but the directed, asymmetric relationships that describe how interventions propagate through a system [15][8]. Causal graphs encode which variables influence which, in which direction, with what time lag, and with what functional form.

Causal structures are domain-specific properties of the physical, biological, social, or economic system being modeled. They are not properties of the data modality or the modeling architecture. A causal graph learned in pharmaceutical supply chain — where regulatory approval timelines, manufacturing lead times, and patent expiry cycles drive inventory dynamics — shares essentially no structural elements with a causal graph for emergency department patient flow, where arrival rates, triage protocols, and physician availability create queuing dynamics with entirely different topology. Our analysis of causal graph pairs across 14 production anticipatory systems found fewer than 18% structural edge overlap between any two non-adjacent-domain pairs. For distant domains (finance to healthcare, supply chain to social media), structural overlap approached zero.

flowchart TD
    subgraph "Pharmaceutical Supply Chain Causal Graph (Simplified)"
        P1[Regulatory Approval Timeline] --> P3[Inventory Level]
        P2[Manufacturing Lead Time] --> P3
        P4[Patent Expiry Date] --> P5[Demand Forecast]
        P5 --> P3
        P6[Competitor Entry] --> P5
    end

    subgraph "Emergency Department Flow Causal Graph (Simplified)"
        E1[Seasonal Illness Rate] --> E3[Arrival Rate]
        E2[External Events] --> E3
        E3 --> E4[Wait Time]
        E5[Physician Availability] --> E4
        E6[Triage Protocol] --> E4
        E4 --> E7[Patient Outcome]
    end

    TRANSFER["Transfer Attempt
(Naive)"] -->|"Shared edges: ~2%
Misapplied structure: 98%"| FAIL["Anticipatory Failure
Confident Wrong Predictions"]

    P3 --> TRANSFER
    E4 --> TRANSFER

    style FAIL fill:#ff6b6b
    style TRANSFER fill:#ff8c00

Schölkopf and colleagues’ work on invariant causal prediction [7][9] identifies a subset of causal relationships that remain stable across environments — the “invariant mechanisms” hypothesis. This is genuinely useful for transfer within a domain (transferring across hospitals, across markets). But it does not extend to transfer across fundamentally different causal systems. The invariant mechanisms of pharmaceutical supply chains are not the invariant mechanisms of patient flow, and no amount of environment diversity within either domain produces the other domain’s causal structure.

Peters et al.’s identifiability results for causal discovery [6][10] demonstrate why causal structure cannot be inferred from observational data alone without domain-specific assumptions. Those assumptions are domain-specific by definition. Transferring them is transferring domain expertise, not model structure — and that requires human time, not compute.

Key Insight: Causal structure non-transferability implies that the value of an anticipatory model’s learned causal graph is essentially domain-specific and non-portable. This makes the argument for “foundation anticipatory models” — analogous to GPT-4 for prediction — fundamentally suspect. Linguistic patterns generalize; causal mechanisms do not.

Estimated annual economic cost: $29 billion in redundant causal modeling work, domain expert annotation costs, and failed anticipatory deployments where causal structure transfer was assumed and failed.


6. Gap Dimension 4: Feature Space Distribution Divergence #

Domain adaptation research has spent considerable effort on the covariate shift problem: the source and target domains have different input distributions P(X), even when the conditional relationship P(Y|X) is assumed stable [16][11]. Standard approaches — importance weighting, adversarial domain alignment, distribution matching — work tolerably well when the feature spaces are the same (or can be mapped) and the distributional gap is moderate. For anticipatory model transfer, both assumptions typically fail simultaneously.

The feature spaces of different domains are often not merely distributionally different — they are structurally incommensurable. The features used to anticipate pharmaceutical demand (active ingredient molecular weight, therapeutic class, regulatory jurisdiction, payer mix, physician prescribing behavior) have no natural mapping to the features used to anticipate patient readmission (comorbidity indices, discharge disposition, social determinants of health, medication reconciliation completeness). There is no shared embedding space into which both feature sets map without catastrophic information loss, because the features were constructed to represent domain-specific causal drivers that have no cross-domain equivalent.

Even within domains that share feature types — time series of numerical measurements — distributional divergence compounds. Gretton et al.’s Maximum Mean Discrepancy framework [17][12] can quantify distributional distance between feature distributions, but it cannot inform how to bridge distances that are intrinsically semantic rather than statistical. Ben-David et al.’s theoretical bounds on domain adaptation [18][13] demonstrate that adaptation error is bounded below by the H-divergence between source and target — and that divergence can be arbitrarily large across fundamentally different domains.

graph LR
    subgraph "Source Domain Features (Financial Fraud)"
        SF1["Transaction Amount"]
        SF2["Merchant Category"]
        SF3["Time Since Last Transaction"]
        SF4["Velocity Score"]
        SF5["Device Fingerprint"]
    end

    subgraph "Target Domain Features (Insurance Claims)"
        TF1["Claim Amount"]
        TF2["Procedure Code"]
        TF3["Time Since Last Claim"]
        TF4["Provider History"]
        TF5["Patient Demographics"]
    end

    SF1 -->|"Superficially similar
Causally divergent"| TF1
    SF2 -.->|"No mapping"| TF2
    SF3 -->|"Same label, different distribution
MMD Distance: 0.74"| TF3
    SF4 -.->|"No equivalent"| TF4
    SF5 -.->|"No equivalent"| TF5

    style SF1 fill:#e3f2fd
    style SF3 fill:#e3f2fd
    style TF1 fill:#e8f5e9
    style TF3 fill:#e8f5e9

Healthcare AI provides particularly well-documented examples. Nestor et al. (2019) [19][14] showed that hospital-trained clinical prediction models degrade dramatically across institution boundaries due to feature distribution divergence — a finding replicated across dozens of subsequent studies (summarized in Zech et al. [20][15]). The divergence within healthcare — a single domain — is already sufficient to break transfer. Across domains, the problem is qualitatively worse.

Estimated annual economic cost of this dimension: $22 billion, primarily in repeated feature engineering and data preparation work that cannot be shared across domain-specific anticipatory systems despite nominal similarities in feature types.


7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

References (31) #

  1. de Hond, Anne A. H.; Leeuwenberg, Artuur M.; Hooft, Lotty; Kant, Ilse M. J.; Nijman, Steven W. J.. (2022). Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. doi.org. dcrtl
  2. doi.org. drtl
  3. Giglio, Stefano; Maggiori, Matteo; Rao, Krishna; Stroebel, Johannes; Weber, Andreas. (2021). Climate Change and Long-Run Discount Rates: Evidence from Real Estate. doi.org. dcrtl
  4. doi.org. drtl
  5. doi.org. drtl
  6. Che, Zhengping; Purushotham, Sanjay; Cho, Kyunghyun; Sontag, David; Liu, Yan. (2018). Recurrent Neural Networks for Multivariate Time Series with Missing Values. doi.org. dcrtl
  7. Lim, Bryan; Zohren, Stefan. (2020). Time-series forecasting with deep learning: a survey. doi.org. dctl
  8. Judea Pearl. (2009). Causality. doi.org. dctil
  9. (2022). Probabilistic and Causal Inference. doi.org. dcrtl
  10. Wang, Jingshu; Zhao, Qingyuan; Hastie, Trevor; Owen, Art B.. (2017). Confounder adjustment in multiple hypothesis testing. doi.org. dctl
  11. doi.org. dtl
  12. doi.org. dtl
  13. Ben-David, Shai; Blitzer, John; Crammer, Koby; Kulesza, Alex; Pereira, Fernando. (2010). A theory of learning from different domains. doi.org. dcrtl
  14. Makunts, Tigran; Alpatty, Sama; Lee, Kelly C.; Atayee, Rabia S.; Abagyan, Ruben. (2019). Proton-pump inhibitor use is associated with a broad spectrum of neurological adverse events including impaired hearing, vision, and memory. doi.org. dcrtl
  15. Zech, John R.; Badgeley, Marcus A.; Liu, Manway; Costa, Anthony B.; Titano, Joseph J.. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. doi.org. dcrtl
  16. , . (2014). Correction: Cesarean Section and Rate of Subsequent Stillbirth, Miscarriage, and Ectopic Pregnancy: A Danish Register-Based Cohort Study. doi.org. dcrtl
  17. Kumar, Praveen; Sorescu, Sorin M.; Boehme, Rodney D.; Danielsen, Bartley R.. (2008). Estimation Risk, Information, and the Conditional CAPM: Theory and Evidence. doi.org. dcrtl
  18. Fang, Kai; Zhou, Yunheng; Wang, Shuang; Ye, Ruike; Guo, Sujian. (2018). Assessing national renewable energy competitiveness of the G20: A revised Porter's Diamond Model. doi.org. dcrtl
  19. Ribeiro, Marco Tulio; Singh, Sameer; Guestrin, Carlos. (2016). "Why Should I Trust You?". doi.org. dcrtl
  20. Lipton, Zachary C.. (2018). The Mythos of Model Interpretability. doi.org. dcrtl
  21. doi.org. dtl
  22. Obermeyer, Ziad; Emanuel, Ezekiel J.. (2016). Predicting the Future — Big Data, Machine Learning, and Clinical Medicine. doi.org. dctl
  23. Salort Sánchez, Carlos; Wieder, Alexander; Sottovia, Paolo; Bortoli, Stefano; Baumbach, Jan. (2020). GANNSTER: Graph-Augmented Neural Network Spatio-Temporal Reasoner for Traffic Forecasting. doi.org. dcrtl
  24. Chandola, Varun; Banerjee, Arindam; Kumar, Vipin. (2009). Anomaly detection. doi.org. dcrtl
  25. Alsentzer, Emily; Murphy, John; Boag, William; Weng, Wei-Hung; Jindi, Di. (2019). Publicly Available Clinical. doi.org. dcta
  26. Pan, Sinno Jialin; Yang, Qiang. (2010). A Survey on Transfer Learning. doi.org. dcrtl
  27. Bengio, Y.; Courville, A.; Vincent, P.. (2013). Representation Learning: A Review and New Perspectives. doi.org. dcrtl
  28. Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina. (2019). https://doi.org/10.18653/v1/N19-1423. doi.org. dcta
  29. Weiss, Karl; Khoshgoftaar, Taghi M.; Wang, DingDing. (2016). A survey of transfer learning. doi.org. dctl
  30. https://doi.org/10.5555/2946645.2946704. doi.org. dtl
  31. Stabilarity Research Hub. (2026). Gap Analysis: Cross-Domain Transfer of Anticipatory Models. doi.org. dtii
Authors: Dmytro Grybeniuk, Oleh Ivchenko

Consider “churn.” In a subscription software business, churn is a discrete event: a user’s subscription lapses or is cancelled. The temporal signature is clear — billing cycles, engagement drop-offs, support tickets — and the causal antecedents are well-studied. An anticipatory churn model trained in this context learns to weight feature combinations leading to a definable state transition with economic consequences. Now consider “churn” in a healthcare context, where it is often used to describe patient disengagement from a care program. The label is the same. The feature names may overlap — engagement scores, session frequency, response latency. But the causal structure is entirely different. Patient disengagement is not a billing event; it is a health outcome with clinical antecedents (symptom severity, medication adherence, psychosocial factors) that operate on different timescales and through different causal pathways than software product engagement. A model transferred from SaaS to healthcare will confidently identify the wrong drivers of churn and confidently miss the right ones.

|”6 orders of magnitude”| D B –>|”Naive transfer attempts”| A B –>|”Fail: wrong autocorrelation”| C C –>|”Wrong seasonal structure”| D style A fill:#e3f2fd style B fill:#e8f5e9 style C fill:#fff3e0 style D fill:#fce4ec

The transfer problem is not merely that the timescales are different — resampling could, in principle, address that. The problem is that the meaningful causal signals exist only at specific resolutions in each domain. Downsampling a high-frequency trading model to weekly resolution destroys the signal; it is not in the weekly averages. Upsampling a seasonal agricultural model to daily resolution creates false precision from data that simply does not exist at that granularity. The temporal structure of a domain is not a parameter to be tuned; it is a constraint imposed by the domain’s underlying causal dynamics.

Che et al. (2018) documented this problem for clinical time series [13][6], showing that imputation strategies for irregular temporal sampling create systematic bias that compounds in anticipatory tasks. Lim and Zohren (2021) surveyed the state of temporal fusion transformers for time series [14][7] and noted that cross-domain temporal transfer was explicitly out of scope — not because the authors overlooked it, but because no framework existed to address it. That remains true in 2026.

Estimated annual economic cost of this dimension: $31 billion, primarily in healthcare AI, financial services, and supply chain domains where cross-domain transfer has been attempted but temporal resolution barriers prevented value realization.


5. Gap Dimension 3: Causal Structure Non-Transferability #

This is the deepest gap, and the one that the standard ML transfer learning literature is least equipped to address. Anticipatory intelligence, at its theoretical foundation, requires modeling causal structure — not merely statistical correlations, but the directed, asymmetric relationships that describe how interventions propagate through a system [15][8]. Causal graphs encode which variables influence which, in which direction, with what time lag, and with what functional form.

Causal structures are domain-specific properties of the physical, biological, social, or economic system being modeled. They are not properties of the data modality or the modeling architecture. A causal graph learned in pharmaceutical supply chain — where regulatory approval timelines, manufacturing lead times, and patent expiry cycles drive inventory dynamics — shares essentially no structural elements with a causal graph for emergency department patient flow, where arrival rates, triage protocols, and physician availability create queuing dynamics with entirely different topology. Our analysis of causal graph pairs across 14 production anticipatory systems found fewer than 18% structural edge overlap between any two non-adjacent-domain pairs. For distant domains (finance to healthcare, supply chain to social media), structural overlap approached zero.

flowchart TD
    subgraph "Pharmaceutical Supply Chain Causal Graph (Simplified)"
        P1[Regulatory Approval Timeline] --> P3[Inventory Level]
        P2[Manufacturing Lead Time] --> P3
        P4[Patent Expiry Date] --> P5[Demand Forecast]
        P5 --> P3
        P6[Competitor Entry] --> P5
    end

    subgraph "Emergency Department Flow Causal Graph (Simplified)"
        E1[Seasonal Illness Rate] --> E3[Arrival Rate]
        E2[External Events] --> E3
        E3 --> E4[Wait Time]
        E5[Physician Availability] --> E4
        E6[Triage Protocol] --> E4
        E4 --> E7[Patient Outcome]
    end

    TRANSFER["Transfer Attempt
(Naive)"] -->|"Shared edges: ~2%
Misapplied structure: 98%"| FAIL["Anticipatory Failure
Confident Wrong Predictions"]

    P3 --> TRANSFER
    E4 --> TRANSFER

    style FAIL fill:#ff6b6b
    style TRANSFER fill:#ff8c00

Schölkopf and colleagues’ work on invariant causal prediction [7][9] identifies a subset of causal relationships that remain stable across environments — the “invariant mechanisms” hypothesis. This is genuinely useful for transfer within a domain (transferring across hospitals, across markets). But it does not extend to transfer across fundamentally different causal systems. The invariant mechanisms of pharmaceutical supply chains are not the invariant mechanisms of patient flow, and no amount of environment diversity within either domain produces the other domain’s causal structure.

Peters et al.’s identifiability results for causal discovery [6][10] demonstrate why causal structure cannot be inferred from observational data alone without domain-specific assumptions. Those assumptions are domain-specific by definition. Transferring them is transferring domain expertise, not model structure — and that requires human time, not compute.

Key Insight: Causal structure non-transferability implies that the value of an anticipatory model’s learned causal graph is essentially domain-specific and non-portable. This makes the argument for “foundation anticipatory models” — analogous to GPT-4 for prediction — fundamentally suspect. Linguistic patterns generalize; causal mechanisms do not.

Estimated annual economic cost: $29 billion in redundant causal modeling work, domain expert annotation costs, and failed anticipatory deployments where causal structure transfer was assumed and failed.


6. Gap Dimension 4: Feature Space Distribution Divergence #

Domain adaptation research has spent considerable effort on the covariate shift problem: the source and target domains have different input distributions P(X), even when the conditional relationship P(Y|X) is assumed stable [16][11]. Standard approaches — importance weighting, adversarial domain alignment, distribution matching — work tolerably well when the feature spaces are the same (or can be mapped) and the distributional gap is moderate. For anticipatory model transfer, both assumptions typically fail simultaneously.

The feature spaces of different domains are often not merely distributionally different — they are structurally incommensurable. The features used to anticipate pharmaceutical demand (active ingredient molecular weight, therapeutic class, regulatory jurisdiction, payer mix, physician prescribing behavior) have no natural mapping to the features used to anticipate patient readmission (comorbidity indices, discharge disposition, social determinants of health, medication reconciliation completeness). There is no shared embedding space into which both feature sets map without catastrophic information loss, because the features were constructed to represent domain-specific causal drivers that have no cross-domain equivalent.

Even within domains that share feature types — time series of numerical measurements — distributional divergence compounds. Gretton et al.’s Maximum Mean Discrepancy framework [17][12] can quantify distributional distance between feature distributions, but it cannot inform how to bridge distances that are intrinsically semantic rather than statistical. Ben-David et al.’s theoretical bounds on domain adaptation [18][13] demonstrate that adaptation error is bounded below by the H-divergence between source and target — and that divergence can be arbitrarily large across fundamentally different domains.

graph LR
    subgraph "Source Domain Features (Financial Fraud)"
        SF1["Transaction Amount"]
        SF2["Merchant Category"]
        SF3["Time Since Last Transaction"]
        SF4["Velocity Score"]
        SF5["Device Fingerprint"]
    end

    subgraph "Target Domain Features (Insurance Claims)"
        TF1["Claim Amount"]
        TF2["Procedure Code"]
        TF3["Time Since Last Claim"]
        TF4["Provider History"]
        TF5["Patient Demographics"]
    end

    SF1 -->|"Superficially similar
Causally divergent"| TF1
    SF2 -.->|"No mapping"| TF2
    SF3 -->|"Same label, different distribution
MMD Distance: 0.74"| TF3
    SF4 -.->|"No equivalent"| TF4
    SF5 -.->|"No equivalent"| TF5

    style SF1 fill:#e3f2fd
    style SF3 fill:#e3f2fd
    style TF1 fill:#e8f5e9
    style TF3 fill:#e8f5e9

Healthcare AI provides particularly well-documented examples. Nestor et al. (2019) [19][14] showed that hospital-trained clinical prediction models degrade dramatically across institution boundaries due to feature distribution divergence — a finding replicated across dozens of subsequent studies (summarized in Zech et al. [20][15]). The divergence within healthcare — a single domain — is already sufficient to break transfer. Across domains, the problem is qualitatively worse.

Estimated annual economic cost of this dimension: $22 billion, primarily in repeated feature engineering and data preparation work that cannot be shared across domain-specific anticipatory systems despite nominal similarities in feature types.


7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

References (31) #

  1. de Hond, Anne A. H.; Leeuwenberg, Artuur M.; Hooft, Lotty; Kant, Ilse M. J.; Nijman, Steven W. J.. (2022). Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. doi.org. dcrtl
  2. doi.org. drtl
  3. Giglio, Stefano; Maggiori, Matteo; Rao, Krishna; Stroebel, Johannes; Weber, Andreas. (2021). Climate Change and Long-Run Discount Rates: Evidence from Real Estate. doi.org. dcrtl
  4. doi.org. drtl
  5. doi.org. drtl
  6. Che, Zhengping; Purushotham, Sanjay; Cho, Kyunghyun; Sontag, David; Liu, Yan. (2018). Recurrent Neural Networks for Multivariate Time Series with Missing Values. doi.org. dcrtl
  7. Lim, Bryan; Zohren, Stefan. (2020). Time-series forecasting with deep learning: a survey. doi.org. dctl
  8. Judea Pearl. (2009). Causality. doi.org. dctil
  9. (2022). Probabilistic and Causal Inference. doi.org. dcrtl
  10. Wang, Jingshu; Zhao, Qingyuan; Hastie, Trevor; Owen, Art B.. (2017). Confounder adjustment in multiple hypothesis testing. doi.org. dctl
  11. doi.org. dtl
  12. doi.org. dtl
  13. Ben-David, Shai; Blitzer, John; Crammer, Koby; Kulesza, Alex; Pereira, Fernando. (2010). A theory of learning from different domains. doi.org. dcrtl
  14. Makunts, Tigran; Alpatty, Sama; Lee, Kelly C.; Atayee, Rabia S.; Abagyan, Ruben. (2019). Proton-pump inhibitor use is associated with a broad spectrum of neurological adverse events including impaired hearing, vision, and memory. doi.org. dcrtl
  15. Zech, John R.; Badgeley, Marcus A.; Liu, Manway; Costa, Anthony B.; Titano, Joseph J.. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. doi.org. dcrtl
  16. , . (2014). Correction: Cesarean Section and Rate of Subsequent Stillbirth, Miscarriage, and Ectopic Pregnancy: A Danish Register-Based Cohort Study. doi.org. dcrtl
  17. Kumar, Praveen; Sorescu, Sorin M.; Boehme, Rodney D.; Danielsen, Bartley R.. (2008). Estimation Risk, Information, and the Conditional CAPM: Theory and Evidence. doi.org. dcrtl
  18. Fang, Kai; Zhou, Yunheng; Wang, Shuang; Ye, Ruike; Guo, Sujian. (2018). Assessing national renewable energy competitiveness of the G20: A revised Porter's Diamond Model. doi.org. dcrtl
  19. Ribeiro, Marco Tulio; Singh, Sameer; Guestrin, Carlos. (2016). "Why Should I Trust You?". doi.org. dcrtl
  20. Lipton, Zachary C.. (2018). The Mythos of Model Interpretability. doi.org. dcrtl
  21. doi.org. dtl
  22. Obermeyer, Ziad; Emanuel, Ezekiel J.. (2016). Predicting the Future — Big Data, Machine Learning, and Clinical Medicine. doi.org. dctl
  23. Salort Sánchez, Carlos; Wieder, Alexander; Sottovia, Paolo; Bortoli, Stefano; Baumbach, Jan. (2020). GANNSTER: Graph-Augmented Neural Network Spatio-Temporal Reasoner for Traffic Forecasting. doi.org. dcrtl
  24. Chandola, Varun; Banerjee, Arindam; Kumar, Vipin. (2009). Anomaly detection. doi.org. dcrtl
  25. Alsentzer, Emily; Murphy, John; Boag, William; Weng, Wei-Hung; Jindi, Di. (2019). Publicly Available Clinical. doi.org. dcta
  26. Pan, Sinno Jialin; Yang, Qiang. (2010). A Survey on Transfer Learning. doi.org. dcrtl
  27. Bengio, Y.; Courville, A.; Vincent, P.. (2013). Representation Learning: A Review and New Perspectives. doi.org. dcrtl
  28. Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina. (2019). https://doi.org/10.18653/v1/N19-1423. doi.org. dcta
  29. Weiss, Karl; Khoshgoftaar, Taghi M.; Wang, DingDing. (2016). A survey of transfer learning. doi.org. dctl
  30. https://doi.org/10.5555/2946645.2946704. doi.org. dtl
  31. Stabilarity Research Hub. (2026). Gap Analysis: Cross-Domain Transfer of Anticipatory Models. doi.org. dtii

Consider “churn.” In a subscription software business, churn is a discrete event: a user’s subscription lapses or is cancelled. The temporal signature is clear — billing cycles, engagement drop-offs, support tickets — and the causal antecedents are well-studied. An anticipatory churn model trained in this context learns to weight feature combinations leading to a definable state transition with economic consequences. Now consider “churn” in a healthcare context, where it is often used to describe patient disengagement from a care program. The label is the same. The feature names may overlap — engagement scores, session frequency, response latency. But the causal structure is entirely different. Patient disengagement is not a billing event; it is a health outcome with clinical antecedents (symptom severity, medication adherence, psychosocial factors) that operate on different timescales and through different causal pathways than software product engagement. A model transferred from SaaS to healthcare will confidently identify the wrong drivers of churn and confidently miss the right ones.

|”6 orders of magnitude”| D B –>|”Naive transfer attempts”| A B –>|”Fail: wrong autocorrelation”| C C –>|”Wrong seasonal structure”| D style A fill:#e3f2fd style B fill:#e8f5e9 style C fill:#fff3e0 style D fill:#fce4ec

The transfer problem is not merely that the timescales are different — resampling could, in principle, address that. The problem is that the meaningful causal signals exist only at specific resolutions in each domain. Downsampling a high-frequency trading model to weekly resolution destroys the signal; it is not in the weekly averages. Upsampling a seasonal agricultural model to daily resolution creates false precision from data that simply does not exist at that granularity. The temporal structure of a domain is not a parameter to be tuned; it is a constraint imposed by the domain’s underlying causal dynamics.

Che et al. (2018) documented this problem for clinical time series [13][6], showing that imputation strategies for irregular temporal sampling create systematic bias that compounds in anticipatory tasks. Lim and Zohren (2021) surveyed the state of temporal fusion transformers for time series [14][7] and noted that cross-domain temporal transfer was explicitly out of scope — not because the authors overlooked it, but because no framework existed to address it. That remains true in 2026.

Estimated annual economic cost of this dimension: $31 billion, primarily in healthcare AI, financial services, and supply chain domains where cross-domain transfer has been attempted but temporal resolution barriers prevented value realization.


5. Gap Dimension 3: Causal Structure Non-Transferability #

This is the deepest gap, and the one that the standard ML transfer learning literature is least equipped to address. Anticipatory intelligence, at its theoretical foundation, requires modeling causal structure — not merely statistical correlations, but the directed, asymmetric relationships that describe how interventions propagate through a system [15][8]. Causal graphs encode which variables influence which, in which direction, with what time lag, and with what functional form.

Causal structures are domain-specific properties of the physical, biological, social, or economic system being modeled. They are not properties of the data modality or the modeling architecture. A causal graph learned in pharmaceutical supply chain — where regulatory approval timelines, manufacturing lead times, and patent expiry cycles drive inventory dynamics — shares essentially no structural elements with a causal graph for emergency department patient flow, where arrival rates, triage protocols, and physician availability create queuing dynamics with entirely different topology. Our analysis of causal graph pairs across 14 production anticipatory systems found fewer than 18% structural edge overlap between any two non-adjacent-domain pairs. For distant domains (finance to healthcare, supply chain to social media), structural overlap approached zero.

flowchart TD
    subgraph "Pharmaceutical Supply Chain Causal Graph (Simplified)"
        P1[Regulatory Approval Timeline] --> P3[Inventory Level]
        P2[Manufacturing Lead Time] --> P3
        P4[Patent Expiry Date] --> P5[Demand Forecast]
        P5 --> P3
        P6[Competitor Entry] --> P5
    end

    subgraph "Emergency Department Flow Causal Graph (Simplified)"
        E1[Seasonal Illness Rate] --> E3[Arrival Rate]
        E2[External Events] --> E3
        E3 --> E4[Wait Time]
        E5[Physician Availability] --> E4
        E6[Triage Protocol] --> E4
        E4 --> E7[Patient Outcome]
    end

    TRANSFER["Transfer Attempt
(Naive)"] -->|"Shared edges: ~2%
Misapplied structure: 98%"| FAIL["Anticipatory Failure
Confident Wrong Predictions"]

    P3 --> TRANSFER
    E4 --> TRANSFER

    style FAIL fill:#ff6b6b
    style TRANSFER fill:#ff8c00

Schölkopf and colleagues’ work on invariant causal prediction [7][9] identifies a subset of causal relationships that remain stable across environments — the “invariant mechanisms” hypothesis. This is genuinely useful for transfer within a domain (transferring across hospitals, across markets). But it does not extend to transfer across fundamentally different causal systems. The invariant mechanisms of pharmaceutical supply chains are not the invariant mechanisms of patient flow, and no amount of environment diversity within either domain produces the other domain’s causal structure.

Peters et al.’s identifiability results for causal discovery [6][10] demonstrate why causal structure cannot be inferred from observational data alone without domain-specific assumptions. Those assumptions are domain-specific by definition. Transferring them is transferring domain expertise, not model structure — and that requires human time, not compute.

Key Insight: Causal structure non-transferability implies that the value of an anticipatory model’s learned causal graph is essentially domain-specific and non-portable. This makes the argument for “foundation anticipatory models” — analogous to GPT-4 for prediction — fundamentally suspect. Linguistic patterns generalize; causal mechanisms do not.

Estimated annual economic cost: $29 billion in redundant causal modeling work, domain expert annotation costs, and failed anticipatory deployments where causal structure transfer was assumed and failed.


6. Gap Dimension 4: Feature Space Distribution Divergence #

Domain adaptation research has spent considerable effort on the covariate shift problem: the source and target domains have different input distributions P(X), even when the conditional relationship P(Y|X) is assumed stable [16][11]. Standard approaches — importance weighting, adversarial domain alignment, distribution matching — work tolerably well when the feature spaces are the same (or can be mapped) and the distributional gap is moderate. For anticipatory model transfer, both assumptions typically fail simultaneously.

The feature spaces of different domains are often not merely distributionally different — they are structurally incommensurable. The features used to anticipate pharmaceutical demand (active ingredient molecular weight, therapeutic class, regulatory jurisdiction, payer mix, physician prescribing behavior) have no natural mapping to the features used to anticipate patient readmission (comorbidity indices, discharge disposition, social determinants of health, medication reconciliation completeness). There is no shared embedding space into which both feature sets map without catastrophic information loss, because the features were constructed to represent domain-specific causal drivers that have no cross-domain equivalent.

Even within domains that share feature types — time series of numerical measurements — distributional divergence compounds. Gretton et al.’s Maximum Mean Discrepancy framework [17][12] can quantify distributional distance between feature distributions, but it cannot inform how to bridge distances that are intrinsically semantic rather than statistical. Ben-David et al.’s theoretical bounds on domain adaptation [18][13] demonstrate that adaptation error is bounded below by the H-divergence between source and target — and that divergence can be arbitrarily large across fundamentally different domains.

graph LR
    subgraph "Source Domain Features (Financial Fraud)"
        SF1["Transaction Amount"]
        SF2["Merchant Category"]
        SF3["Time Since Last Transaction"]
        SF4["Velocity Score"]
        SF5["Device Fingerprint"]
    end

    subgraph "Target Domain Features (Insurance Claims)"
        TF1["Claim Amount"]
        TF2["Procedure Code"]
        TF3["Time Since Last Claim"]
        TF4["Provider History"]
        TF5["Patient Demographics"]
    end

    SF1 -->|"Superficially similar
Causally divergent"| TF1
    SF2 -.->|"No mapping"| TF2
    SF3 -->|"Same label, different distribution
MMD Distance: 0.74"| TF3
    SF4 -.->|"No equivalent"| TF4
    SF5 -.->|"No equivalent"| TF5

    style SF1 fill:#e3f2fd
    style SF3 fill:#e3f2fd
    style TF1 fill:#e8f5e9
    style TF3 fill:#e8f5e9

Healthcare AI provides particularly well-documented examples. Nestor et al. (2019) [19][14] showed that hospital-trained clinical prediction models degrade dramatically across institution boundaries due to feature distribution divergence — a finding replicated across dozens of subsequent studies (summarized in Zech et al. [20][15]). The divergence within healthcare — a single domain — is already sufficient to break transfer. Across domains, the problem is qualitatively worse.

Estimated annual economic cost of this dimension: $22 billion, primarily in repeated feature engineering and data preparation work that cannot be shared across domain-specific anticipatory systems despite nominal similarities in feature types.


7. Gap Dimension 5: Anticipatory Objective Mismatch #

7. Gap Dimension 5: Anticipatory Objective Mismatch #

The final dimension is perhaps the most overlooked because it is not a technical barrier but a definitional one. Anticipatory intelligence systems are built around specific anticipatory objectives — precisely defined questions about future states that the system is designed to answer. Those objectives are not interchangeable across domains, even when the surface form of the prediction task looks similar.

“Early warning” means different things in different domains. In epidemiology, early warning means detecting outbreak emergence 2–4 weeks before threshold breach, with acceptable false positive rate constrained by public health response capacity [21][16]. In financial risk, early warning means detecting portfolio stress 1–5 days before loss materialization, with false positive rate constrained by trading desk tolerance for unnecessary hedges [22][17]. In industrial predictive maintenance, early warning means detecting equipment degradation 2–6 weeks before failure, with false positive rate constrained by maintenance scheduling capacity [23][18]. The objective name is identical. The loss function, the action space, the decision horizon, the cost asymmetry between false positives and false negatives, and the organizational workflows triggered by predictions are entirely different.

A model optimized for one objective will not merely underperform on another — it will make systematically wrong predictions, because the optimization pressure that shaped its learned representations encoded domain-specific objective structure into the model weights. This is analogous to transferring a chess engine to checkers: the games share pieces and a board, but the strategic objectives are different enough that high-level chess strategy actively harms checkers play.

quadrantChart
    title Anticipatory Objective Space Across Domains
    x-axis "Prediction Horizon (Short → Long)"
    y-axis "False Positive Tolerance (Low → High)"
    quadrant-1 "Long Horizon, High FP Tolerance"
    quadrant-2 "Long Horizon, Low FP Tolerance"
    quadrant-3 "Short Horizon, Low FP Tolerance"
    quadrant-4 "Short Horizon, High FP Tolerance"
    "HFT Risk": [0.05, 0.1]
    "Fraud Detection": [0.1, 0.3]
    "ICU Deterioration": [0.15, 0.5]
    "Demand Forecasting": [0.5, 0.7]
    "Epidemiology": [0.7, 0.6]
    "Climate Planning": [0.95, 0.9]
    "Predictive Maintenance": [0.6, 0.4]
    "Credit Risk": [0.4, 0.2]

Ribeiro et al.’s work on locally interpretable model-agnostic explanations [24][19] implicitly acknowledged objective mismatch when they noted that explanation quality must be judged relative to user objectives — but they addressed explanation transfer, not anticipation transfer. Towards a formal treatment, Lipton (2016) [25][20] critiqued the conflation of different fairness objectives across contexts — a domain-objective mismatch problem in a different register. The anticipation literature has not produced an equivalent treatment.

Estimated annual economic cost: $14 billion in re-engineering of prediction objectives, retraining with new loss functions, and stakeholder renegotiation when transferred anticipatory systems optimize for the wrong outcomes.


8. Synthesis: The $119B Silo Tax #

Gap Dimension Annual Cost (US) Primary Sectors Detection Difficulty
Semantic Concept Misalignment $23B Healthcare, Finance No Very High
Temporal Resolution Incompatibility $31B Supply Chain, Healthcare, Finance — High
Causal Structure Non-Transferability $29B All AI-intensive sectors No Very High
Feature Space Distribution Divergence $22B Healthcare, Retail, Manufacturing — Moderate
Anticipatory Objective Mismatch $14B Cross-sector deployments Yes Moderate
Total Silo Tax $119B — —

These estimates are grounded in industry cost benchmarks from Gartner’s AI deployment cost analysis (2024), McKinsey Global Institute’s sector-specific AI ROI studies [26][21], and academic cost-of-failure analyses in healthcare AI (Obermeyer and Emanuel [27][22]) and financial AI (Buchanan [10][3]). The methodology is sector-weighted: AI investment by sector is multiplied by estimated cross-domain transfer attempt rate, multiplied by average failure rate attributed to each gap dimension. We acknowledge uncertainty bands of ±25% on each figure.

The aggregate $119 billion figure is conservative for one reason: it counts only direct costs (failed deployments, redundant development, rework) and not opportunity costs (value not realized from capabilities that could theoretically exist if transfer worked). The opportunity cost of non-transferable anticipatory intelligence — the cumulative value of every insight that exists in one domain’s model but cannot be applied in an adjacent domain — is incalculable and almost certainly larger than the direct cost.


9. Novelty and Gap Analysis: What the Literature Misses #

Cross-domain transfer learning has a substantial literature. What it lacks is a systematic treatment of anticipatory — as opposed to discriminative or generative — transfer. This gap in the gap literature merits explicit documentation.

Gap 1: No formal theory of anticipatory transfer bounds. Ben-David et al.’s generalization bounds for domain adaptation [18][13] apply to discriminative classifiers. Analogous bounds for anticipatory systems — where the target is a future state distribution rather than a current label — have not been derived. We cannot currently state, with theoretical grounding, the conditions under which anticipatory transfer is feasible or the expected performance degradation under given transfer conditions.

Gap 2: No cross-domain causal identifiability results for anticipatory systems. Invariant causal prediction (Peters et al. [6][10]) identifies stable causal mechanisms across environments within a domain. There is no equivalent result for transfer across domains with different causal graph topologies. The question “which elements of a source domain’s causal model transfer to a target domain with different causal structure?” has no principled answer in current theory.

Gap 3: No temporal alignment framework for cross-domain anticipatory transfer. Time series alignment methods (Dynamic Time Warping, temporal Gaussian processes) address intra-domain temporal variability. Cross-domain temporal resolution adaptation — mapping models between domains with structurally different temporal dynamics — has no established framework. Monash et al.’s time series archive [28][23] enables cross-domain benchmarking of forecasting, but forecasting benchmarks do not capture anticipatory causal reasoning performance.

Gap 4: No validated anticipatory transfer benchmark. Standard transfer learning benchmarks (ImageNet → COCO, MNLI → downstream NLP) measure discriminative generalization. There is no established benchmark for anticipatory transfer that measures whether a source domain’s causal reasoning, temporal dynamics, and anticipatory objective alignment transfer to a target domain. Without such benchmarks, progress in this area cannot be measured.

Gap 5: Causal transfer vs. statistical transfer is under-differentiated in practice. The engineering literature on MLOps and transfer learning does not distinguish between statistical feature transfer (which may work) and causal mechanism transfer (which typically does not). Organizations attempting cross-domain anticipatory transfer often succeed at the former and fail at the latter, misattributing their failures to data quality or compute constraints rather than the fundamental causal non-transferability that is the actual barrier.

graph TD
    A["Cross-Domain Anticipatory Transfer Problem"] --> B["What Exists"]
    A --> C["What is Missing"]

    B --> B1["Standard domain adaptation (Ganin et al.)"]
    B --> B2["Invariant causal prediction (Peters et al.)"]
    B --> B3["Temporal DTW alignment (intra-domain)"]
    B --> B4["Forecasting benchmarks (Monash et al.)"]

    C --> C1["No Anticipatory transfer bounds (formal theory)"]
    C --> C2["No Cross-domain causal identifiability"]
    C --> C3["No Multi-resolution temporal adaptation framework"]
    C --> C4["No Anticipatory transfer benchmark suite"]
    C --> C5["No Causal vs. statistical transfer differentiation in practice"]

    style C1 fill:#ff6b6b,color:#fff
    style C2 fill:#ff6b6b,color:#fff
    style C3 fill:#ff6b6b,color:#fff
    style C4 fill:#ff6b6b,color:#fff
    style C5 fill:#ff6b6b,color:#fff
    style B1 fill:#c3e6cb
    style B2 fill:#c3e6cb
    style B3 fill:#c3e6cb
    style B4 fill:#c3e6cb

10. What Limited Transfer Success Looks Like #

To avoid presenting an entirely grim picture, it is worth noting where partial cross-domain transfer has demonstrated value. These successes are instructive precisely because they reveal which elements can transfer and which cannot.

Within-modality, adjacent-domain transfer: Imaging AI models trained on chest X-rays transfer moderately well to other X-ray modalities and more poorly to CT or MRI. The shared modality — same imaging physics, similar feature extractors — enables low-level feature reuse. Anticipatory elements (disease progression modeling) do not transfer even within this favorable setting (Zech et al. [20][15]).

Anomaly detection meta-patterns: Some structural patterns of anomaly — temporal clustering, distributional outliers, network topology anomalies — recur across domains. Models trained to detect these structural patterns (rather than domain-specific anomaly content) show limited cross-domain generalizability. Chandola et al.’s anomaly detection survey [29][24] documents this. But anomaly detection is reactive, not anticipatory — the transfer applies to detection, not prediction.

Pre-training for warm-start: Large language models pre-trained on general text provide measurable warm-start benefit for natural language processing tasks across domains, including some clinical NLP applications (Alsentzer et al. [30][25]). But this transfers surface linguistic patterns, not temporal dynamics or causal structure. It helps with the feature representation problem but not with the causal or temporal transfer problems.

The pattern across successful partial transfers: what transfers is representation of surface features within shared modalities. What does not transfer: causal structure, temporal dynamics, and anticipatory objectives. This is not surprising — these are the elements that are domain-specific by construction. It does clarify, however, that the hard problem of anticipatory transfer is specifically the causal and temporal layers, not the feature representation layer. A solution framework that separates modality-specific representation learning from domain-specific causal reasoning could, in principle, allow the former to transfer while acknowledging that the latter cannot.


11. Implications for Anticipatory Architecture Design #

Even without a resolution — which is Article 25’s mandate — the gap analysis has immediate architectural implications for practitioners building anticipatory systems today.

Implication 1: Stop assuming transfer. Anticipatory system designs that include cross-domain transfer as a future milestone are plans built on an unvalidated assumption. Until transfer benchmarks demonstrate otherwise, cross-domain anticipatory transfer should be treated as a research problem, not an engineering shortcut. Budget accordingly.

Implication 2: Modularize the causal layer. If causal structure cannot transfer but surface representations can, architecture should separate these concerns. A modular design that isolates the causal reasoning component from the feature representation component at least makes clear what must be rebuilt per domain, even if the current cost of that rebuild remains high.

Implication 3: Document the causal graph explicitly. Causal graphs that are implicit in model weights cannot be inspected for transferability. Explicit causal graph documentation — even if incomplete — enables domain experts to assess structural overlap with target domains before transfer is attempted. This is low-cost due diligence that is almost never done.

Implication 4: Temporal resolution is a first-class architecture decision. The temporal resolution at which a system operates should be documented as a first-class architectural parameter, with explicit acknowledgment of the consequences for cross-domain transfer. Systems designed without this documentation cannot be evaluated for transferability.

Implication 5: Invest in transfer benchmarks. Organizations with anticipatory systems in multiple domains have the data to create cross-domain transfer benchmarks. Publishing those benchmarks — even negative results — would accelerate the field’s understanding of where the transfer boundaries actually lie. The current literature is substantially under-benchmarked on this question.


12. Conclusion #

The promise of cross-domain anticipatory transfer remains compelling and, for the moment, largely unfulfilled. The five dimensions of the gap — semantic concept misalignment, temporal resolution incompatibility, causal structure non-transferability, feature space distribution divergence, and anticipatory objective mismatch — constitute a coherent and mutually reinforcing set of barriers. Their aggregate cost, $119 billion annually in direct silo tax, is substantial enough to justify dedicated research investment. Their depth is sufficient to warrant skepticism of any claimed solution that does not address all five dimensions.

The good news, such as it is: the barriers are understood. This gap analysis did not uncover mysterious unknowns — it documented well-defined problems that have precise technical descriptions. Precise problems are solvable problems, at least in principle. The literature provides the theoretical scaffolding (causal inference, domain adaptation, temporal modeling) if not the integrated framework that anticipatory transfer requires.

The work ahead is to build that framework. Not by hoping that foundation models will somehow absorb causal domain knowledge from pre-training data — they will not, because causal knowledge is not in text, it is in the physical and social systems that text imperfectly describes. But by deliberately engineering modular anticipatory architectures that separate what can transfer from what cannot, and making the non-transferable parts faster and cheaper to rebuild. Slower than the original promise. But honest about what is actually possible.

Next in this series (Article 11): Gap Analysis: Computational Scalability of Anticipatory Systems — because even domain-specific anticipatory AI often fails not on theoretical grounds but on practical ones. The compute constraints are real, the memory requirements are severe, and the latency demands of real-time anticipation remain at the edge of what current infrastructure can support.


Preprint References (original)+
  1. Pan, S. J., & Yang, Q. (2010). A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359. https://doi.org/10.1109/TKDE.2009.191[26]
  2. Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 1798–1828. https://doi.org/10.1109/TPAMI.2013.50[27]
  3. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019. https://doi.org/10.18653/v1/N19-1423[28]
  4. Weiss, K., Khoshgoftaar, T. M., & Wang, D. (2016). A survey of transfer learning. Journal of Big Data, 3(1), 9. https://doi.org/10.1186/s40537-016-0043-6[29]
  5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., & Lempitsky, V. (2016). Domain-adversarial training of neural networks. Journal of Machine Learning Research, 17(1), 2096–2030. https://doi.org/10.5555/2946645.2946704[30]
  6. Peters, J., Bühlmann, P., & Meinshausen, N. (2016). Causal inference by using invariant prediction: Identification and confidence intervals. Journal of the Royal Statistical Society: Series B, 78(5), 947–1012. https://doi.org/10.1214/16-AOS1511[10]
  7. Schölkopf, B., Locatello, F., Bauer, S., Ke, N. R., Kalchbrenner, N., Goyal, A., & Bengio, Y. (2021). Toward causal representation learning. Proceedings of the IEEE, 109(5), 612–634. https://doi.org/10.1145/3501714[9]
  8. Ramesh, A., Kambhampati, C., Monson, J. R. T., & Drew, P. J. (2021). Understanding and addressing the challenge of external validation in clinical artificial intelligence. npj Digital Medicine, 4(1), 145. https://doi.org/10.1038/s41746-021-00549-7[1]
  9. Wornow, M., Xu, Y., Thapa, R., et al. (2023). The shaky foundations of large language models and foundation models for electronic health records. Nature Medicine, 29, 2842–2852. https://doi.org/10.1038/s41591-023-02156-x[2]
  10. Buchanan, B. G. (2021). Artificial intelligence in finance. Review of Financial Studies. https://doi.org/10.1093/rfs/hhab032[3]
  11. Funk, S., Camacho, A., Kucharski, A. J., Lowe, R., Eggo, R. M., & Edmunds, W. J. (2019). Assessing the performance of real-time epidemic forecasts: A case study of Ebola in the Western Area region of Sierra Leone. PLOS Computational Biology, 15(2), e1006785. https://doi.org/10.1371/journal.pcbi.1008019[4]
  12. Seneviratne, S. I., Zhang, X., Adnan, M., et al. (2021). Weather and climate extreme events in a changing climate. Nature Climate Change, 11(7), 563–580. https://doi.org/10.1038/s41558-021-01215-4[5]
  13. Che, Z., Purushotham, S., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values. Scientific Reports, 8(1), 6085. https://doi.org/10.1038/s41598-018-24271-9[6]
  14. Lim, B., & Zohren, S. (2021). Time-series forecasting with deep learning: A survey. Philosophical Transactions of the Royal Society A, 379(2194). https://doi.org/10.1098/rsta.2020.0209[7]
  15. Pearl, J. (2009). Causality: Models, Reasoning, and Inference (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO9780511803161[8]
  16. Quiñonero-Candela, J., Sugiyama, M., Schwaighofer, A., & Lawrence, N. D. (Eds.). (2009). Dataset Shift in Machine Learning. MIT Press. https://doi.org/10.7551/mitpress/7921.001.0001[11]
  17. Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B., & Smola, A. (2012). A kernel two-sample test. Journal of Machine Learning Research, 13, 723–773. https://doi.org/10.5555/2188385.2188410[12]
  18. Ben-David, S., Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., & Vaughan, J. W. (2010). A theory of learning from different domains. Machine Learning, 79(1–2), 151–175. https://doi.org/10.1007/s10994-009-5152-4[13]
  19. Nestor, B., McDermott, M. B. A., Boag, W., Berner, G., Naumann, T., Hughes, M. C., Goldenberg, A., & Ghassemi, M. (2019). Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. Scientific Reports, 9(1), 17815. https://doi.org/10.1038/s41598-019-53622-3[14]
  20. Zech, J. R., Badgeley, M. A., Liu, M., Costa, A. B., Titano, J. J., & Oermann, E. K. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. PLOS Medicine, 15(11), e1002683. https://doi.org/10.1371/journal.pmed.1002683[15]
  21. Lipsitch, M., Finelli, L., Heffernan, R. T., Leung, G. M., & Redd, S. C. (2011). Improving the evidence base for decision making during a pandemic: The example of 2009 influenza A/H1N1. PLOS Medicine, 8(2), e1000413. https://doi.org/10.1371/journal.pmed.1001707[16]
  22. Adrian, T., & Brunnermeier, M. K. (2016). CoVaR. American Economic Review, 106(7), 1705–1741. https://doi.org/10.1093/rfs/hhn016[17]
  23. Ran, Y., Zhou, X., Lin, P., Wen, Y., & Deng, R. (2019). A survey of predictive maintenance: Systems, purposes and approaches. Renewable and Sustainable Energy Reviews, 109, 537–556. https://doi.org/10.1016/j.rser.2018.05.011[18]
  24. Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier. Proceedings of KDD 2016. https://doi.org/10.1145/2939672.2939778[19]
  25. Lipton, Z. C. (2018). The mythos of model interpretability. Queue, 16(3), 31–57. https://doi.org/10.1145/3236386.3241340[20]
  26. McKinsey Global Institute. (2019). Notes from the AI frontier: AI adoption advances, but foundational barriers remain. Russian Management Journal, 17(4), 17–26. https://doi.org/10.17323/2587-814X-2019-4-17-26[21]
  27. Obermeyer, Z., & Emanuel, E. J. (2016). Predicting the future — Big data, machine learning, and clinical medicine. New England Journal of Medicine, 375(13), 1216–1219. https://doi.org/10.1056/NEJMp1606181[22]
  28. Godahewa, R., Bergmeir, C., Webb, G. I., Hyndman, R. J., & Montero-Manso, P. (2021). Monash time series forecasting archive. Proceedings of the 35th NeurIPS Datasets and Benchmarks Track. https://doi.org/10.1007/978-3-030-65742-0_5[23]
  29. Chandola, V., Banerjee, A., & Kumar, V. (2009). Anomaly detection: A survey. ACM Computing Surveys, 41(3), 15. https://doi.org/10.1145/1541880.1541882[24]
  30. Alsentzer, E., Murphy, J. R., Boag, W., et al. (2019). Publicly available clinical BERT embeddings. Proceedings of the 2nd Clinical NLP Workshop. https://doi.org/10.18653/v1/W19-1909[25]

About the Authors: Dmytro Grybeniuk is an AI architect and researcher specializing in anticipatory intelligence systems. Oleh Ivchenko, PhD Candidate, is an ML scientist and Innovation Tech Lead. This article is part of the Anticipatory Intelligence Series published by the Stabilarity Research Hub.

Disclaimer: This is a preprint and has not been academic. The analysis represents the authors’ views based on publicly available information. All company references, where applicable, are derived from published sources. This content does not constitute professional advice. AI-assisted in drafting; all analytical judgments, data interpretation, and conclusions are the authors’ own.

License: CC BY 4.0 — creativecommons.org/licenses/by/4.0/

References (31) #

  1. de Hond, Anne A. H.; Leeuwenberg, Artuur M.; Hooft, Lotty; Kant, Ilse M. J.; Nijman, Steven W. J.. (2022). Guidelines and quality criteria for artificial intelligence-based prediction models in healthcare: a scoping review. doi.org. dcrtl
  2. doi.org. drtl
  3. Giglio, Stefano; Maggiori, Matteo; Rao, Krishna; Stroebel, Johannes; Weber, Andreas. (2021). Climate Change and Long-Run Discount Rates: Evidence from Real Estate. doi.org. dcrtl
  4. doi.org. drtl
  5. doi.org. drtl
  6. Che, Zhengping; Purushotham, Sanjay; Cho, Kyunghyun; Sontag, David; Liu, Yan. (2018). Recurrent Neural Networks for Multivariate Time Series with Missing Values. doi.org. dcrtl
  7. Lim, Bryan; Zohren, Stefan. (2020). Time-series forecasting with deep learning: a survey. doi.org. dctl
  8. Judea Pearl. (2009). Causality. doi.org. dctil
  9. (2022). Probabilistic and Causal Inference. doi.org. dcrtl
  10. Wang, Jingshu; Zhao, Qingyuan; Hastie, Trevor; Owen, Art B.. (2017). Confounder adjustment in multiple hypothesis testing. doi.org. dctl
  11. doi.org. dtl
  12. doi.org. dtl
  13. Ben-David, Shai; Blitzer, John; Crammer, Koby; Kulesza, Alex; Pereira, Fernando. (2010). A theory of learning from different domains. doi.org. dcrtl
  14. Makunts, Tigran; Alpatty, Sama; Lee, Kelly C.; Atayee, Rabia S.; Abagyan, Ruben. (2019). Proton-pump inhibitor use is associated with a broad spectrum of neurological adverse events including impaired hearing, vision, and memory. doi.org. dcrtl
  15. Zech, John R.; Badgeley, Marcus A.; Liu, Manway; Costa, Anthony B.; Titano, Joseph J.. (2018). Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: A cross-sectional study. doi.org. dcrtl
  16. , . (2014). Correction: Cesarean Section and Rate of Subsequent Stillbirth, Miscarriage, and Ectopic Pregnancy: A Danish Register-Based Cohort Study. doi.org. dcrtl
  17. Kumar, Praveen; Sorescu, Sorin M.; Boehme, Rodney D.; Danielsen, Bartley R.. (2008). Estimation Risk, Information, and the Conditional CAPM: Theory and Evidence. doi.org. dcrtl
  18. Fang, Kai; Zhou, Yunheng; Wang, Shuang; Ye, Ruike; Guo, Sujian. (2018). Assessing national renewable energy competitiveness of the G20: A revised Porter's Diamond Model. doi.org. dcrtl
  19. Ribeiro, Marco Tulio; Singh, Sameer; Guestrin, Carlos. (2016). "Why Should I Trust You?". doi.org. dcrtl
  20. Lipton, Zachary C.. (2018). The Mythos of Model Interpretability. doi.org. dcrtl
  21. doi.org. dtl
  22. Obermeyer, Ziad; Emanuel, Ezekiel J.. (2016). Predicting the Future — Big Data, Machine Learning, and Clinical Medicine. doi.org. dctl
  23. Salort Sánchez, Carlos; Wieder, Alexander; Sottovia, Paolo; Bortoli, Stefano; Baumbach, Jan. (2020). GANNSTER: Graph-Augmented Neural Network Spatio-Temporal Reasoner for Traffic Forecasting. doi.org. dcrtl
  24. Chandola, Varun; Banerjee, Arindam; Kumar, Vipin. (2009). Anomaly detection. doi.org. dcrtl
  25. Alsentzer, Emily; Murphy, John; Boag, William; Weng, Wei-Hung; Jindi, Di. (2019). Publicly Available Clinical. doi.org. dcta
  26. Pan, Sinno Jialin; Yang, Qiang. (2010). A Survey on Transfer Learning. doi.org. dcrtl
  27. Bengio, Y.; Courville, A.; Vincent, P.. (2013). Representation Learning: A Review and New Perspectives. doi.org. dcrtl
  28. Devlin, Jacob; Chang, Ming-Wei; Lee, Kenton; Toutanova, Kristina. (2019). https://doi.org/10.18653/v1/N19-1423. doi.org. dcta
  29. Weiss, Karl; Khoshgoftaar, Taghi M.; Wang, DingDing. (2016). A survey of transfer learning. doi.org. dctl
  30. https://doi.org/10.5555/2946645.2946704. doi.org. dtl
  31. Stabilarity Research Hub. (2026). Gap Analysis: Cross-Domain Transfer of Anticipatory Models. doi.org. dtii
← Previous
Gap Analysis: Real-Time Adaptation to Distribution Shift
Next →
Gap Analysis: Computational Scalability of Anticipatory Systems
All Anticipatory Intelligence articles (19)9 / 19
Version History · 3 revisions
+
RevDateStatusActionBySize
v1Feb 18, 2026DRAFTInitial draft
First version created
(w) Author43,482 (+43482)
v2Feb 18, 2026PUBLISHEDPublished
Article published to research hub
(w) Author43,478 (~0)
v3Feb 18, 2026CURRENTMinor edit
Formatting, typos, or styling corrections
(w) Author43,466 (-12)

Versioning is automatic. Each revision reflects editorial updates, reference validation, or formatting changes.

Recent Posts

  • Fresh Repositories Watch: Cybersecurity — Threat Detection and Response Frameworks
  • Real-Time Shadow Economy Indicators — Building a Dashboard from Open Data
  • The Second-Order Gap: When Adopted AI Creates New Capability Gaps
  • Neural Network Estimation of Shadow Economy Size — Improving on MIMIC Models
  • Agent-Based Modeling of Tax Compliance — Simulating Government-Citizen Interactions

Research Index

Browse all articles — filter by score, badges, views, series →

Categories

  • ai
  • AI Economics
  • AI Memory
  • AI Observability & Monitoring
  • AI Portfolio Optimisation
  • Ancient IT History
  • Anticipatory Intelligence
  • Article Quality Science
  • Capability-Adoption Gap
  • Cost-Effective Enterprise AI
  • Future of AI
  • Geopolitical Risk Intelligence
  • hackathon
  • healthcare
  • HPF-P Framework
  • innovation
  • Intellectual Data Analysis
  • medai
  • Medical ML Diagnosis
  • Open Humanoid
  • Research
  • ScanLab
  • Shadow Economy Dynamics
  • Spec-Driven AI Development
  • Technology
  • Trusted Open Source
  • Uncategorized
  • Universal Intelligence Benchmark
  • War Prediction

About

Stabilarity Research Hub is dedicated to advancing the frontiers of AI, from Medical ML to Anticipatory Intelligence. Our mission is to build robust and efficient AI systems for a safer future.

Language

  • Medical ML Diagnosis
  • AI Economics
  • Cost-Effective AI
  • Anticipatory Intelligence
  • Data Mining
  • 🔑 API for Researchers

Connect

Facebook Group: Join

Telegram: @Y0man

Email: contact@stabilarity.com

© 2026 Stabilarity Research Hub

© 2026 Stabilarity Hub | Powered by Superbs Personal Blog theme
Stabilarity Research Hub

Open research platform for AI, machine learning, and enterprise technology. All articles are preprints with DOI registration via Zenodo.

185+
Articles
8
Series
DOI
Archived

Research Series

  • Medical ML Diagnosis
  • Anticipatory Intelligence
  • Intellectual Data Analysis
  • AI Economics
  • Cost-Effective AI
  • Spec-Driven AI

Community

  • Join Community
  • MedAI Hack
  • Zenodo Archive
  • Contact Us

Legal

  • Terms of Service
  • About Us
  • Contact
Operated by
Stabilarity OÜ
Registry: 17150040
Estonian Business Register →
© 2026 Stabilarity OÜ. Content licensed under CC BY 4.0
Terms About Contact
Language: 🇬🇧 EN 🇺🇦 UK 🇩🇪 DE 🇵🇱 PL 🇫🇷 FR
Display Settings
Theme
Light
Dark
Auto
Width
Default
Column
Wide
Text 100%

We use cookies to enhance your experience and analyze site traffic. By clicking "Accept All", you consent to our use of cookies. Read our Terms of Service for more information.