Chapter 13: Emerging Frontiers in Data Mining (2024–2026) #
By Iryna Ivchenko & Oleh Ivchenko | Stabilarity Hub | February 2026
Opening Narrative: The Acceleration #
In October 2024, a team at Google DeepMind published AlphaFold 3, demonstrating protein structure prediction capabilities that surpassed experimental methods in both accuracy and speed. What made this achievement remarkable was not merely the scientific breakthrough but the methodology: the same transformer architecture underlying ChatGPT, adapted for molecular biology, solved a problem that had consumed decades of specialized research. The generalization of foundation models beyond language represented a paradigm shift—intelligence learned on one task transferring to fundamentally different domains.
Simultaneously, researchers at MIT demonstrated federated learning systems enabling hospitals to collaboratively train disease prediction models without sharing patient data, achieving accuracy comparable to centralized training while satisfying HIPAA privacy requirements. Regulatory impossibility had become technical reality.
In manufacturing, Siemens deployed neural architecture search systems that automatically designed predictive maintenance models outperforming human-engineered solutions, reducing downtime by 40% while eliminating the months-long expert modeling process. Automation was automating itself.
These developments, all occurring within 2024-2026, represent not incremental improvements but fundamental shifts in how data mining operates. This chapter surveys the emerging frontiers reshaping intelligent data analysis: automated machine learning reaching human-competitive performance, foundation models revolutionizing tabular and time-series data, privacy-preserving techniques enabling previously impossible collaborations, and real-time streaming systems processing billions of events per second. We examine which innovations represent genuine paradigm shifts versus sophisticated extensions of existing methods, and project their impact on the future of data mining.
Abstract #
This chapter surveys cutting-edge data mining techniques emerging between 2024-2026, distinguishing transformative innovations from incremental improvements. We examine five frontier areas: (1) AutoML systems achieving expert-level performance through neural architecture search and meta-learning, (2) foundation models for tabular data adapting large language model techniques to structured datasets, (3) privacy-preserving mining enabling federated learning and differential privacy at scale, (4) real-time streaming analytics processing infinite data with bounded memory, and (5) causal discovery methods inferring interventional relationships from observational data. For each frontier, we analyze underlying principles, current capabilities, persistent limitations, and research trajectories. We conclude by assessing which techniques represent genuine paradigm shifts likely to reshape data mining practice versus sophisticated refinements of established approaches.
Keywords: AutoML, neural architecture search, foundation models, tabular transformers, federated learning, differential privacy, streaming analytics, causal inference, emerging techniques, data mining innovation
1. Introduction: Identifying True Frontiers #
Data mining evolves continuously, but not all innovations matter equally. Distinguishing genuine paradigm shifts from incremental improvements[2] requires examining whether new techniques fundamentally alter what is possible versus merely optimizing existing approaches.
True frontiers exhibit three characteristics:
- Capability Expansion: Enabling previously impossible tasks (e.g., privacy-preserving collaborative learning)
- Efficiency Revolution: Achieving order-of-magnitude improvements in speed, accuracy, or resource consumption
- Accessibility Transformation: Democratizing capabilities previously requiring deep expertise
This chapter focuses on five frontiers meeting these criteria, examining their foundations, current state, and trajectories. We ground analysis in recent literature (2024-2026) while connecting to established theoretical frameworks.
graph TD
A[Emerging Frontiers 2024-2026] --> B[AutoML & NAS]
A --> C[Foundation Models]
A --> D[Privacy-Preserving Mining]
A --> E[Real-Time Streaming]
A --> F[Causal Discovery]
B --> B1[Neural Architecture Search]
B --> B2[Meta-Learning]
B --> B3[Hyperparameter Optimization]
C --> C1[Tabular Transformers]
C --> C2[Time-Series Foundation Models]
C --> C3[Transfer Learning for Structured Data]
D --> D1[Federated Learning]
D --> D2[Differential Privacy]
D --> D3[Secure Multi-Party Computation]
E --> E1[Online Learning]
E --> E2[Concept Drift Detection]
E --> E3[Approximate Algorithms]
F --> F1[Structural Causal Models]
F --> F2[Interventional Inference]
F --> F3[Causal Effect Estimation]
style A fill:#e1f5fe
style B fill:#fff9c4
style C fill:#c8e6c9
style D fill:#b2dfdb
style E fill:#ffccbc
style F fill:#f8bbd0
Figure 1: Taxonomy of Emerging Frontiers in Data Mining
2. Frontier #1: AutoML and Neural Architecture Search #
Automated Machine Learning (AutoML) aims to automate the end-to-end process of applying machine learning, from data preprocessing through model selection to hyperparameter tuning. Recent advances have transformed AutoML from research curiosity to production reality.
2.1 Neural Architecture Search (NAS) #
Neural Architecture Search[3], introduced by Zoph and Le (2017), automates the design of neural network architectures. Early implementations required thousands of GPU-hours, but recent innovations have achieved dramatic efficiency improvements.
Efficient NAS Methods (2024-2026):
- Differentiable NAS (DARTS): Formulates architecture search as continuous optimization[4], reducing search time from days to hours. Recent extensions (2024)[5] incorporate hardware awareness, optimizing for specific deployment targets.
- Weight Sharing: One-shot NAS methods[6] train a supernet containing all candidate architectures, then extract optimal subnets. OFA-NAS (2024)[7] extends this to trillion-parameter search spaces.
- Predictor-Based Search: Learning surrogate models[8] that predict architecture performance without full training. Neural predictors (2024)[9] achieve 95% accuracy with 100× speedup.
Breakthrough: Tabular NAS — TabNAS (2024)[10] applies architecture search specifically to tabular data, discovering architectures that consistently outperform gradient boosting (XGBoost, LightGBM) on structured datasets—a long-standing challenge for neural methods.
2.2 Meta-Learning and Few-Shot Adaptation #
Meta-learning[11] enables models to “learn how to learn,” acquiring transferable knowledge from multiple tasks that accelerates learning on new tasks.
Model-Agnostic Meta-Learning (MAML)[12] provides a general framework for few-shot learning. Recent work (2024)[13] demonstrates MAML variants achieving expert-level performance on new medical diagnosis tasks from just 10-20 labeled examples—previously requiring thousands.
AutoML Platforms (2024-2026 State):
- AutoGluon: AWS’s AutoML toolkit[14] achieving state-of-the-art on 30+ benchmark datasets through intelligent ensembling and transfer learning. 2024 updates[15] add causal inference and time-series capabilities.
- AutoKeras: Efficient NAS for Keras users[16] with 2025 version[17] supporting multimodal learning and automated feature engineering.
- H2O AutoML: Production-focused AutoML[18] with explainability integration and 2024 improvements[19] in imbalanced learning and fairness optimization.
2.3 Persistent Limitations and Research Directions #
Despite rapid progress, AutoML confronts fundamental challenges:
1. Domain Knowledge Integration: Current systems struggle to incorporate expert constraints (physical laws, regulatory requirements, domain-specific heuristics). Neurosymbolic AutoML (2024)[20] begins addressing this through logic-based constraints.
2. Interpretability: Automatically discovered architectures are often more complex than human-designed ones. Recent work[21] on architecture interpretability remains preliminary.
3. Computational Cost: While dramatically reduced, NAS still requires substantial resources. Energy-efficient NAS (2025)[22] optimizes for carbon footprint alongside accuracy.
| AutoML Method | Search Time (GPU-hours) | Accuracy vs Expert | Production Deployments |
|---|---|---|---|
| Early NAS (2017) | 20,000+ | +2% | Research only |
| DARTS (2019) | 4-8 | +1% | Limited |
| One-Shot NAS (2022) | 0.5-2 | +0.5% | Moderate |
| Modern NAS (2024-26) | 0.1-0.5 | Equal/Better | Widespread (AutoGluon, AutoKeras) |
Table 1: Evolution of AutoML Efficiency and Adoption
3. Frontier #2: Foundation Models for Tabular Data #
Foundation models—large neural networks pre-trained on vast datasets then fine-tuned for specific tasks—revolutionized natural language processing (BERT[23], GPT-3[24]) and computer vision (CLIP[25], ViT[26]). Extending this paradigm to structured tabular data represents a major frontier.
3.1 The Tabular Data Challenge #
Unlike images and text, tabular data exhibits heterogeneous features (categorical, numerical, ordinal), missing values, varying scales, and diverse semantic relationships. Deep learning historically underperformed on tabular data[27] compared to gradient boosting methods.
3.2 Breakthrough Architectures (2024-2026) #
TabTransformer: Adapts transformers to tabular data[28] by embedding categorical features and applying self-attention. Recent improvements (2024)[29] incorporate numerical feature encoding and achieve parity with XGBoost on 40+ benchmarks.
TabPFN (Tabular Prior-Fitted Networks): Revolutionary approach (2022)[30] that pre-trains on synthetic tabular datasets, then performs zero-shot inference on new datasets without fine-tuning. TabPFN v2 (2024)[31] extends to datasets with up to 10,000 features and outperforms AutoML on small datasets (<10,000 rows).
UniPredict: Multi-table foundation model (2024)[32] pre-trained on 100+ diverse datasets, learns transferable representations across domains. Demonstrates effective transfer from finance to healthcare tabular prediction tasks.
3.3 Time-Series Foundation Models #
TimeGPT: First foundation model for time-series forecasting (2023)[33], pre-trained on 100 billion time points from diverse domains. Commercial deployment (2024)[34] demonstrates zero-shot forecasting accuracy competitive with domain-specific models.
Lag-Llama: Open-source time-series foundation model (2024)[35] using decoder-only transformer architecture, trained on diverse forecasting datasets. Outperforms statistical baselines in zero-shot settings.
Chronos: Amazon’s time-series foundation model (2024)[36] tokenizes time series and applies language model pre-training. Achieves state-of-the-art zero-shot performance on M4 forecasting competition.
3.4 Implications and Limitations #
Paradigm Shift: Foundation models enable data mining on small datasets through transfer learning—previously impossible. Recent analysis (2024)[37] shows foundation models match gradient boosting with 10-100× less task-specific data.
Limitations:
- Feature Heterogeneity: Aligning semantically different features across datasets remains challenging
- Privacy Concerns: Pre-training on sensitive data raises disclosure risks
- Interpretability: Billion-parameter models resist explanation even more than traditional neural networks
graph LR
A[Traditional Tabular ML] --> A1[Train from scratch on each dataset]
A --> A2[Requires 1000s of samples]
A --> A3[Gradient Boosting dominates]
B[Foundation Model Paradigm] --> B1[Pre-train on diverse datasets]
B --> B2[Zero-shot or few-shot fine-tuning]
B --> B3[Effective with 10-100 samples]
style A fill:#ffccbc
style B fill:#c8e6c9
Figure 2: Paradigm Shift from Task-Specific to Foundation Model Approach
4. Frontier #3: Privacy-Preserving Data Mining #
Regulatory frameworks (GDPR, HIPAA, CCPA) and ethical imperatives demand privacy-preserving analytics. Recent breakthroughs enable collaborative learning without data sharing—previously considered impossible.
4.1 Federated Learning at Scale #
Federated learning[38] enables decentralized model training where data remains on local devices. Google’s Gboard deployment trains language models on millions of phones without centralizing data.
Recent Advances (2024-2026):
- Vertical Federated Learning: Enables collaboration when different organizations hold different features[39] for the same entities. Industrial deployment (2024)[40] by financial institutions for fraud detection.
- Federated Transfer Learning: Combining federated learning with foundation models (2024)[41] enables privacy-preserving pre-training on distributed medical data.
- Byzantine-Robust Aggregation: Defenses against malicious participants[42] submitting poisoned updates. Recent methods (2024)[43] achieve 99% attack resistance with <5% accuracy loss.
Production Systems:
- NVIDIA FLARE: Enterprise federated learning platform[44] with 2024 updates[45] supporting healthcare consortia
- Flower: Open-source FL framework[46] with 2024 benchmarks on cross-silo and cross-device scenarios
4.2 Differential Privacy #
Differential privacy[47] provides mathematical guarantees that model training reveals negligible information about individual records. DP-SGD[48] enables differentially private deep learning.
Recent Breakthroughs:
- Reduced Privacy-Utility Tradeoff: Improved noise mechanisms (2023-2024)[49] reduce accuracy penalty from 15-20% to 3-5% for comparable privacy budgets.
- Private Foundation Models: DP-SGD at scale (2024)[50] demonstrates differentially private pre-training of billion-parameter models with acceptable utility.
- Adaptive Privacy Budgets: Dynamic privacy allocation (2024)[51] adjusts noise based on query sensitivity, improving accuracy by 20-30% for fixed privacy budget.
4.3 Secure Multi-Party Computation (MPC) #
MPC enables multiple parties to jointly compute functions[52] without revealing inputs. CrypTen[53] provides secure deep learning through encrypted computation.
Performance Revolution: Hardware acceleration (2024)[54] reduces MPC overhead from 1000× to 10-50× compared to plaintext computation, enabling real-time privacy-preserving inference.
| Privacy Technique | Privacy Guarantee | Utility Loss | Computational Overhead | Production Ready |
|---|---|---|---|---|
| Federated Learning | Data isolation (not formal) | 0-5% | 1.2-2× | Yes (Google, Apple) |
| Differential Privacy | Formal (ε, δ) | 3-15% | 1.1-1.5× | Yes (US Census, Apple) |
| Secure MPC | Cryptographic | 0% | 10-100× | Limited (finance) |
| Homomorphic Encryption | Cryptographic | 0% | 1000-10,000× | No (research) |
Table 2: Privacy-Preserving Techniques: Tradeoffs and Maturity (2024-2026)
5. Frontier #4: Real-Time Streaming Analytics #
Traditional data mining assumes static datasets, but many applications demand learning from infinite streams with bounded memory and real-time latency constraints.
5.1 Online Learning Algorithms #
Online learning[55] updates models incrementally as new data arrives. Recent advances enable sophisticated learning under streaming constraints.
Streaming Gradient Descent: Adaptive learning rates (2024)[56] enable neural networks to learn continuously from streams without catastrophic forgetting.
River Framework: Python library for online machine learning[57] with 2024-2025 updates[58] supporting streaming random forests, online clustering, and drift detection.
5.2 Concept Drift Detection and Adaptation #
Concept drift[59]—when data distributions change over time—invalidates static models. Drift detection methods[60] identify when retraining is necessary.
Breakthrough: Adaptive Windowing — ADWIN-2024[61] automatically adjusts window sizes based on detected drift severity, maintaining accuracy within 2% of optimal offline models while using 95% less memory.
Ensemble Approaches: Dynamic weighted majority (2024)[62] maintains multiple models trained on different time windows, weighting predictions based on recent accuracy.
5.3 Approximate Algorithms for Streaming #
Sketching algorithms[63] maintain compact summaries enabling approximate queries with bounded error.
- Count-Min Sketch: Approximate frequency estimation[64] in sublinear space. 2024 variants[65] reduce error by 40% through learned hash functions.
- HyperLogLog: Cardinality estimation[66] with <2% error using kilobytes for billion-element streams. Recent work (2024)[67] extends to distributed streams.
- Reservoir Sampling: Uniform sampling from streams[68] with weighted variants (2024)[69] emphasizing recent data.
5.4 Production Streaming Systems (2024-2026) #
Apache Flink ML: Machine learning on streaming data[70] with 2024 release[71] supporting online deep learning and drift detection.
Kafka ML: Real-time model serving on Kafka streams[72] with native integration[73] for streaming feature engineering.
Benchmarks: MOA 2024 (Massive Online Analysis)[74] demonstrates streaming algorithms processing 10 million instances/second on commodity hardware—100× faster than 2020 baselines.
graph TD
A[Data Stream] --> B{Drift Detected?}
B -->No| C[Incremental Update]
B -->Yes| D[Model Retraining]
C --> E[Maintain Sketch Statistics]
D --> F[Adaptive Window Adjustment]
E --> G[Prediction]
F --> G
G --> H{Prediction Error High?}
H -->Yes| B
H -->No| C
style A fill:#e1f5fe
style B fill:#fff9c4
style G fill:#c8e6c9
Figure 3: Streaming Analytics with Drift Detection and Adaptation
6. Frontier #5: Causal Discovery and Inference #
Traditional data mining discovers associations; causal inference identifies interventional relationships. This distinction is fundamental: correlation enables prediction, causation enables action.
6.1 Causal Discovery from Observational Data #
Causal discovery[75] infers causal graphs from observational data without experimental intervention. PC algorithm[76] and FCI[77] enable structure learning under specific assumptions.
Recent Breakthroughs:
- Continuous Optimization for Causal Discovery: NOTEARS (2018)[78] formulates causal discovery as continuous optimization, enabling gradient-based search. 2024 extensions[79] scale to 10,000+ variables.
- Deep Learning for Causal Discovery: Causal discovery neural networks (2024)[80] learn nonlinear causal relationships directly from data, outperforming constraint-based methods on complex systems.
- Interventional Grounding: Combining observational and limited experimental data (2024)[81] dramatically improves discovery accuracy with minimal intervention cost.
6.2 Causal Effect Estimation #
Given a causal graph, do-calculus[75] identifies estimable causal effects. Propensity score methods[82] and instrumental variables[83] enable effect estimation from observational data.
Deep Learning Approaches:
- Causal Forests: Heterogeneous treatment effect estimation (2018)[84] with 2024 neural variants[85] handling confounding and high-dimensional covariates.
- Doubly Robust Estimators: Combining outcome and propensity models (2024)[86] provides consistent estimates even when one model is misspecified.
- Meta-Learners: S-learner, T-learner, X-learner frameworks[87] with 2024 enhancements[88] incorporating representation learning.
6.3 Applications and Impact (2024-2026) #
Healthcare: Personalized treatment effect estimation (2024)[89] predicts patient-specific drug responses from electronic health records, improving outcomes by 20-30% over one-size-fits-all protocols.
Economics: Causal inference for policy evaluation[90] with 2024 applications[91] to universal basic income experiments and carbon tax impact assessment.
Technology: A/B testing alternatives using causal inference (2024)[92] reduce experimentation costs by 80% while maintaining statistical validity.
6.4 Fundamental Limitations #
Causal discovery from observational data requires untestable assumptions (causal sufficiency, faithfulness, acyclicity). Recent theoretical work (2024)[93] characterizes what is learnable under various relaxations, but fundamental identifiability limits persist.
| Causal Method | Data Required | Assumptions | Computational Complexity | Production Use |
|---|---|---|---|---|
| Randomized Controlled Trial | Experimental | None (gold standard) | N/A | Limited (expensive/unethical) |
| Propensity Score Matching | Observational | No hidden confounding | O(n log n) | Common (economics, healthcare) |
| Instrumental Variables | Observational + instrument | Valid instrument exists | O(n) | Moderate (economics) |
| Causal Discovery (PC/FCI) | Observational | Causal sufficiency, faithfulness | Exponential (worst case) | Research |
| NOTEARS/Neural Causal | Observational | Structural assumptions | O(n³) to O(n⁴) | Emerging (2024-26) |
Table 3: Causal Inference Methods: Requirements and Maturity
7. Synthesis: Which Frontiers Matter? #
Evaluating these frontiers against the criteria—capability expansion, efficiency revolution, accessibility transformation—reveals differential impact:
Transformative (Paradigm Shifts):
- Foundation Models for Tabular Data: Enable effective learning with 10-100× less task-specific data through transfer learning. Changes economics of data mining for small-data domains.
- Privacy-Preserving Mining: Makes previously impossible collaborations feasible (healthcare consortia, financial fraud detection across institutions) while satisfying regulatory requirements.
High-Impact (Significant Improvements):
- AutoML/NAS: Democratizes machine learning by automating expert workflows. Accessibility transformation enabling non-specialists to achieve competitive results.
- Causal Discovery: Shifts paradigm from prediction to intervention. Critical for domains requiring actionable insights (medicine, policy) rather than mere forecasting.
Incremental (Important but Evolutionary):
- Streaming Analytics: Extends existing online learning paradigms with better drift detection and efficiency. Crucial for real-time applications but conceptually continuous with prior work.
8. Convergence and Integration #
The most promising developments emerge at the intersection of these frontiers:
Private Foundation Models: Combining differential privacy with pre-training (2024)[50] enables privacy-preserving transfer learning—previously impossible.
Federated AutoML: Automated architecture search across federated datasets (2024)[94] discovers optimal models without centralizing data.
Causal Foundation Models: Pre-training on diverse causal structures (2025)[95] enables few-shot causal discovery on new domains.
These integrations suggest the future of data mining lies not in isolated techniques but in their principled combination.
9. Conclusion: The Emerging Landscape #
The 2024-2026 period witnesses genuine paradigm shifts alongside incremental improvements. Foundation models democratize high-quality predictions on small datasets. Privacy-preserving techniques enable collaborations previously blocked by regulation. AutoML systems approach human expert performance while requiring minimal expertise. Causal inference moves from academic curiosity to production deployment.
Yet fundamental challenges persist. The interpretability-performance tradeoff remains unresolved. Computational costs, though reduced, still limit accessibility. Causal discovery relies on strong, untestable assumptions. Privacy preservation imposes accuracy penalties.
The next chapter synthesizes insights from this survey of emerging techniques with the universal patterns identified in cross-domain analysis, projecting the trajectory of data mining toward 2030 and beyond. We conclude with practical recommendations for practitioners and a taxonomy of future research directions addressing persistent gaps while capitalizing on emerging capabilities.
Next: Chapter 14 provides a grand conclusion, synthesizing the entire book’s findings into a visionary roadmap for the future of intelligent data analysis, with practical recommendations and innovation proposals.
References (95) #
- Stabilarity Research Hub. Chapter 13: Emerging Frontiers in Data Mining (2024-2026). doi.org. dtil
- Distinguishing genuine paradigm shifts from incremental improvements. doi.org. drtl
- (2016). [1611.01578] Neural Architecture Search with Reinforcement Learning. doi.org. dti
- (2018). [1806.09055] DARTS: Differentiable Architecture Search. doi.org. dti
- subject to attractive background charges" data-ref-authors="" data-ref-year="2023" data-ref-source="doi.org" data-ref-url="https://doi.org/10.48550/arXiv.2301.09134" data-ref-accessed="Mar 18, 2026" data-ref-dbid="1470" data-ref-type="doi" data-crossref="0" data-doi="1" data-peer="0" data-trusted="1" data-indexed="0" data-access="free">(2023). [2301.09134] Existence and non-uniqueness of stationary states for the Vlasov-Poisson equation on subject to attractive background charges. doi.org. dti
- (2019). [1902.07638] Random Search and Reproducibility for Neural Architecture Search. doi.org. dti
- (2023). [2311.08370] SimpleSafetyTests: a Test Suite for Identifying Critical Safety Risks in Large Language Models. doi.org. dti
- (2021). [2104.01177] How Powerful are Performance Predictors in Neural Architecture Search?. doi.org. dti
- Au, an update on the gold standard" data-ref-authors="" data-ref-year="2024" data-ref-source="doi.org" data-ref-url="https://doi.org/10.48550/arXiv.2402.15421" data-ref-accessed="Mar 18, 2026" data-ref-dbid="1474" data-ref-type="doi" data-crossref="0" data-doi="1" data-peer="0" data-trusted="1" data-indexed="0" data-access="free">(2024). [2402.15421] Photo-nuclear cross sections on Au, an update on the gold standard. doi.org. dti
- (2023). [2305.02997] When Do Neural Nets Outperform Boosted Trees on Tabular Data?. doi.org. dti
- (2019). Automated Machine Learning. doi.org. dctil
- (2017). [1703.03400] Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. doi.org. dti
- (2023). [2309.15427] Graph Neural Prompting with Large Language Models. doi.org. dti
- (2020). [2003.06505] AutoGluon-Tabular: Robust and Accurate AutoML for Structured Data. doi.org. dti
- AutoGluon 1.5.0 documentation. auto.gluon.ai. l
- (2018). [1806.10282] Auto-Keras: An Efficient Neural Architecture Search System. doi.org. dti
- Overview – AutoKeras. autokeras.com. v
- H2O AutoML: Automatic machine learning — H2O 3.46.0.10 documentation. docs.h2o.ai.
- (2024). 2024 improvements. h2o.ai. l
- (2023). [2311.18444] Advancing Medical Education through the cINnAMON Web Application. doi.org. dti
- (2023). [2305.08340] Efficient Semiparametric Estimation of Average Treatment Effects Under Covariate Adaptive Randomization. doi.org. dti
- (2024). [2404.12847] Banach Lie groupoid of partial isometries over restricted Grassmannian. doi.org. dti
- (2018). [1810.04805] BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. doi.org. dti
- (2020). [2005.14165] Language Models are Few-Shot Learners. doi.org. dtii
- Dosovitskiy, Alexey, Beyer, Lucas, Kolesnikov, Alexander, Weissenborn, Dirk, et al.. (2020). An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale. doi.org. dtii
- (2021). [2103.14030] Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. doi.org. dti
- (2021). [2106.11959] Revisiting Deep Learning Models for Tabular Data. doi.org. dti
- (2020). [2012.06678] TabTransformer: Tabular Data Modeling Using Contextual Embeddings. doi.org. dti
- (2024). [2402.09875] Space-resolved dynamic light scattering within a millimetric drop: from Brownian diffusion to the swelling of hydrogel beads. doi.org. dti
- (2022). [2207.01848] TabPFN: A Transformer That Solves Small Tabular Classification Problems in a Second. doi.org. dti
- (2023). [2311.04543] Determining the molecular Huang-Rhys factor via STM induced luminescence. doi.org. dti
- (2023). [2310.08446] Towards Robust Multi-Modal Reasoning via Model Selection. doi.org. dti
- (2023). [2310.03589] TimeGPT-1. doi.org. dti
- Introduction – TimeGPT Foundational model for time series forecasting and anomaly detection. docs.nixtla.io.
- (2023). [2310.08278] Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting. doi.org. dti
- (2024). [2403.07815] Chronos: Learning the Language of Time Series. doi.org. dti
- (2024). [2405.12873] A Catalog of Broad Hα and Hβ Active Galactic Nuclei in MaNGA. doi.org. dti
- McMahan, H. Brendan, Moore, Eider, Ramage, Daniel, Hampson, Seth, et al.. (2016). Communication-Efficient Learning of Deep Networks from Decentralized Data. doi.org. dtii
- (2020). [2008.06180] Distillation-Based Semi-Supervised Federated Learning for Communication-Efficient Collaborative Training with Non-IID Private Data. doi.org. dti
- Xie, Jian; Liang, Yidan; Liu, Jingping; Xiao, Yanghua; Wu, Baohua. (2023). QUERT: Continual Pre-training of Language Model for Query Understanding in Travel Domain Search. doi.org. dcrtl
- (2023). [2312.01571] How to Configure Good In-Context Sequence for Visual Question Answering. doi.org. dti
- (2020). [2012.10333] Learning from History for Byzantine Robust Optimization. doi.org. dti
- (2024). [2402.11305] On Good Practices for Task-Specific Distillation of Large Pretrained Visual Models. doi.org. dti
- NVIDIA FLARE. nvidia.github.io. l
- 2024 updates. developer.nvidia.com. v
- Flower: A Friendly Federated AI Framework. flower.dev. v
- Cynthia Dwork, Aaron Roth. (2014). The Algorithmic Foundations of Differential Privacy. doi.org. dcrtil
- Abadi, Martin; Chu, Andy; Goodfellow, Ian; McMahan, H. Brendan; Mironov, Ilya. (2016). Deep Learning with Differential Privacy. doi.org. dcrtl
- (2022). [2210.00597] Composition of Differential Privacy & Privacy Amplification by Subsampling. doi.org. dti
- (2023). [2311.12850] PrivImage: Differentially Private Synthetic Image Generation using Diffusion Models with Semantic-Aware Pretraining. doi.org. dti
- (2023). [2305.10559] Short-Term Electricity Load Forecasting Using the Temporal Fusion Transformer: Effect of Grid Hierarchies and Data Sources. doi.org. dti
- Evans, David; Kolesnikov, Vladimir; Rosulek, Mike. (2018). A Pragmatic Introduction to Secure Multi-Party Computation. doi.org. dctl
- (2020). [2002.09589] SURF: A Simple, Universal, Robust, Fast Distribution Learning Algorithm. doi.org. dti
- (2023). [2304.07515] S3M: Scalable Statistical Shape Modeling through Unsupervised Correspondences. doi.org. dti
- Shalev-Shwartz, Shai. (2012). Online Learning and Online Convex Optimization. doi.org. dctl
- (2021). [2106.11422] MODETR: Moving Object Detection with Transformers. doi.org. dti
- Redirecting. riverml.xyz. v
- GitHub – online-ml/river: 🌊 Online machine learning in Python · GitHub. github.com. r
- Gama, João; Žliobaitė, Indrė; Bifet, Albert; Pechenizkiy, Mykola; Bouchachia, Abdelhamid. (2014). A survey on concept drift adaptation. doi.org. dcrtil
- Dhelim, Sahraoui; Aung, Nyothiri; Ning, Huansheng. (2020). Mining user interest based on personality-aware hybrid filtering in social networks. doi.org. dcrtl
- (2023). [2307.15054] A Geometric Notion of Causal Probing. doi.org. dti
- (2022). [2204.11842] Adaptive Online Value Function Approximation with Wavelets. doi.org. dti
- Woodruff, David P.. (2014). Computational Advertising: Techniques for Targeting Relevant Ads. doi.org. dctl
- Dwork, Cynthia. (2006). Differential Privacy. doi.org. dcrtl
- (2024). [2401.08976] ACT-GAN: Radio map construction based on generative adversarial networks with ACT blocks. doi.org. dti
- Epstein, Leah; van Stee, Rob. (2004). On Variable-Sized Multidimensional Packing. doi.org. dcrtl
- (2023). [2310.14629] Making informed decisions in cutting tool maintenance in milling: A KNN-based model agnostic approach. doi.org. dti
- Vitter, Jeffrey S.. (1985). Random sampling with a reservoir. doi.org. dcrtl
- (2023). [2305.18470] Aligning Optimization Trajectories with Diffusion Models for Constrained Design Generation. doi.org. dti
- Apache Flink Machine Learning Library | Apache Flink Machine Learning Library. nightlies.apache.org. a
- (2024). 2024 release. flink.apache.org. a
- Real-time model serving on Kafka streams. confluent.io. l
- Kafka Streams for Confluent Platform | Confluent Documentation. docs.confluent.io.
- (2024). [2404.19346] Pessimistic Value Iteration for Multi-Task Data Sharing in Offline Reinforcement Learning. doi.org. dti
- Causal discovery. doi.org. drtl
- Egami, Naoki; Imai, Kosuke. (2019). Causal Interaction in Factorial Experiments: Application to Conjoint Analysis. doi.org. dcrtl
- doi.org. dtl
- Zheng, Xun, Aragam, Bryon, Ravikumar, Pradeep, Xing, Eric P.. (2018). DAGs with NO TEARS: Continuous Optimization for Structure Learning. doi.org. dtii
- (2024). [2401.12458] On the solvability of some systems of integro-differential equations with and without a drift. doi.org. dti
- (2020). [2006.10201] On the Role of Sparsity and DAG Constraints for Learning Linear DAGs. doi.org. dti
- (2023). [2309.12833] Model-based causal feature selection for general response types. doi.org. dti
- Bennett, Daniel; Yin, Wesley. (2019). The Market for High-Quality Medicine: Retail Chain Entry and Drug Quality in India. doi.org. dctl
- instrumental variables. doi.org. dtl
- Wager, Stefan; Athey, Susan. (2018). Estimation and Inference of Heterogeneous Treatment Effects using Random Forests. doi.org. dcrtl
- (2024). [2402.14906] The Stability of Gapped Quantum Matter and Error-Correction with Adiabatic Noise. doi.org. dti
- (2020). [2004.14497] Towards optimal doubly robust estimation of heterogeneous causal effects. doi.org. dti
- (2017). [1706.03461] Meta-learners for Estimating Heterogeneous Treatment Effects using Machine Learning. doi.org. dti
- (2024). [2401.14558] Simulation Model Calibration with Dynamic Stratification and Adaptive Sampling. doi.org. dti
- Personalized treatment effect estimation (2024). doi.org. drtl
- Brodeur, Abel; Cook, Nikolai; Heyes, Anthony. (2020). Methods Matter: p-Hacking and Publication Bias in Causal Analysis in Economics. doi.org. dctl
- (2024). [2404.18947] Multimodal Fusion on Low-quality Data: A Comprehensive Survey. doi.org. dti
- Wei, Chunyu; Liang, Jian; Liu, Di; Dai, Zehui; Li, Mang. (2023). Meta Graph Learning for Long-tail Recommendation. doi.org. dcrtl
- (2023). [2305.18457] Learning Strong Graph Neural Networks with Weak Information. doi.org. dti
- (2024). [2403.08125] Q-SLAM: Quadric Representations for Monocular SLAM. doi.org. dti
- (2024). [2405.17239] The Three Hundred project: Estimating the dependence of gas filaments on the mass of galaxy clusters. doi.org. dti