Difference between revisions of "Functional Data Analysis"
| (4 intermediate revisions by the same user not shown) | |||
| Line 1: | Line 1: | ||
| − | + | * Channel: [https://t.me/+XyXmEXRlrXB9dZKD The chat-link FDA group] | |
| − | * | ||
| − | |||
| − | |||
| − | |||
| − | |||
| − | == | + | ==Introduction== |
| − | |||
| − | |||
| − | |||
The statistical analysis of spatial time series requires additional methods of data analysis. First, we suppose time is continuous, put the state space changes <math>\frac{d\mathbf{x}}{dt}</math>, and use neural ordinary and stochastic differential equations. Second, we analyze a multivariate and multidimensional time series and use the tensor representation and tensor analysis. Third, since the time series have significant cross-correlation, we model them in the Riemannian space. Fourth, medical time series are periodic; the base model is the pendulum model, <math>\frac{d^2x}{dt^2}=-c\sin{x}</math>. We use physics-informed neural networks to approximate data. Fifth, the practical experiments involve multiple data sources. We use canonical correlation analysis with a latent state space. This space aligns the source and target spaces and generates data in the source and target manifolds. | The statistical analysis of spatial time series requires additional methods of data analysis. First, we suppose time is continuous, put the state space changes <math>\frac{d\mathbf{x}}{dt}</math>, and use neural ordinary and stochastic differential equations. Second, we analyze a multivariate and multidimensional time series and use the tensor representation and tensor analysis. Third, since the time series have significant cross-correlation, we model them in the Riemannian space. Fourth, medical time series are periodic; the base model is the pendulum model, <math>\frac{d^2x}{dt^2}=-c\sin{x}</math>. We use physics-informed neural networks to approximate data. Fifth, the practical experiments involve multiple data sources. We use canonical correlation analysis with a latent state space. This space aligns the source and target spaces and generates data in the source and target manifolds. | ||
| Line 15: | Line 7: | ||
This field of Machine Learning applies to any field where the measurements have continuous time and space data acquired from multimodal sources: climate modeling, neural interfaces, solid-state physics, electronics, fluid dynamics, and many more. We will carefully collect both the theory and its practice. | This field of Machine Learning applies to any field where the measurements have continuous time and space data acquired from multimodal sources: climate modeling, neural interfaces, solid-state physics, electronics, fluid dynamics, and many more. We will carefully collect both the theory and its practice. | ||
| − | == | + | ===Foundation models for scientific research=== |
| − | + | Foundation AI models are universal models to solve a wide set of problems. This project proposes to investigate the theoretical properties of foundation models. The domain to model is a spatial-time series. These data are used in various scientific disciplines and serve to generalise scientific knowledge and make forecasts. The essential problems, formulated as user requests that solve a foundation model, are <i>forecasting</i> and <i>generation</i> of time series; <i>analysis</i> and <i>classification</i> of time series; <i>detection of change point</i>, and <i>causal inference</i>. To solve these problems, the foundation AI models are trained on massive datasets. The main goal of this project is to compare various architectures of foundation models to find an optimal architecture that solves the listed problems for a wide range of spatial time series. | |
| + | |||
| + | == Topics to discuss== | ||
# State Space Models, Convolution, SSA, SSM (Spectral Submanifolds) | # State Space Models, Convolution, SSA, SSM (Spectral Submanifolds) | ||
# Neural and Controlled ODE, Neural PDE, Geometric Learning | # Neural and Controlled ODE, Neural PDE, Geometric Learning | ||
| Line 23: | Line 17: | ||
# Riemmannian models; time series generation | # Riemmannian models; time series generation | ||
# AI for science: mathematical modelling principles | # AI for science: mathematical modelling principles | ||
| − | + | # <it>Left behind:<\it> data-driven tensor analysis, differential forms, and spinors | |
| − | |||
=== State of the Art in 2025=== | === State of the Art in 2025=== | ||
| − | + | The NeurIPS workshop "Foundational models for science" reflected this theme in 2024: | |
# Foundation Models for Science: Progress, Opportunities, and Challenges [https://neurips.cc/virtual/2024/workshop/84714 URL] | # Foundation Models for Science: Progress, Opportunities, and Challenges [https://neurips.cc/virtual/2024/workshop/84714 URL] | ||
# Foundation Models for the Earth system [https://neurips.cc/virtual/2024/107817 UPL, no paper] | # Foundation Models for the Earth system [https://neurips.cc/virtual/2024/107817 UPL, no paper] | ||
| Line 35: | Line 28: | ||
# VSMNO: Solving PDE by Utilizing Spectral Patterns of Different Neural Operators [https://openreview.net/pdf?id=oCT8pYix5e NIPS] | # VSMNO: Solving PDE by Utilizing Spectral Patterns of Different Neural Operators [https://openreview.net/pdf?id=oCT8pYix5e NIPS] | ||
| − | === | + | === Physics problem Simulations === |
# The Well: a Large-Scale Collection of Diverse Physics Simulations for Machine Learning [https://arxiv.org/pdf/2412.00568 ArXiv], [https://polymathic-ai.org/the_well/data_format/ Code] | # The Well: a Large-Scale Collection of Diverse Physics Simulations for Machine Learning [https://arxiv.org/pdf/2412.00568 ArXiv], [https://polymathic-ai.org/the_well/data_format/ Code] | ||
# Polymatic Advancing Science through Multi‑Disciplinary AI [https://polymathic-ai.org/ blog] | # Polymatic Advancing Science through Multi‑Disciplinary AI [https://polymathic-ai.org/ blog] | ||
| Line 53: | Line 46: | ||
# Longitudinal predictive modeling of tau progression along the structural connectome by Joyita Dutta [https://www.sciencedirect.com/science/article/pii/S1053811921004031?via%3Dihub 2021] | # Longitudinal predictive modeling of tau progression along the structural connectome by Joyita Dutta [https://www.sciencedirect.com/science/article/pii/S1053811921004031?via%3Dihub 2021] | ||
| − | == Key reviews == | + | === Key reviews on AI for Science === |
# 2018. Diffusion Convolutional Recurrent Neural Network [https://arxiv.org/pdf/1707.01926 ICLR] | # 2018. Diffusion Convolutional Recurrent Neural Network [https://arxiv.org/pdf/1707.01926 ICLR] | ||
# 2021. Neural Partial Differential Equations with Functional Convolution [https://openreview.net/pdf?id=D4A-v0kltaX ICLR] | # 2021. Neural Partial Differential Equations with Functional Convolution [https://openreview.net/pdf?id=D4A-v0kltaX ICLR] | ||
| Line 60: | Line 53: | ||
# 2022. Time Series Forecasting Using Manifold Learning, Radial Basis Function Interpolation and Geometric Harmonics [https://doi.org/10.1063/5.0094887 doi] (all basic models + superpositions review) | # 2022. Time Series Forecasting Using Manifold Learning, Radial Basis Function Interpolation and Geometric Harmonics [https://doi.org/10.1063/5.0094887 doi] (all basic models + superpositions review) | ||
| − | === Catch-up === | + | === Catch-up on LLM === |
If you are not familiar with the LLM and GPT: | If you are not familiar with the LLM and GPT: | ||
# Build an LLM from scratch by Sebastian Raschka, 2025 [https://github.com/rasbt/LLMs-from-scratch/ github] | # Build an LLM from scratch by Sebastian Raschka, 2025 [https://github.com/rasbt/LLMs-from-scratch/ github] | ||
| Line 66: | Line 59: | ||
For fun, the vibe coding: [https://cursor.com/home 1], [https://windsurf.com/ 2], [https://www.augmentcode.com/ 3], [https://www.augmentcode.com/install 4], also [https://arxiv.org/abs/2501.09223 Foundations of LLM] and [https://www.youtube.com/@AndrejKarpathy Karpathy's] project | For fun, the vibe coding: [https://cursor.com/home 1], [https://windsurf.com/ 2], [https://www.augmentcode.com/ 3], [https://www.augmentcode.com/install 4], also [https://arxiv.org/abs/2501.09223 Foundations of LLM] and [https://www.youtube.com/@AndrejKarpathy Karpathy's] project | ||
[https://github.com/karpathy/nanoGPT nanoGPT] | [https://github.com/karpathy/nanoGPT nanoGPT] | ||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
<!---Your goal is to enhance your abilities to''' convey messages''' to the reader in the ''' language of applied mathematics'''. The main part of your MS thesis work is the theoretical foundations of Machine Learning, where you present your personal results supported by the necessary theory. ---> | <!---Your goal is to enhance your abilities to''' convey messages''' to the reader in the ''' language of applied mathematics'''. The main part of your MS thesis work is the theoretical foundations of Machine Learning, where you present your personal results supported by the necessary theory. ---> | ||
| − | + | <!---Structure of seminars | |
| − | |||
| − | |||
The semester lasts 12 weeks, and six couple of weeks are for homework. | The semester lasts 12 weeks, and six couple of weeks are for homework. | ||
* Odd week: introduction to the topic and a handout of a theme for the homework. | * Odd week: introduction to the topic and a handout of a theme for the homework. | ||
* Every week: a discussion of the essay, collecting the list of improvements to each essay. | * Every week: a discussion of the essay, collecting the list of improvements to each essay. | ||
| − | * Odd week: a discussion of the improved essay, putting the essays into a joint structure. | + | * Odd week: a discussion of the improved essay, putting the essays into a joint structure.---> |
| − | + | <!---Scoring | |
| − | + | The group activity is evaluated by cross-ranking with the Kemeni median score. The personal talks give a score.---> | |
| − | The group activity is evaluated by cross-ranking with the Kemeni median score. The personal talks give a score. | ||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
<!-- | <!-- | ||
| − | + | <!--The homework | |
The course gives two credits, so it requires time. The result is a two-page essay. It delivers an introduction to the designated topic. It could be automatically generated or collected from Wikipedia. The main requirement is that you be responsible for each statement in your essay. Each formula is yours. | The course gives two credits, so it requires time. The result is a two-page essay. It delivers an introduction to the designated topic. It could be automatically generated or collected from Wikipedia. The main requirement is that you be responsible for each statement in your essay. Each formula is yours. | ||
| − | |||
The essay carries a comprehensive and strict answer to the topic question; illustrative plots are welcome. The result is ready to compile in a joint manuscript after the Even week. So please use the LaTeX template. | The essay carries a comprehensive and strict answer to the topic question; illustrative plots are welcome. The result is ready to compile in a joint manuscript after the Even week. So please use the LaTeX template. | ||
| − | |||
The style is set theory, algebra, analysis, and Bayesian statistics. Category theory and homotopy theory are welcome. | The style is set theory, algebra, analysis, and Bayesian statistics. Category theory and homotopy theory are welcome. | ||
| − | |||
This course gives you two credits, so it is 76/10 = ''' 5 hours of weekly ''' homework. | This course gives you two credits, so it is 76/10 = ''' 5 hours of weekly ''' homework. | ||
--> | --> | ||
| − | <!-- | + | <!--Templated and links |
| − | |||
* The Git Hub to download the essays | * The Git Hub to download the essays | ||
* The overleaf to compile the joint manuscript | * The overleaf to compile the joint manuscript | ||
* The LaTeX template for an essay --> | * The LaTeX template for an essay --> | ||
| − | + | <!--Requirements for the text and the discussion | |
| − | |||
# Comprehensive explanation of the method or the question we discuss | # Comprehensive explanation of the method or the question we discuss | ||
# Only the principle, no experiments | # Only the principle, no experiments | ||
| Line 200: | Line 86: | ||
# The picture is obligatory | # The picture is obligatory | ||
# However, a brief reference to some deep learning structure is welcome | # However, a brief reference to some deep learning structure is welcome | ||
| − | # Talk could be a slide or | + | # Talk could be a slide or the text itself |
# The list of references with doi | # The list of references with doi | ||
# Tell how it was generated | # Tell how it was generated | ||
| − | # Observing a gap, put a note about it (to question later) | + | # Observing a gap, put a note about it (to question later)--> |
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
==Table of topics for seminars== | ==Table of topics for seminars== | ||
| Line 221: | Line 99: | ||
# Riemannian spaces | # Riemannian spaces | ||
| − | + | These items comprise the stochastic-deterministic decomposition. So the questions include three parts: | |
# deterministic model, | # deterministic model, | ||
# generative model, | # generative model, | ||
# stochastic-deterministic decomposition method. | # stochastic-deterministic decomposition method. | ||
| − | |||
=== Multimodal data === | === Multimodal data === | ||
| Line 236: | Line 113: | ||
# Comparative analysis of variants of CCA, like PLS and others | # Comparative analysis of variants of CCA, like PLS and others | ||
# Functional PCA | # Functional PCA | ||
| − | + | # Canonical Correlation Analysis: forecasting model and loss function with variants- | |
| − | + | # CCA parameter estimation algorithm | |
Talks | Talks | ||
| Line 247: | Line 124: | ||
=== Continous models === | === Continous models === | ||
| − | |||
# Neural ODE | # Neural ODE | ||
# Continous state space models | # Continous state space models | ||
# Continous normalizing flows | # Continous normalizing flows | ||
# Adjoint method and continuous backpropagation | # Adjoint method and continuous backpropagation | ||
| − | # Neural Delayed Differential Equations | + | # Neural Delayed Differential Equations |
| + | # Neural CDE (PID control is welcome) | ||
# Neural PDE | # Neural PDE | ||
# S4 and Hippo models [https://doi.org/10.48550/arXiv.2206.12037], [https://github.com/HazyResearch/state-spaces] (LSSL, SaShiMi, DSS, HTTYH, S4D, and S4ND) | # S4 and Hippo models [https://doi.org/10.48550/arXiv.2206.12037], [https://github.com/HazyResearch/state-spaces] (LSSL, SaShiMi, DSS, HTTYH, S4D, and S4ND) | ||
| Line 262: | Line 139: | ||
# Adjoint method and continuous backpropagation [https://github.com/intsystems/IDA/blob/main-2024/essay-2-cont/boeva/essay2_final.pdf Galina] | # Adjoint method and continuous backpropagation [https://github.com/intsystems/IDA/blob/main-2024/essay-2-cont/boeva/essay2_final.pdf Galina] | ||
# Riemannian continuous models [https://github.com/intsystems/IDA/blob/main-2024/essay-2-cont/vladimirov/main-final.pdf Eduard] | # Riemannian continuous models [https://github.com/intsystems/IDA/blob/main-2024/essay-2-cont/vladimirov/main-final.pdf Eduard] | ||
| + | |||
====SINDy==== | ====SINDy==== | ||
# Learning partial differential equations via data discovery and sparse optimization by Hayden Schaeffer, 2017 [https://doi.org/10.1098/rspa.2016.0446 DOI], [https://robotics.caltech.edu/wiki/images/b/bc/LearningPDEs.pdf PDF] | # Learning partial differential equations via data discovery and sparse optimization by Hayden Schaeffer, 2017 [https://doi.org/10.1098/rspa.2016.0446 DOI], [https://robotics.caltech.edu/wiki/images/b/bc/LearningPDEs.pdf PDF] | ||
| Line 336: | Line 214: | ||
==Practical spatial-time series== | ==Practical spatial-time series== | ||
| + | |||
| + | ===Datasets=== | ||
| + | # ClimateSet, 2023 [https://arxiv.org/pdf/2311.03721 ArXiv] | ||
# A guide to state–space modeling of ecological time series, 2021 [https://doi.org/10.1002/ecm.1470 PDF], (Bayesian Kalman) | # A guide to state–space modeling of ecological time series, 2021 [https://doi.org/10.1002/ecm.1470 PDF], (Bayesian Kalman) | ||
# Kalman Filtering and Smoothing, 2025 [https://arxiv.org/pdf/2405.08971 ArXiv] (Riemannian Kalman) | # Kalman Filtering and Smoothing, 2025 [https://arxiv.org/pdf/2405.08971 ArXiv] (Riemannian Kalman) | ||
| − | |||
| − | |||
| − | |||
| − | |||
===General=== | ===General=== | ||
# Artificial Intelligence for Science in Quantum, Atomistic, and Continuum Systems [https://arxiv.org/abs/2307.08423 arxiv 2023] | # Artificial Intelligence for Science in Quantum, Atomistic, and Continuum Systems [https://arxiv.org/abs/2307.08423 arxiv 2023] | ||
| Line 350: | Line 227: | ||
# Differential Geometry of Curves and Surfaces: Textbook, 2016 by Kristopher Tapp [https://link.springer.com/book/10.1007/978-3-319-39799-3] | # Differential Geometry of Curves and Surfaces: Textbook, 2016 by Kristopher Tapp [https://link.springer.com/book/10.1007/978-3-319-39799-3] | ||
| − | === | + | ===Basic literature=== |
# Understanding Deep Learning ''by Simon J.D. Prince'' [https://udlbook.github.io/udlbook/ mit 2023] | # Understanding Deep Learning ''by Simon J.D. Prince'' [https://udlbook.github.io/udlbook/ mit 2023] | ||
# Deep Learning by ''C.M. and H. Bishops'' [https://www.bishopbook.com/ Springer 2024] (online version) | # Deep Learning by ''C.M. and H. Bishops'' [https://www.bishopbook.com/ Springer 2024] (online version) | ||
| Line 359: | Line 236: | ||
# A Tutorial on Independent Component Analysis [https://arxiv.org/abs/1404.2986 arxiv, 2014] | # A Tutorial on Independent Component Analysis [https://arxiv.org/abs/1404.2986 arxiv, 2014] | ||
# On the Stability of Multilinear Dynamical Systems [https://arxiv.org/abs/2105.01041 arxiv 2022] | # On the Stability of Multilinear Dynamical Systems [https://arxiv.org/abs/2105.01041 arxiv 2022] | ||
| − | # Tensor-based Regression Models and Applications ''by Ming Hou'' Thèse [https://core.ac.uk/download/pdf/442636056.pdf Uni-Laval 2017] | + | # Tensor-based Regression Models and Applications ''by Ming Hou'' Thèse [https://core.ac.uk/download/pdf/442636056.pdf Uni-Laval 2017] |
# Tensor Canonical Correlation Analysis for Multi-view Dimension Reduction [https://arxiv.org/pdf/1502.02330] (Semkin) | # Tensor Canonical Correlation Analysis for Multi-view Dimension Reduction [https://arxiv.org/pdf/1502.02330] (Semkin) | ||
#Tensor Learning in Multi-view Kernel PCA [https://link.springer.com/chapter/10.1007/978-3-030-01421-6_21 arxiv 2018] | #Tensor Learning in Multi-view Kernel PCA [https://link.springer.com/chapter/10.1007/978-3-030-01421-6_21 arxiv 2018] | ||
| Line 386: | Line 263: | ||
# NeuPDE: Neural Network Based Ordinary and Partial Differential Equations for Modeling Time-Dependent Data [https://arxiv.org/abs/1908.03190 arxiv 2019] | # NeuPDE: Neural Network Based Ordinary and Partial Differential Equations for Modeling Time-Dependent Data [https://arxiv.org/abs/1908.03190 arxiv 2019] | ||
# Physics-based deep learning [https://www.physicsbaseddeeplearning.org/intro-teaser.html code] | # Physics-based deep learning [https://www.physicsbaseddeeplearning.org/intro-teaser.html code] | ||
| − | # PINN by Steve Burton [https://www.youtube.com/watch?v=g-S0m2zcKUg&list=PLMrJAkhIeNNQ0BaKuBKY43k4xMo6NSbBa&index=3 yt] | + | # PINN by Steve Burton [https://www.youtube.com/watch?v=g-S0m2zcKUg&list=PLMrJAkhIeNNQ0BaKuBKY43k4xMo6NSbBa&index=3 yt] |
# Process Model Inversion in the Data-Driven Engineering Context for Improved Parameter Sensitivities [https://www.mdpi.com/2227-9717/10/9/1764 mdpi processes 2022] ('''nice connection pictures''') | # Process Model Inversion in the Data-Driven Engineering Context for Improved Parameter Sensitivities [https://www.mdpi.com/2227-9717/10/9/1764 mdpi processes 2022] ('''nice connection pictures''') | ||
# Physics-based Deep Learning [https://www.physicsbaseddeeplearning.org/intro.html github] | # Physics-based Deep Learning [https://www.physicsbaseddeeplearning.org/intro.html github] | ||
| Line 479: | Line 356: | ||
#[https://en.wikipedia.org/wiki/Fisher_information Fisher information] | #[https://en.wikipedia.org/wiki/Fisher_information Fisher information] | ||
# also dobrushin stratonovich wasserstein | # also dobrushin stratonovich wasserstein | ||
| − | # also fluid | + | # also fluid dynamics, transportation theory |
===Tutorials=== | ===Tutorials=== | ||
| Line 520: | Line 397: | ||
# State-space representation [https://en.wikipedia.org/wiki/State-space_representation wiki] | # State-space representation [https://en.wikipedia.org/wiki/State-space_representation wiki] | ||
# Phase space [https://en.wikipedia.org/wiki/Phase_space wiki] | # Phase space [https://en.wikipedia.org/wiki/Phase_space wiki] | ||
| + | |||
| + | === Collection=== | ||
| + | # Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting [https://arxiv.org/pdf/2405.16312 arxiv] | ||
| + | # Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics [arxiv] | ||
| + | # srush/annotated-s4 | ||
| + | # Modeling Nonlinear Dynamics from Equations and Data by George Haller [https://epubs.siam.org/doi/book/10.1137/1.9781611978353 book] [https://www.youtube.com/watch?v=mhcZaBMeA-U youtube] | ||
Latest revision as of 15:13, 10 February 2026
- Channel: The chat-link FDA group
Contents
- 1 Introduction
- 2 Topics to discuss
- 3 Table of topics for seminars
- 4 Practical spatial-time series
- 5 Basics
- 6 Tools
- 7 State Space Reconstruction
Introduction
The statistical analysis of spatial time series requires additional methods of data analysis. First, we suppose time is continuous, put the state space changes \(\frac{d\mathbf{x}}{dt}\), and use neural ordinary and stochastic differential equations. Second, we analyze a multivariate and multidimensional time series and use the tensor representation and tensor analysis. Third, since the time series have significant cross-correlation, we model them in the Riemannian space. Fourth, medical time series are periodic; the base model is the pendulum model, \(\frac{d^2x}{dt^2}=-c\sin{x}\). We use physics-informed neural networks to approximate data. Fifth, the practical experiments involve multiple data sources. We use canonical correlation analysis with a latent state space. This space aligns the source and target spaces and generates data in the source and target manifolds.
Applications
This field of Machine Learning applies to any field where the measurements have continuous time and space data acquired from multimodal sources: climate modeling, neural interfaces, solid-state physics, electronics, fluid dynamics, and many more. We will carefully collect both the theory and its practice.
Foundation models for scientific research
Foundation AI models are universal models to solve a wide set of problems. This project proposes to investigate the theoretical properties of foundation models. The domain to model is a spatial-time series. These data are used in various scientific disciplines and serve to generalise scientific knowledge and make forecasts. The essential problems, formulated as user requests that solve a foundation model, are forecasting and generation of time series; analysis and classification of time series; detection of change point, and causal inference. To solve these problems, the foundation AI models are trained on massive datasets. The main goal of this project is to compare various architectures of foundation models to find an optimal architecture that solves the listed problems for a wide range of spatial time series.
Topics to discuss
- State Space Models, Convolution, SSA, SSM (Spectral Submanifolds)
- Neural and Controlled ODE, Neural PDE, Geometric Learning
- Operator Learning, Physics-informed learning, and multimodeling
- Spatial-Temporal Graph Modeling: Graph convolution and metric tensors
- Riemmannian models; time series generation
- AI for science: mathematical modelling principles
- <it>Left behind:<\it> data-driven tensor analysis, differential forms, and spinors
State of the Art in 2025
The NeurIPS workshop "Foundational models for science" reflected this theme in 2024:
- Foundation Models for Science: Progress, Opportunities, and Challenges URL
- Foundation Models for the Earth system UPL, no paper
- Foundation Methods for foundation models for scientific machine learning URL, no paper
- AI-Augmented Climate simulators and emulators URL, no paper
- Provable in-context learning of linear systems and linear elliptic PDEs with transformers NIPS
- VSMNO: Solving PDE by Utilizing Spectral Patterns of Different Neural Operators NIPS
Physics problem Simulations
- The Well: a Large-Scale Collection of Diverse Physics Simulations for Machine Learning ArXiv, Code
- Polymatic Advancing Science through Multi‑Disciplinary AI blog
- Long Term Memory: The Foundation of AI Self-Evolution ArXiv
- Distilling Free-Form Natural Laws from Experimental Data, 2009 Science, comment, medium
- Deep learning for universal linear embeddings of nonlinear dynamics nature
- A comparison of data-driven approaches to build low-dimensional ocean models, 2021 by Pavel Berloff ArXiv, talk by Daniil Dorin for S.V. Fortova
- Applications of Deep Learning to Ocean Data Inference and Subgrid Parameterization by Thomas Bolton and Laure Zanna, 2018 preprint, talk by Nilita Kiselev
- On energy-aware hybrid models by Shevchenko,2024 doi, talk by Mariya Nikitina
- Science: NASA satellites and computers have provided us with these mesmerizing swirls that cover our planet—but this isn’t star stuff. Each color represents a different aerosol that was floating in the atmosphere above our heads from 1 August to 14 September 2024 video
Spatial-Temporal Graph Modeling
- Graph WaveNet for Deep Spatial-Temporal Graph Modeling ArXiv
- Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting ICLR
- Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting ArXiv SSMTool
- State Space Reconstruction for Multivariate Time Series Prediction ArXiv](Denis)
- Longitudinal predictive modeling of tau progression along the structural connectome by Joyita Dutta 2021
Key reviews on AI for Science
- 2018. Diffusion Convolutional Recurrent Neural Network ICLR
- 2021. Neural Partial Differential Equations with Functional Convolution ICLR
- 2018. Graph WaveNet for Deep Spatial-Temporal Graph Modeling ArXiV
- 2021. Neural Rough Differential Equations for Long Time Series (comparison)
- 2022. Time Series Forecasting Using Manifold Learning, Radial Basis Function Interpolation and Geometric Harmonics doi (all basic models + superpositions review)
Catch-up on LLM
If you are not familiar with the LLM and GPT:
- Build an LLM from scratch by Sebastian Raschka, 2025 github
- Agentic Design Patterns by Antonio Gulli, 2025 docx
For fun, the vibe coding: 1, 2, 3, 4, also Foundations of LLM and Karpathy's project nanoGPT
Table of topics for seminars
In these ten weeks, we will discuss the next five topics:
- Multimodal data
- Continous time and space models
- Physics-informed models
- Multilinear models
- Riemannian spaces
These items comprise the stochastic-deterministic decomposition. So the questions include three parts:
- deterministic model,
- generative model,
- stochastic-deterministic decomposition method.
Multimodal data
First series
- Canonical Correlation Analysis
- CCA in tensor representation
- Kernel CCA in Hilbert and L2[a,b] spaces
- CCA versus Cross-Attention Transformers
- Generative CCA, diffusion, and flow
- Comparative analysis of variants of CCA, like PLS and others
- Functional PCA
- Canonical Correlation Analysis: forecasting model and loss function with variants-
- CCA parameter estimation algorithm
Talks
- Canonical Correlation Analysis in tensor representation Marat
- Kernel CCA in Hilbert and L2[a,b] spaces Bair
- CCA versus Cross-Attention Transformers Eduard
- Generative CCA, diffusion, and flow Galina, Galina
- Functional PCA Parviz
Continous models
- Neural ODE
- Continous state space models
- Continous normalizing flows
- Adjoint method and continuous backpropagation
- Neural Delayed Differential Equations
- Neural CDE (PID control is welcome)
- Neural PDE
- S4 and Hippo models [1], [2] (LSSL, SaShiMi, DSS, HTTYH, S4D, and S4ND)
- Riemannian continuous models
Talks
- Continuous state space models Bair
- Continuous normalizing flows Marat
- Adjoint method and continuous backpropagation Galina
- Riemannian continuous models Eduard
SINDy
- Learning partial differential equations via data discovery and sparse optimization by Hayden Schaeffer, 2017 DOI, PDF
- Data-driven discovery of partial differential equations by Rudy et al. 2017 Science, []
- Supporting Information for: Discovering governing equations from data by Steven L. Brunton et al. [pnas.1517384113.sapp.pdf PDF]
- SINDy-PI: a robust algorithm for parallel implicit sparse identification of nonlinear dynamics by Kadierdan Kaheman et al., 2020 DOI
- Ensemble-SINDy by Fasel et al. 2021 DOI
Connection to NeurODE
- Hybrid Models: Combining Neural ODEs with Discrete Layers medium
- ODE manual. Linearization. Особые точки нелинейных систем на плоскости by Ilya Shurov URL Equilibrium points wiki
Physics-Informed models
Third series
- PINNs as multimodels
- Spherical harmonics in p dimensions (an IMU example is welcome)
- PDF and Physics-Informed learning
- Integral Transforms in Physics-Informed learning
Talks
Multilinear models and topology
Fourth series
- Clifford or Geometric algebra in machine learning
- Tensor models, tensor decomposition, and approximation (tensor PLS or CCA)
- Machine learning models for tensors: Field Equation (Yang-Mills Equations_
- Machine learning models for theoretical physics (Maxwell’s Equations, Navier-Stokes)
- Persistent homology and dimensionality reduction (say, arXiv:2302.03447 with embedding delays)
Talks
- Tensor models, tensor decomposition, and approximation Eduard
- Machine learning models for theoretical physics (Maxwell’s Equations, Navier-Stocks) Galina
Generative and Riemannian models
Fifth series
- Generative Riemannian models. How do we extract and use the distribution?
- Generative Canonical Correlation Analysis and its connection with the Riemannian spaces in the latent part
- Scoring-based Riemannian models. How do we extract and use the distribution?
- Generative convolutional models for tensors. Is there a continuous-time? (A variant is the Riemannian Residual Networks).
- Riemannian continuous normalizing flows. How do we generate a time series of a given distribution?
Talks
Operator learning
An additional topic to summarise all the above. See the introduction in
- Neural operators wiki
- Operator Learning: Convolutional Neural Operators blog
- Convolutional Neural Operators for robust and accurate learning of PDEs arxiv 2023
- Representation Equivalent Neural Operators: a Framework for Alias-free Operator Learning arxiv 2023
- PID: Proportional-Integral-Differential-equation modeling with operator learning
Discussed literature
- Generative CCA, diffusion, and flow by Galina [3] [4] [5] [6]
- Kernel CCA in Hilbert and L2[a,b] spaces by Bair [7] [8]
- CCA versus Cross-Attention Transformers by Eduard [9] [10] [11]
- Ajoint method and continuous backpropagation by Galina [12]
- Continuous normalizing flows by Galina [13]
- Tensor models by Eduard [14] [15] [16]
- Navier-Stokes [17] [18] [19]
- Classics versus quantum by Galina
- Schroedinger vs. Navier–Stokes 2016
- Many-particle quantum hydrodynamics: Exact equations and pressure tensors 2019
- Quantum hydrodynamics, Wigner transforms, the classical limit 1995
- Geometry of Nonadiabatic Quantum Hydrodynamics 2019
- Theory of quantum friction 2014
- Minimal quantum viscosity from fundamental physical constants
- Fluid Dynamics with Incompressible Schrödinger Flow 2017
- Гидродинамика Шрёдингера на пальцах
- Riemannian continuous normalizing flows by Galina [20] [21] [22]
Practical spatial-time series
Datasets
- ClimateSet, 2023 ArXiv
- A guide to state–space modeling of ecological time series, 2021 PDF, (Bayesian Kalman)
- Kalman Filtering and Smoothing, 2025 ArXiv (Riemannian Kalman)
General
- Artificial Intelligence for Science in Quantum, Atomistic, and Continuum Systems arxiv 2023
- Algebra, Topology, Differential Calculus, and Optimization Theory For Computer Science and Machine Learning upenn 2024
- The Elements of Differentiable Programming arxiv 2024
- The list from the previous year 2023.
- Differential Geometry of Curves and Surfaces: Textbook, 2016 by Kristopher Tapp [23]
Basic literature
- Understanding Deep Learning by Simon J.D. Prince mit 2023
- Deep Learning by C.M. and H. Bishops Springer 2024 (online version)
- A Geometric Approach to Differential Forms by David Bachman arxiv 2013
- Advanced Calculus: Geometric View by James J. Callahan pdf 2010, collection
- Geometric Deep Learning by Michael M. Bronstein arxiv 2021
Linear and bilinear models
- A Tutorial on Independent Component Analysis arxiv, 2014
- On the Stability of Multilinear Dynamical Systems arxiv 2022
- Tensor-based Regression Models and Applications by Ming Hou Thèse Uni-Laval 2017
- Tensor Canonical Correlation Analysis for Multi-view Dimension Reduction [24] (Semkin)
- Tensor Learning in Multi-view Kernel PCA arxiv 2018
- Tensor decomposition of EEG signals: A brief review 2015
Spherical Harmonics
- Spherical Harmonic Transforms: In JAX and PyTorch Medium 2024
- Spherical Harmonics in p Dimensions arxiv 2012
- Physics of simple pendulum: a case study of nonlinear dynamics RG 2008
- Time series forecasting using manifold learning, 2021 arxiv
- Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics 2022 Chaos AIP
State Space Models
- Missing Slice Recovery for Tensors Using a Low-rank Model in Embedded Space arxiv 2018
- Legendre Memory Units: Continuous-Time Representation in Recurrent Neural Networks by A.R. Voelker et al., 2019 NeurIPS
SSM Generative Models
- Masked Autoregressive Flow for Density Estimation arxiv 2017
SSM+Riemann+Gaussian process regression
- Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics by Ioannis G. Kevrekidis,3 and Constantinos Siettos, 2022 pdf
Physics-Informed Neural Networks
- Neural partial differential equations with functional convolution ICLP
- Solving PDEs by variational physics-informed neural networks: an a posteriori error analysis PDF plus several links to the books on the subject inside
- Predicting the nonlinear dynamics of spatiotemporal PDEs via physics-informed informer networks PDF
- Three ways to solve partial differential equations with neural networks — A review arxiv 2021
- NeuPDE: Neural Network Based Ordinary and Partial Differential Equations for Modeling Time-Dependent Data arxiv 2019
- Physics-based deep learning code
- PINN by Steve Burton yt
- Process Model Inversion in the Data-Driven Engineering Context for Improved Parameter Sensitivities mdpi processes 2022 (nice connection pictures)
- Physics-based Deep Learning github
- Integral Transforms in a Physics-Informed (Quantum) Neural Network setting arxiv 2022
- Lectures ny Stephen Brunton AI/ML+Physics, Part 4, Basic PDEs, PDE Overview
PINN Libraries
- PINA Gianluigi Rozza at SISSA MathLab www see tutorials and solvers
Riemmanian models
- Riemannian Continuous Normalizing Flows arxiv 2020
- Residual Riemannian Networks arxiv 2023
Continous time, Neural ODE
- Neural Spatio-Temporal Point Processes by Ricky Chen et al. iclr 2021 (likelihood for time and space)
- Neural Ordinary Differential Equations by Ricky Chen et al. arxiv 2018 torchdiffeq github
- Neural Controlled Differential Equations for Irregular Time Series 'Patrick Kidger et al. arxiv 2020github
- On Neural Differential Equations by Patrick Kidger arxiv 2021
- Diffusion Normalizing Flow arxiv 2021
- Differentiable Programming for Differential Equations: A Review arxiv 2024
- (code tutorial) Deep Implicit Layers - Neural ODEs, Deep Equilibrium Models, and Beyond nips 2020
- (code tutorial) 2021
- Neural CDE and tensors IEEE, IEEE
- Latent ODEs for Irregularly-Sampled Time Series 2019
- Apprentissage et calcul scientifique by Emmanuel Franck www draft of a texbook, chapter 11.4
- Adjoint State Method, Backpropagation and Neural ODEs by Ilya Schurov www
Graph and PDEs
- Fourier Neural Operator for Parametric Partial Differential Equations arxiv 2020
- Masked Attention is All You Need for Graphs arxiv 2024
Neural SDE
- Approximation of Stochastic Quasi-Periodic Responses of Limit Cycles in Non-Equilibrium Systems under Periodic Excitations and Weak Fluctuations mdpi entropy 2017 (great illustrations on the stochastic nature of a simple phase trajectory)
- Approximation of Stochastic Quasi-Periodic Responses of Limit Cycles in Non-Equilibrium Systems under Periodic Excitations and Weak Fluctuations mdpi entropy 2017 (great illustrations on the stochastic nature of a simple phase trajectory)
- Neural SDEs for Conditional Time Series Generation arxiv 2023 code github LSTM - CSig-WGAN
- Neural SDEs as Infinite-Dimensional GANs 2021
- Efficient and Accurate Gradients for Neural SDEs by Patrick Kidger arxiv 2021 code diffrax
Chains and homology
- Operator Learning: Algorithms and Analysis arxiv 2024
- Hi-res weather: Operator learning arxiv 2022
- Homotopy theory for beginners by J.M. Moeller ku.dk 2015 (is it a pertinent link?)
- Explorations in Homeomorphic Variational Auto-Encoding arxiv 2018
- Special Finite Elements for Dipole Modelling master thesis Bauer 2011
- Selecting embedding delays: An overview of embedding techniques and a new method using persistent homology arxiv 2023 (Denis)
- (code) Clifford Algebra for Python https://clifford.readthedocs.io/en/v1.1.0/
Appendix
- Neural Memory Networks stanford reports 2019
- An Elementary Introduction to Information Geometry by Frank Nielsen [An Elementary Introduction to Information Geometry Frank Nielsen mdpi entropy
- The Many Faces of Information Geometry by Frank Nielsen ams 2022 (short version)
- Geometric Clifford Algebra Networks arxiv 3022
- Clifford Algebras and Dimensionality Reduction for Signal Separation by M. Guillemard Uni-Hamburg 2010code
- Special Finite Elements for Dipole Modelling by Martin Bauer Master Thesis Erlangen 2012 diff p-form must read
- Bayesian model selection for complex dynamic systems 2018
- Visualizing 3-Dimensional Manifolds by Dugan J. Hammock 2013 umass
- At the Interface of Algebra and Statistics by T-D. Bradley arxiv 2020
- Time Series Handbook by Borja, 2021 github
- Physics-informed machine learning Nature reviews: Physics 2021
- Integral Transforms in a Physics-Informed (Quantum) Neural Network setting: Applications & Use-Cases arxiv 2022
- Deep Efficient Continuous Manifold Learning for Time Series Modeling arxiv 2021
Causality
- Toward Causal Representation Learning 2021
- See the Sugihara collection
Basics
Collection of wiki-links
Time Series
- Spectral submanifold (with nonlinear dimensional reduction like som)
- Lagrangian coherent structure (software below)
Signal Processing
- Estimation of signal parameters via rotational invariance techniques
- Reproducing kernel Hilbert space
- Kernel principal component analysis
- Gram matrix
- Generalized pencil-of-function method
- Wavelet transform
Differential Geometry
- Pushforward (differential)
- Ffibers, Bundles, Sheaves
- Homology
- Topological data analysis
- Conditional mutual information
- Convergent cross mapping
- Differential form
- The total derivative as a differential form
- #Riemannian_metrics Riemannian_metrics
- Multidimensional Differential and Integral Calculus: A Practical Approach (textbook)
Probabilistical Decompisition
- Wasserstein metric
- Mutual information
- Jacobian
- Fisher information
- also dobrushin stratonovich wasserstein
- also fluid dynamics, transportation theory
Tutorials
- Connected papers search
- Operator Learning via Physics-Informed DeepONet: Let’s Implement It From Scratch Medium
Tools
- icebeem
- ivae
- fmri-component
- analysis/blob/master/VAE_for_fMRI/dataset/train/Bystrova0_y-axis.png
- Neural ODE in Matlab
- pyRiemann
- causality inference peps
- LMM grok-1 with weights
Turbulence
Physics and Engineering of Turbulence
- Fundamentals of Fluid_Mechanics, 2013 PDF
- Introduction ot Fluid Mechanics, 2004 by R. Fox et al. PDF
- Computational fluid dynamics, 1995 by John D. Anderson, Jr. PDF
- Fluid-dynamic drag, 1965 by S.F. Hoerner PDF
- TorchDyn: A Neural Differential Equations Library arXiv github
State Space Reconstruction
(out of this topic)
- The false nearest neighbors algorithm by Carl Rhodes doi 1997, Carl's another paper
- Use of False Nearest Neighbours for Selecting Variables andEmbedding Parameters for State Space Reconstruction by Anna Krakovská et al. doi 2014
- Estimating a Minimum Embedding Dimension by False Nearest Neighbors Method without an Arbitrary Threshold doi 2022
Author’s Name: Kohki Nakane1,a), Akihiro Sugiura2, Hiroki Takada1
- ODE. Differential manifolds by Vladimir Arnold (last chapter of the textbook)
- ODE by Ilya Shchurov www
- State-space representation wiki
- Phase space wiki
Collection
- Time-SSM: Simplifying and Unifying State Space Models for Time Series Forecasting arxiv
- Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics [arxiv]
- srush/annotated-s4
- Modeling Nonlinear Dynamics from Equations and Data by George Haller book youtube