Functional Data Analysis
Channels
When
- September 4, 11, 18, 25 — Thursdays at 10:30 · m1p.org/go_zoom
 - October (most likely) — Saturdays at 10:30
 
Foundation models for spatial-time series
Foundation AI models are universal models for a wide set of problems. This project investigates their theoretical properties on spatial-time series—data used across sciences to generalize knowledge and make forecasts. Core user-level tasks: forecasting and generation of time series; analysis and classification; change-point detection; causal inference. These models are trained on massive datasets. Our goal is to compare architectures to find an optimal one that solves the above for a broad range of spatial time series.
Functional data analysis
We assume continuous time and study state-space changes $\frac{d\mathbf{x}}{dt}$ via neural ODEs/SDEs. We analyze multivariate/multidimensional series with tensor representations; model strong cross-correlations in Riemannian spaces. Many medical series are periodic; the base model is the pendulum $\frac{d^2 x}{dt^2} = -c\sin x$. We use physics-informed neural networks (PINNs). Practical experiments involve multiple sources; we use canonical correlation analysis with a latent state space to align source/target manifolds and enable generation in both.
Applications
Any field with continuous time/space data from multimodal sources: climate, neural interfaces, solid-state physics, electronics, fluid dynamics, and more. We collect both theory and practice.
Fall 2025: Foundation models for time series
Topics to discuss
- State Space Models, Convolution, SSA, SSM (Spectral Submanifolds)
 - Neural & Controlled ODE, Neural PDE, Geometric Learning
 - Operator Learning, Physics-informed learning, multimodeling
 - Spatio-Temporal Graph Modeling: graph convolution & metric tensors
 - Riemannian models; time series generation
 - AI for science: mathematical modeling principles
 
Outside the course: data-driven tensor analysis, differential forms, spinors.
State of the Art in 2025
In December 2024, a NeurIPS workshop “Foundational models for science” reflected this theme:
- Foundation Models for Science: Progress, Opportunities, and Challenges — URL
 - Foundation Models for the Earth system — UPL, no paper
 - Foundation Methods for foundation models for scientific machine learning — URL, no paper
 - AI-Augmented Climate simulators and emulators — URL, no paper
 - Provable in-context learning of linear systems and linear elliptic PDEs with transformers — NIPS
 - VSMNO: Solving PDE by Utilizing Spectral Patterns of Different Neural Operators — NIPS PDF
 
March 2025 Physics Problem Simulations
- The Well: a Large-Scale Collection of Diverse Physics Simulations for ML — arXiv · Code
 - Polymathic: Advancing Science through Multi-Disciplinary AI — blog
 - Long Term Memory: The Foundation of AI Self-Evolution — arXiv
 - Distilling Free-Form Natural Laws from Experimental Data (2009) — Science · comment · medium
 - Deep learning for universal linear embeddings of nonlinear dynamics — Nature
 - A comparison of data-driven approaches to low-dimensional ocean models (2021) — arXiv
 - Applications of DL to Ocean Data Inference & Subgrid Parameterization (2018) — preprint
 - On energy-aware hybrid models (2024) — doi
 
Spatial-Temporal Graph Modeling
- Graph WaveNet — arXiv
 - Diffusion Convolutional Recurrent Neural Network (DCRNN) — ICLR
 - Time-SSM: Simplifying & Unifying State Space Models — arXiv
 - State Space Reconstruction for Multivariate Time Series — arXiv
 - Longitudinal predictive modeling of tau progression — NeuroImage 2021
 
Work arrangements
| Week | Date | Theme | Delivery | 
|---|---|---|---|
| 1 | Sep 4 | Preliminary discussion — pdf | |
| 2 | Sep 11 | Problem statement — pdf | |
| 3 | Sep 18 | Preliminary solution | Group talk & discussion | 
| 4 | Sep 25 | Minimum deployment | Group report | 
| 5 | Oct 4+ | FDA | Personal talks | 
| 13 | Nov 29 | Final discussion | Group talks | 
Structure of seminars
The semester lasts 12 weeks; six alternate weeks are for homework.
- Odd week: topic intro + homework theme handout.
 - Every week: essay discussion; collect improvement list.
 - Odd week: discuss improved essays; integrate into a joint structure.
 
Scoring
Group activity: cross-ranking with Kemeny median. Personal talks contribute to score.
Week 3 — Homework 1
- Form a group.
 - Discuss goals and a solution ([see the problem statement]).
 - Review solution approaches.
 - Select an LLM-GPT.
 - 
    
Run the code; verify it works.
- Store code in the group repository.
 - Store slides/report as well.
 
 - 
    
Make a 10-minute talk covering:
- Functionality and architecture of the model.
 - Why you selected this model.
 - Alternative models considered.
 
 
Requirements for the text & discussion
- Comprehensive explanation of the discussed method/question.
 - Principles only; no experiments.
 - ~Two pages.
 - Target reader: 2nd–3rd-year student.
 - One figure is mandatory.
 - Brief reference to DL structure is welcome.
 - Talk may be a slide or the text itself.
 - References with DOIs.
 - State how it was generated.
 - Note observed gaps to revisit later.
 
Style remarks for the essays
Automatic text generation raises the bar for clarity and authorship. Use generative tools to train persuasion skills; write for a thesis defense committee.