Overview¶
In this application, we are interested in modeling temperatures, and in particular, temperature extremes.
Quick Walk-Through¶
Datasets.¶
The first set of datasets will feature a range of different
Global Mean Surface Temperature Anomaly. This is our first univariate time series that we will be using to explore data-driven modeling techniques. It is a basic dataset which captures the change in mean surface temperature over the Earths surface. It has already been pre-processed to remove all artifacts and
Equations
Sub-Topics¶
Basics.
We will feature some basics of the Bayesian modeling workflow.
There will be some basics like defining a prior, likelihood, and posterior.
We will also have.
All code will be done with a high-level PPL, i.e., numpyro.
Bayesian Hierarchical Modeling. For many small datasets, we have a lot of hierarchical dependencies. By introducing the notion of a fully-pooled, non-pooled, and partially-pooled model for a simple model, we will introduce the notion of a latent-variable model. This will be an underlying approach for all remaining models throughout these applications.
Game of Dependencies. Many people do not have a solid grasp about dependencies within spatiotemporal data. We will walk-through this step-by-step by slowly introducing more complex representations, e.g., IID, time conditioning, dynamical models, and state-space models.
Extreme Values. Extreme values are a worlds apart from the standard mean and standard deviation. In this section, we will introduce this from fir We will introduce this from the perspective of Temporal Point Processes (TPPs). Some staple distributions will include the Poisson Process, the GEVD, and the GPD.
Marked Temporal Point Processes. The end-goal for extreme values (and many schemes in general) is to make predictions, i.e., what is the probability that there will be an extreme event, λ, with a certain intensity, , at a certain location, , or time, , given historical observations, , and some covariate parameter, .
Useful Skills¶
Unstructured Measurements. We will immediately start dealing with unstructured datasets which are ubiquitous in the geosciences. We will learn how to model data with an unstructured representation. We will also learn how we can transform them to a more structured representation which is more useful for other other applications.
Real-Time Applications. We will be dealing with time series directly. This gives us the advantage that we will be able to train our models, use them for predictions, and then update them as we obtain more observations.
Immediate Applicability. We will have some immediate applicability because extreme events are things that we have to deal with on a daily basis.
Topics¶
Datasets¶
We will use some standard datasets that are found within the AEMET database.
- Covariante,
- Global Mean Surface Temperature Anomaly (GMSTA)
- Measurement, :
- Daily Mean Temperature (TMax)
- Daily Maximum Temperature (TMean)
Global Mean Surface Temperature¶
This is an anomaly dataset.
Time
Weather Station Data Spain¶
Our coordinates are as follows.
We can also showcase the discretized version
Time
Spatial Coordinates
Variables
Extreme Events¶
We will use some standard techniques to extract the extreme events from the time series.
- Block Maximum (BM)
- Peak-Over-Threshold (POT)
- Temporal Point Process (TPP)
Data Likelihood¶
We need to decide which likelihood we want to use. For the mean data, we can assume that it follows a Gaussian distribution or a long-tailed T-Student distribution. However, for the maximum values, we need to assume an EVD like the GEVD, GPD or TPP.
- Mean -> Gaussian, T-Student
- Maximum -> GEVD, GPD, TPP
Parameterizations¶
We have to decide the parameterization we wish to use.
- IID
- Hierarchical
- Strong-Constrained Dynamical, i.e., Neural ODE/PDE
- Weak-Constrained Dynamical, i.e., SSM, EnsKF
State Estimation¶
- Weak-Constrained
- Gradient-Based - 4DVar
- Derivative-Free - Ensemble Kalman Filter
- Minimization-Based - DEQ
Predictions¶
Our final task is to use a our new model to make predictions.
- Feature Representation
- Interpolation
- HindCasting
- Forecasting
Questions¶
Parameter Estimation
- Naive Density Estimator,
- Can we maximize the data likelihood?,
- What are the distribution of the parameters?
- What parametric form fits the best?
- How sensitive is it to the initial condition?
- What is the predictive uncertainty?,
- Can we introduce structure to reduce the complexity? LocalLinear, AutoRegressive, Cycle
Representation
- What temporal resolution for GEVD?, e.g. Decade, Year, Season, Month
- What temporal resolution for GPD?, e.g., Day, 3 Days, 5 Days, 7 Days
- What spatial resolution for GEVD?, e.g., Individual Stations, Spatial Clusters, Coarse Grid, Fine Grid
Predictions
- Can we forecast? Does it make sense?
- Can we hindcast? Does it makes sense?
Appendix¶
There are various things we need to go over which will help us deal with extreme events.