Combined Approximation (Taylor + MacLaurin) - Relative/Normalized Radiance
Introduction: The Operational Standard¶
The relative combined linearized model represents the most widely used formulation in operational atmospheric methane detection. This model applies two sequential approximations—Taylor expansion of optical depth followed by MacLaurin approximation of the exponential—to the normalized Beer-Lambert law, resulting in a fully linear relationship between observations and methane enhancement.
Historical Context and Adoption¶
This formulation emerged from the practical needs of satellite and airborne hyperspectral methane detection in the 2010s. Early systems struggled with:
Computational constraints: Processing millions of pixels in real-time
Calibration uncertainties: Absolute radiometric calibration errors of 5-10%
Scene variability: Surface reflectance variations of 2-10× across footprints
The relative combined model addressed all three challenges by providing a fast, calibration-independent, linear detection algorithm with acceptable accuracy for moderate plumes.
Key operational systems using this approach:
AVIRIS-NG (NASA/JPL airborne)
GHGSat constellation (commercial satellites)
Sentinel-5P TROPOMI (in some retrieval algorithms)
EnMAP methane product
Carbon Mapper constellation
Relationship to Model Family¶
This model sits at the “sweet spot” in the approximation hierarchy:
Exact ←────────── Approximation Level ──────────→ Simplified │ │ Nonlinear Taylor Only Combined MacLaurin Only (Models 1, 1B) (Models 3, 3B) (Models 4, 4B) (Models 2, 2B) │ │ │ │ Most Accurate Good Accuracy Practical Theoretical Slowest Moderate Speed Fast Fastest Iterative Iterative Closed-Form Closed-Form All Plumes τ < 0.3 τ < 0.1 τ < 0.05
Model 4B combines:
Taylor expansion (Model 3B): Linearizes optical depth around background → handles concentration perturbations
MacLaurin approximation (Model 2B): Linearizes exponential transmittance → enables closed-form solution
Normalization: Divides by background radiance → eliminates calibration dependencies
Result: A fully linear, calibration-independent, closed-form detection algorithm.
The Two Approximations: Physical Justification¶
Approximation 1: Taylor Expansion of Optical Depth¶
Mathematical form:
Physical justification:
Optical depth is defined as:
where is path coordinate. For well-mixed plumes over a sensor footprint:
This is exactly linear in VMR (no approximation needed for itself). The Taylor expansion becomes exact:
because (second derivative vanishes).
Key insight: This isn’t really an approximation—it’s an exact representation of the linear relationship between concentration and optical depth. The “approximation” comes from assuming:
Spatial homogeneity: VMR constant over footprint
Vertical well-mixing: Concentration uniform through boundary layer
Path-averaged properties: Single effective ,
When it breaks down:
Vertically stratified plumes (elevated sources)
Strong horizontal gradients within pixel
Sub-pixel plume structure
Typical accuracy: 1-5% error for boundary layer plumes, 10-20% for elevated plumes
Approximation 2: MacLaurin Expansion of Exponential¶
Mathematical form:
Physical justification:
The exponential represents the multiplicative attenuation through the atmosphere. For weak absorption:
Full series:
Truncate at first order:
Truncation error:
Relative error:
(for small )
Numerical accuracy table:
| Fractional Absorption | Relative Error | Acceptable? | |
|---|---|---|---|
| 0.01 | 1.0% | 0.005% | ✓ Excellent |
| 0.05 | 4.9% | 0.13% | ✓ Very good |
| 0.10 | 9.5% | 0.5% | ✓ Good (threshold) |
| 0.15 | 13.9% | 1.1% | ⚠ Marginal |
| 0.20 | 18.1% | 2.0% | ✗ Poor |
| 0.30 | 25.9% | 4.7% | ✗ Very poor |
Operational threshold: Most systems require (< 10% absorption) for this approximation.
Physical interpretation of validity:
The MacLaurin approximation assumes weak absorption where:
Most photons survive:
Linear regime: Fractional change proportional to optical depth
No saturation: Absorption doesn’t significantly deplete the radiation field
When it breaks down:
Strong absorption lines (line centers at high resolution)
Very large plumes (>2000 ppm for typical SWIR conditions)
Long path lengths (thick boundary layers, oblique viewing)
Typical accuracy: 0.5% for , degrades rapidly beyond
Combined Effect: Why Two Approximations Work Together¶
Synergistic simplification:
Starting from exact normalized radiance:
Apply both approximations:
Result: Linear in !
Computational advantage:
Exact model: Requires iterative nonlinear optimization (5-20 iterations)
Taylor only (Model 3B): Still requires iteration (exponential is nonlinear)
MacLaurin only (Model 2B): Linear but uses total optical depth (less accurate)
Combined: Closed-form solution (single matrix operation)
Speed comparison:
Exact: ~10-100 ms/pixel (iterative)
Taylor only: ~5-20 ms/pixel (fewer iterations)
Combined: ~0.1-1 ms/pixel (direct solve)
MacLaurin only: ~0.1-1 ms/pixel (but worse accuracy)
For a 1000×1000 pixel hyperspectral cube:
Exact: ~3-28 hours
Taylor only: ~1.5-6 hours
Combined: ~2-17 minutes ← Practical for operations
MacLaurin only: ~2-17 minutes (but unacceptable errors)
This 100× speedup enables real-time processing essential for operational systems.
Accuracy vs. Computational Cost Trade-off¶
The Pareto frontier: Accuracy ↑ │ Exact Nonlinear ● │ ╲ │ ╲ │ ╲ Taylor Only │ ● │ ╲╲ │ ╲╲ │ Combined ●●●● ← Optimal trade-off │ ╲╲╲╲ │ ╲╲╲╲ │ MacLaurin Only ● └────────────────────────────────────────→ Speed
Model 4B occupies the “knee” of the curve: Maximum gain in speed for minimum loss in accuracy.
Quantitative comparison for ppm (typical moderate plume):
| Model | Retrieval Error | Computation Time | Operational? |
|---|---|---|---|
| Exact Nonlinear | 0% (reference) | 50 ms/pixel | ✗ Too slow |
| Taylor Only | ~2-5% | 10 ms/pixel | ⚠ Borderline |
| Combined (4B) | ~5-10% | 0.5 ms/pixel | ✓ Optimal |
| MacLaurin Only | ~15-20% | 0.5 ms/pixel | ✗ Too inaccurate |
Decision rationale for operational systems:
5-10% retrieval error is acceptable (within other uncertainties like wind speed for flux estimation)
100× speedup is essential for real-time large-area surveys
Combined model provides best balance
1. Concentration (Background + Enhancement)¶
Same as all models:
No approximation applied to this fundamental relationship.
2. Optical Depth + Jacobian¶
Optical Depth (Taylor Expansion Applied)¶
Background optical depth (exact):
Differential optical depth (Taylor expansion around background):
Since is linear in VMR, the Taylor expansion is exact:
Explicit form:
Units: dimensionless
Physical interpretation:
is the additional optical depth due to plume enhancement only
Independent of background conditions ( doesn’t appear)
Linear in enhancement:
This is where the first “approximation” enters, but it’s really the assumption that:
Validity requires:
Spatially uniform enhancement over sensor footprint
Single effective path length (not vertically varying)
Constant atmospheric state (, uniform → , uniform)
Typical operational conditions where this holds:
Boundary layer plumes (well-mixed below 1-2 km)
Footprint sizes 30-60 m (smaller than typical plume scale)
Moderate wind conditions (promotes mixing)
Where it breaks down:
Elevated plumes with vertical structure
Sub-footprint plume edges (sharp gradients)
Very large footprints (>100 m) relative to plume size
Jacobian of Optical Depth¶
Units: ppm⁻¹
Key property: Constant (independent of VMR, exact for linear )
This constant Jacobian is what enables the closed-form solution later.
3. Transmittance + Jacobian¶
Transmittance (Both Approximations Combined)¶
Normalized transmittance (definition, exact):
Apply MacLaurin approximation:
With Taylor-expanded :
Units: dimensionless
Physical interpretation of combined approximation:
Starting from exact physics:
The combined approximation replaces:
Exponential decay → Linear decay
Multiplicative attenuation → Additive reduction
Why this works for weak absorption:
Plot of vs. :
1.0 ●────────────────
│╲
│ ╲ Exponential
│ ╲ ───
0.9 │ ● Linear approximation
│ ╲ ─ ─ ─
│ ╲
│ ╲
0.8 │ ● Good match
│ ╲ for x < 0.1
│ ╲
0.7 │ ●
└─────────────→ Δτ
0 0.1 0.2 0.3
For : curves nearly identical → approximation excellent.
Saturation effects (what we’re missing):
The exponential captures that:
As : (complete absorption, physical)
Linear : Can become negative for (unphysical)
The MacLaurin approximation assumes we stay far from saturation ().
Operational impact:
For : Excellent approximation, <1% error
For : ~2% error, acceptable for screening
For : >5% error, should use Model 3B or 1B instead
Jacobian of Transmittance¶
Explicit form:
Units: ppm⁻¹
Key property: Constant - does not vary with VMR or optical depth.
Physical consequence:
The constant Jacobian means the model predicts:
No saturation: Sensitivity same at 100 ppm as at 5000 ppm
Linear response: Double the VMR → double the signal
Reality (from exact model):
Saturation occurs: decreases as increases
Nonlinear response: Signal gain decreases for strong plumes
Operational consequence:
For moderate plumes ():
Exact: → Sensitivity reduced by 9.5%
Combined model: Assumes sensitivity at 100% → Slight overestimate
This causes a systematic bias: The combined model will slightly underestimate VMR for moderate-to-strong plumes because it overestimates sensitivity.
Correction factor:
Empirical studies show the bias is approximately:
For : Bias ≈ 0.5% underestimation (negligible). For : Bias ≈ 2% underestimation (small).
This is acceptable given other uncertainties (wind speed, path length, etc. often 10-20%).
4. Beer’s Law + Jacobian¶
Beer’s Law (Normalized, Linearized Form)¶
Starting from normalized radiance:
Apply combined approximation:
Units: dimensionless
This is the fundamental forward model: Linear relationship between normalized radiance and VMR enhancement.
Physical interpretation:
Three components:
Unity baseline: Background normalized to 1 (by definition)
Sensitivity factor: Depends on molecular physics (), atmosphere (), geometry (, AMF)
Plume strength: The unknown to estimate
Advantages of this form:
Calibration-free: No or appears
Scene-independent: Doesn’t depend on or
Linear: Simple structure enables closed-form inversion
Physically interpretable: Each term has clear meaning
Computationally efficient: Single matrix multiply for forward model
Comparison across model hierarchy:
| Model | Forward Model | Complexity | Accuracy |
|---|---|---|---|
| Exact | Nonlinear | Highest | |
| Taylor only | Nonlinear | High | |
| Combined | Linear | Good | |
| MacLaurin only | Linear | Poor |
The combined model achieves the optimal balance.
Jacobian of Normalized Radiance¶
Units: ppm⁻¹
Key insight: The Jacobian is the negative of the sensitivity factor.
Wavelength dependence:
Since varies strongly with wavelength:
The Jacobian is a vector with different values at each wavelength:
Strong absorption lines: Large (high sensitivity)
Weak lines or continuum: Small (low sensitivity)
This spectral structure is what enables detection: The plume signal has a characteristic spectral shape matching the methane absorption spectrum.
Matched filter exploits this by correlating observations with the expected spectral signature.
5. Observations¶
Observation Model¶
With combined approximation:
where:
(unknown, ppm)
(known, ppm⁻¹)
: vector of ones (n_wavelengths,)
: Gaussian noise
Model structure: This is a standard linear Gaussian model:
where is the known mean.
This enables textbook statistical inference:
Maximum likelihood estimation
Least squares
Bayesian inference
Hypothesis testing
All have closed-form analytical solutions.
Constructing Normalized Observations¶
From absolute radiance:
Estimating background :
Method 1: Spatial background (most common)
Select background pixels in scene (no plume)
Options:
Median: Robust to outliers
Mean of lowest 10%: Conservative (avoids plume contamination)
Gaussian mixture model: Identify background cluster
Typical: 50-100 background pixels averaged
Method 2: Temporal background
Same location at different time (before/after plume)
Requires persistent observation or plume variability
Method 3: Model-based
Predict from atmospheric/surface model
Less common (introduces model errors)
Operational consideration:
Choice of background estimation has large impact on performance:
Too few pixels: Noisy background estimate
Contaminated pixels: Biased background (underestimates plumes)
Different surface types: Spectral differences can mimic plumes
Best practice: Use spatially local, spectrally matched background from plume-free regions with median aggregation.
Noise Characteristics¶
Normalized noise covariance:
If absolute noise is :
For diagonal :
Physical interpretation:
Darker scenes (small ): Higher fractional noise
Brighter scenes (large ): Lower fractional noise
Signal-to-noise ratio inversely proportional to scene brightness in absolute units
But fractional SNR is scene-independent in normalized units
Typical values:
Bright scene: (0.1% fractional noise)
Moderate scene: (0.3% fractional noise)
Dark scene: (1% fractional noise)
Operational impact:
Detection easier over bright surfaces (deserts, snow)
Harder over dark surfaces (ocean, dense vegetation)
This is a fundamental physics constraint, not a model limitation
6. Measurement Model¶
Forward Measurement Model (Linear)¶
where:
Linearity is the key:
Actually, better to write:
So the deviation from background is linear in :
Inverse Problem: Maximum Likelihood Estimation¶
Likelihood function:
Log-likelihood:
Maximize by taking derivative:
Solve for :
Units: ppm
This is a closed-form, single-step solution! No iteration, no convergence issues, no initial guess needed.
Computational cost breakdown:
For wavelengths:
Compute innovation: →
Matrix-vector product: → (or if diagonal)
Inner products: →
Division: Scalar / scalar →
Total: general, for diagonal covariance
Typical values:
wavelengths
Diagonal covariance: ~200 operations
Time: ~0.1-1 ms/pixel on modern CPU
For 1M pixel scene: ~2-17 minutes total
Compare to iterative methods:
Gauss-Newton: 10-20 iterations × per iteration
Time: ~50-100× slower
This speed enables real-time operations for airborne and satellite missions.
Statistical Properties of the Estimate¶
Variance:
Standard error:
This is the Cramér-Rao lower bound: No unbiased estimator can have lower variance. The matched filter achieves optimal statistical efficiency.
For white noise ():
Physical interpretation:
Uncertainty decreases with stronger absorption (larger )
Uncertainty decreases with more wavelengths ( increases with )
Uncertainty proportional to noise level ()
Typical uncertainty:
Good conditions: ppm
Moderate conditions: ppm
Poor conditions (dark surface, high noise): ppm
Detection threshold:
For 3σ detection: Need → Minimum detectable enhancement ~150-600 ppm depending on conditions.
7. Taylor Expanded Measurement Model (Useful for 3DVar)¶
Exact Representation (No Additional Approximation)¶
Since the forward model is already linear, the “Taylor expansion” is exact:
This is not an approximation—it’s the exact representation of a linear function.
At background ():
Linearized model:
Innovation:
This is already in the standard linear form for data assimilation.
3DVar Formulation¶
Cost function with prior constraint:
where:
: background (prior) VMR enhancement estimate (typically 0)
: background error variance (prior uncertainty, units: ppm²)
Physical interpretation:
First term: Penalizes deviation from prior estimate (regularization)
Second term: Penalizes misfit to observations (data fitting)
Balance controlled by ratio
Analytical solution:
Taking derivative and setting to zero:
Rearranging:
For no prior plume ():
Posterior variance:
Physical interpretation:
Combines prior uncertainty () and observational constraint ()
Posterior uncertainty always less than or equal to both prior and observation-only uncertainties
As (uninformative prior): (matched filter)
As (perfect prior): (ignore observations)
Operational use:
Include prior when plume location/strength expected (e.g., known facility)
Use uninformative prior () for blind detection → reduces to matched filter
8. Pedagogical Connection to Matched Filter¶
From Linear Model to Matched Filter¶
The matched filter is simply the maximum likelihood estimator for the linear-Gaussian model with no prior constraint:
No approximation involved in this step—it’s a direct consequence of the model structure.
The Matched Filter Formula¶
Define target spectrum:
Note the sign: We define as the negative of so that (absorption causes positive values in standard convention).
Matched filter estimate:
where (note sign: we want positive for absorption).
For white noise:
Physical interpretation:
Numerator: Projection of observation onto target (correlation)
Denominator: Power of target (normalization)
Result: “How much of the target is present in the observation?”
Detection Statistic¶
Define detection statistic:
Relationship to estimate:
Under null hypothesis (, no plume):
Standard deviation:
For white noise:
Detection threshold:
Declare plume if where:
: 3σ (99.7% confidence, 0.3% false alarm rate)
: 4σ (99.99% confidence)
: 5σ (99.9999% confidence)
Operational choice: Typically balancing false alarms vs. missed detections.
Why It’s Called “Matched” Filter¶
The filter is matched to the expected signal in two senses:
Spectral matching: The target has the same spectral shape as the expected plume signature (methane absorption spectrum)
Statistical matching: The weighting by whitens the noise, making the filter optimal in the signal-to-noise ratio sense
Matched filter theorem: For detecting a known signal in Gaussian noise, the matched filter maximizes SNR and minimizes probability of error.
Computational Implementation¶
Pseudocode for operational system:
## Preprocessing (once per scene)
L_bg = median(background_pixels, axis=0) # #Estimate background
y_norm = y_absolute / L_bg # Normalize radiance
## Matched filter (per pixel)
d = 1 - y_norm # Innovation
numerator = t.T @ Sigma_inv @ d # Correlation
denominator = t.T @ Sigma_inv @ t # Normalization
alpha_hat = numerator / denominator # VMR estimate
## Detection
sigma_delta = sqrt(t.T @ Sigma_norm @ t) # Detection threshold
delta = numerator # Detection statistic
detected = (delta > 3 * sigma_delta) # 3-sigma detection
Optimizations:
Precompute (shared across pixels)
Precompute denominator (constant for scene)
Use diagonal → instead of
Vectorize over pixels (process batch simultaneously)
With these optimizations: ~0.1 ms/pixel on GPU, enabling real-time processing.
Summary: Model 4B in Context¶
The Optimal Operational Choice¶
Model 4B achieves the best trade-off for operational methane detection:
| Criterion | Model 4B Performance |
|---|---|
| Accuracy | Good (5-10% error for moderate plumes) |
| Speed | Excellent (100× faster than exact) |
| Calibration | None required (self-calibrating) |
| Robustness | High (immune to multiplicative errors) |
| Implementation | Simple (closed-form solution) |
| Validity range | Moderate plumes (, ~1000 ppm) |
When to Use Model 4B¶
✓ Recommended for:
Operational detection systems
Real-time processing requirements
Large-area surveys (satellites, aircraft)
Moderate plumes (100-1000 ppm typical)
Scenes with calibration uncertainty
Initial screening before detailed retrieval
✗ Not recommended for:
Strong plumes (>2000 ppm) → use Model 3B or 1B
Very weak signals (<50 ppm) → below detection threshold anyway
Flux quantification requiring <5% accuracy → use Model 1 or 1B
Elevated/stratified plumes → spatial assumptions break down
Hierarchy Position¶
Most Accurate ←──────────────────→ Fastest Model 1B Model 3B Model 4B Model 2B (Exact) (Taylor only) (Combined) (MacLaurin only) ●──────────────●──────────────●──────────────● │ │ │ │ 0% error 2-5% error 5-10% error 15-20% error 50 ms/px 10 ms/px 0.5 ms/px 0.5 ms/px Iterative Iterative Closed-form Closed-form All τ τ < 0.3 τ < 0.1 τ < 0.05
Model 4B occupies the “sweet spot” where accuracy is good enough and speed is fast enough for practical operations.
Future Directions¶
Extensions of Model 4B:
Adaptive matched filter: Estimate locally from data
Multi-gas detection: Simultaneous CH₄, CO₂, H₂O retrieval
Spatial regularization: Enforce plume smoothness across pixels
Machine learning augmentation: Train detector on Model 4B + corrections
Hybrid approach: Use Model 4B for detection, Model 1B for quantification
The combined linearized model will remain the foundation of operational methane detection for the foreseeable future.