Skip to main content
Skip to article

Research Note

Gaussian Process Bayesian Inversion

Zhenyu He · Jobs Stroustrup 3 min read

Definition

Bayesian Inversion recasts the inverse problem as posterior inference: with likelihood , prior , and the posterior as the deliverable (not a point estimate, but full uncertainty quantification).

Gaussian Process Prior: . Treat the unknown as a random function over some index space (sphere, space-time grid); the kernel encodes “how similar neighboring points should be.”

Conjugacy: with Gaussian likelihood + GP prior, the posterior is Gaussian with closed-form mean and covariance: This closed-form posterior mean is equivalent to the Tikhonov solution when the Bayesian dual of classical regularization.

Core Arguments

1. Tikhonov ≡ posterior mean under isotropic Gaussian prior. = noise precision / prior precision. Choosing = choosing prior strength.

2. GP kernels = structured priors. Isotropic Gaussian is the weakest. RBF kernel gives “spatial smoothness.” In DSCOVR retrieval:

  • Spatial: RBF on HEALPix sphere, = “how large a scale of structure do we expect on Earth”
  • Temporal: separate RBF
  • Separable to avoid

3. Hierarchical Bayes — hyperparameters as random variables. Either Type-II MLE (marginal likelihood point estimate) or Fully Bayesian: MCMC over . Zhenyu chose MCMC in .

4. Practical notes on MCMC over hyperparameters

  • Affine-invariant ensemble (emcee) for moderate dims
  • Log-space reparameterization (log10_alpha, log10_sigma) for scale-spanning params
  • Mixed integration (Rao-Blackwell): MCMC samples ; given each, is closed-form Gaussian — mixture approximation for final posterior
  • Numerical stability: Cholesky (scipy.linalg.solve(..., assume_a="pos")), not inv; slogdet, not log(det)

Different Perspectives

  • Frequentist vs Bayesian: point (fast) vs full posterior (slow, honest UQ)
  • MCMC vs VI: true posterior slowly vs approximate posterior quickly
  • GP scaling: Cholesky is the ceiling. Separable kernel is one escape; inducing points / random features / KISS-GP / GPyTorch LazyTensor are modern directions

Kawahara / Aizawa Exoplanet Spin-Mapping Lineage

Zhenyu’s DSCOVR Earth retrieval borrows directly from the exoplanet line:

  • Cowan & Agol 2008: feasibility of retrieving surface texture from exoplanet light curves
  • Kawahara & Fujii 2010 / Kawahara 2016: SOT (Spin-Orbit Tomography)
  • Aizawa 2020 ApJ 896 22 (PDF in repo): Dynamic SOT — time-varying maps
  • Zhenyu’s transfer: apply SOT to Earth viewed from L1 (DSCOVR) — a perfect testbed

Applications

Light-curve → exoplanet/Earth surface maps; multi-parameter atmospheric retrieval UQ; climate-observation data assimilation; any problem needing smooth solution + UQ.

Open Questions

  • Non-conjugate likelihoods (Poisson photon counts, Student-t heavy tails) — no closed-form posterior, need more involved MCMC
  • Prior sensitivity: how much does a wrong distort results? Systematic sensitivity study?
  • Deep-learning integration: NTK view — infinite-width NN ≡ GP; deep structured GP priors?
  • Real-time retrieval: sequential Bayesian update (Kalman-family) rather than batch MCMC each hour?

Sources

  • — code + Kawahara/Aizawa reproduction + DSCOVR experiments
  • Undergraduate Research @ Peking University & Caltech/UCR — project background (Yuk Yung / King-Fai Li)
  • Textbook (raw): Bishop PRML 2006 — Ch. 6 GP
  • Paper (raw): Aizawa 2020 ApJ 896 22 — Dynamic SOT
  • Implied papers: Kawahara & Fujii 2010, Cowan & Agol 2008
  • [Future expansion] Kawahara series papers, derivation details, Zhenyu’s own Chinese .docx study notes