Skip to main content
Skip to article

Research Note

Inverse Problems & Regularization

Zhenyu He · Jobs Stroustrup 3 min read

Definition

Inverse Problem: given a forward model (known observation/physics operator , data , unknown , noise ), recover .

Ill-posed (Hadamard 1902): fails one of existence, uniqueness, or stability (small data perturbations yield small solution perturbations).

Discrete Inverse Problem: the discretization often has a very large condition number — small data noise amplifies into large solution perturbations.

Regularization trades strict data fit for stability via a regularizer and parameter : Common choices: Tikhonov or , sparsity (compressive sensing).

Core Arguments: choosing

too small → noise-dominated; too large → regularizer-dominated. Zhenyu compared four methods:

  1. L-Curve: log-log plot of vs ; the “corner” marks best . Geometric intuition, needs no noise estimate; fragile when no clear corner exists.
  2. L-Curve Curvature: numerical curvature automates corner detection; sensitive to noise in the differentiation.
  3. Generalized Cross Validation (GCV) [Golub, Heath, Wahba 1979]: ; closed-form approximation to leave-one-out CV. Statistically optimal for prediction MSE; may be flat; sometimes underregularizes.
  4. Morozov Discrepancy Principle: pick s.t. (noise level). Principled; requires known .

Different Perspectives

  • Hansen (2010): L-Curve preferred when noise is unknown — strongest geometric intuition
  • Golub (1979): GCV statistically optimal under Gaussian noise
  • Zhenyu’s DSCOVR observation: L-Curve ≈ Curvature; GCV tends to underregularize on real data — no silver bullet, use multiple + check physical plausibility

Core Tools

  • SVD + Tikhonov closed form: — the filter-factor view reveals which singular directions are kept vs suppressed
  • Condition number analysis:
  • Picard condition: check decays faster than down to noise — otherwise recovery is fundamentally hopeless
  • Filter factors: Tikhonov → smooth transition; TSVD → step (keep top singular components)

Open Questions

  • Nonlinear inverse problems: most real retrievals are with nonlinear — how do the four methods extend?
  • Bayesian unification: ↔ prior precision; choosing ↔ Type-II MLE/MAP — the tradeoff vs full MCMC over hyperparameters (see Gaussian Process Bayesian Inversion)?
  • Deep-learning “regularization” (dropout, weight decay, augmentation) — formal link to classical Tikhonov?

Applications

Used/planned: DSCOVR retrieval, MODIS×CALIPSO joint aerosol, nuclear winter soot profile retrieval, exoplanet spin-orbit tomography. Transferable to: medical imaging (CT, MRI), geophysical prospection, atmospheric trace gas retrieval (CH4, CO2), astronomical image reconstruction.

Sources

  • — code + comparison experiments + self-study notes
  • Undergraduate Research @ Peking University & Caltech/UCR — project background
  • Textbook (raw): Hansen Discrete Inverse Problems 2010
  • Original paper (raw): Golub, Heath, Wahba 1979 (GCV)
  • Lecture notes (raw): Morozov discrepancy principle
  • [Pending] Further .docx study notes from Zhenyu’s research-angle LLM Wiki, to be merged in later