Research Note
Inverse Problems & Regularization
Definition
Inverse Problem: given a forward model (known observation/physics operator , data , unknown , noise ), recover .
Ill-posed (Hadamard 1902): fails one of existence, uniqueness, or stability (small data perturbations yield small solution perturbations).
Discrete Inverse Problem: the discretization often has a very large condition number — small data noise amplifies into large solution perturbations.
Regularization trades strict data fit for stability via a regularizer and parameter : Common choices: Tikhonov or , sparsity (compressive sensing).
Core Arguments: choosing
too small → noise-dominated; too large → regularizer-dominated. Zhenyu compared four methods:
- L-Curve: log-log plot of vs ; the “corner” marks best . Geometric intuition, needs no noise estimate; fragile when no clear corner exists.
- L-Curve Curvature: numerical curvature automates corner detection; sensitive to noise in the differentiation.
- Generalized Cross Validation (GCV) [Golub, Heath, Wahba 1979]: ; closed-form approximation to leave-one-out CV. Statistically optimal for prediction MSE; may be flat; sometimes underregularizes.
- Morozov Discrepancy Principle: pick s.t. (noise level). Principled; requires known .
Different Perspectives
- Hansen (2010): L-Curve preferred when noise is unknown — strongest geometric intuition
- Golub (1979): GCV statistically optimal under Gaussian noise
- Zhenyu’s DSCOVR observation: L-Curve ≈ Curvature; GCV tends to underregularize on real data — no silver bullet, use multiple + check physical plausibility
Core Tools
- SVD + Tikhonov closed form: — the filter-factor view reveals which singular directions are kept vs suppressed
- Condition number analysis:
- Picard condition: check decays faster than down to noise — otherwise recovery is fundamentally hopeless
- Filter factors: Tikhonov → smooth transition; TSVD → step (keep top singular components)
Open Questions
- Nonlinear inverse problems: most real retrievals are with nonlinear — how do the four methods extend?
- Bayesian unification: ↔ prior precision; choosing ↔ Type-II MLE/MAP — the tradeoff vs full MCMC over hyperparameters (see Gaussian Process Bayesian Inversion)?
- Deep-learning “regularization” (dropout, weight decay, augmentation) — formal link to classical Tikhonov?
Applications
Used/planned: DSCOVR retrieval, MODIS×CALIPSO joint aerosol, nuclear winter soot profile retrieval, exoplanet spin-orbit tomography. Transferable to: medical imaging (CT, MRI), geophysical prospection, atmospheric trace gas retrieval (CH4, CO2), astronomical image reconstruction.
Sources
- — code + comparison experiments + self-study notes
- Undergraduate Research @ Peking University & Caltech/UCR — project background
- Textbook (raw): Hansen Discrete Inverse Problems 2010
- Original paper (raw): Golub, Heath, Wahba 1979 (GCV)
- Lecture notes (raw): Morozov discrepancy principle
- [Pending] Further
.docxstudy notes from Zhenyu’s research-angle LLM Wiki, to be merged in later
Related Pages
- Gaussian Process Bayesian Inversion — Bayesian dual of regularization
- MCMC & Bayesian Inversion — skill page
- Satellite Remote Sensing & Data Processing — application vehicle