slides in pdf

31
/centre for analysis, scientific computing and applications Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Num Optimization with nonnegativity constraints Arie Verhoeven [email protected] CASA Seminar, May 30, 2007

Upload: vuongdung

Post on 31-Dec-2016

247 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Optimization with nonnegativity constraints

Arie [email protected]

CASA Seminar, May 30, 2007

Page 2: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Seminar: Inverse problems

1 Introduction Yves van Gennip February 212 Regularization strategies Miguel Patricio March 33 Regularization by Hans Groot March 21

Galerkin methods4 Inverse eigenvalue problems Marco Veneroni April 45 Image deblurring Willem Dijkstra April 186 Parameter identification Nico van der Aa May 27 Total variation regularization Mark van Kraaij May 238 Optimization with Arie Verhoeven May 30

nonnegativity constraints9 ?? Martijn Slob June 13

10 ?? Marc Noot June 20

Page 3: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 4: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 5: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Formulation of the problem

Discretized mathematical model:

f truei = ftrue(xi).

Discrete Fourier Transform:

gtrue = K · ftrue.

Measurements, e.g. from astronomical imaging, satisfy

di ∼ Poisson(gtruei ) + Normal(0, σ2).

Goal: find f for given d and K.

Page 6: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Least Squares method with Tikhonovregularization

The least squares solution minimizes the functional:

Jls(f) =12‖Kf − d‖2 +

α

2‖f‖2.

The minimizer is given by

f lsα =

[KT K + αI

]−1KT d.

The regularization parameter α is selected to minimize‖f ls

α − ftrue‖. But this approach does not work in general whenftrue is unknown.

Page 7: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

The L-curve method

Define X (α) = log ‖Kfα − d‖2 and Y (α) = log ‖fα‖2. WithTikhonov regularization X , Y are smooth functions. Then α canbe selected which maximizes the curvature function

κ(α) =X (α)Y (α)− X (α)Y (α)

(X (α)2 + Y (α)2)32

.

Note that this selected point corresponds to the "corner" of theL-curve, which plots X against Y . Although this method isnonconvergent, it can be used to improve the value of α withoutknowledge of ftrue.

Page 8: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Nonnegatively constrained minimizationFor astronomical imaging it is well-known that fi ≥ 0. Thus it ismore accurate to solve instead

minf

Jls(f) subject to f ≥ 0.

Even better is to include the stochastic information of themeasurements. Then we solve the following constrainedPoisson likelihood minimization problem

minf

Jlhd(f) subject to f ≥ 0,

where

Jlhd(f) =n∑

i=1

(gi +σ2)+n∑

i=1

((max{di , 0}+σ2) log(gi +σ2))+α

2‖f‖2.

Page 9: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Example

ftrue(x) =

750, 0.1 < x < 0.25,250, 0.3 < x < 0.32,5 · 105(x − 0.75)(0.85− x), 0.75 < x < 0.85,0 otherwise.

Page 10: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Constrained likelihood minimization

Page 11: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 12: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Optimization

Consider the inequality constrained minimization problem

minf∈Rn

J(f) subject to c(f) ≥ 0.

Active set of indices:

A(f) = {i | ci(f) = 0}.

Page 13: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

KKT conditions

Karush-Kuhn-Tucker (first order necessary) conditions forinequality constrained minimization: There exists a vector λsuch that

gradJ(f∗)−m∑

i=1

λ∗i gradci(f∗) = 0

andλ∗i ≥ 0, ci(f

∗) ≥ 0, λ∗i ci(f∗) = 0.

We define the projection of f onto C as

PC(f) = arg minv∈C

‖f − v‖.

The operator PC is well defined and continuous.

Page 14: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Nonnegativity constraints

Now we consider the problem

minf∈Rn

J(f) subject to f ≥ 0.

If J is continuously differentiable and f∗ a local minimizer, itfollows that λ∗ = gradJ(f∗). Then we obtain

∂J∂fi

(f∗) ≥ 0, f ∗i ≥ 0, f ∗i∂J∂fi

(f∗) = 0.

A point which satisfies these conditions is a critical point, but itneeds not to be a minimizer if.e.g. J is not strictly convex,

Page 15: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Nonnegativity constraints II

For nonnegativity constraints, we define the feasible setC = {f | f ≥ 0}. The projected gradient ∇CJ : C → Rn satisfies

[∇CJ]i =

{∂J∂fi

(f) if fi > 0,

min{0, ∂J∂fi

(f)} if fi = 0.

Thus f∗ is a critical point if and only if ∇CJ(f∗) = 0. For given Cwe define

P(f) = arg minv≥0

‖v − f‖.

If f∗ is a local minimizer, then

f∗ = P(f∗ − τgradJ(f∗)) for any τ > 0.

Page 16: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 17: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Gradient projection method

ν := 0;f0 := nonnegative initial guess;begin

pν := −grad J(fν);τν := arg minτ>0 J(P(fν + τpν));fν+1 := P(fν + τνpν);ν := ν + 1;

end

This generalized Steepest Descent method converges linearlyto the global minimizer if J is strictly convex, coercive, andLipschitz continuous.

Page 18: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Projected Newton method

ν := 0;f0 := nonnegative initial guess;begin

gν := grad J(fν);Identify active set Aν ;HR := reduced Hessian at fν ;s := −H−1

R gν ;τν := arg minτ>0 J(P(fν + τs));fν+1 := P(fν + τνs);ν := ν + 1;

end

The reduced Hessian equals

[HR]ij =

{δij if i ∈ A(f) or j ∈ A(f),∂2J

∂fi∂fjotherwise.

If the active set can be correctlyidentified, this algorithm will belocally quadraticallyconvergent.

Page 19: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Gradient projection-reduced Newton method

ν := 0;f0 := nonnegative initial guess;begin

Gradient Projection StagepGP := −grad J(fν);τGP := arg minτ>0 J(P(fν + τpGP));

fGPν := P(fν + τGPpGP);

Reduced Newton StageIdentify active set A(fGP

ν );

gR := reduced gradient at fGPν ;

HR := reduced Hessian at fGPν ;

s := −H−1R gR ;

τRN := arg minτ>0 J(P(fGPν + τs));

fν+1 := P(fGPν + τRNs);

ν := ν + 1;end

We need

[gR(f)]i =

{0 if i ∈ A(f),

∂J∂fi

(f) otherwise.

This algorithm combines theglobal convergence of GradientProjection with the locallyquadratic rate of ProjectedNewton. For large-scaleproblems the linear systemHRs = −gR could be solved byan iterative method like theConjugate Gradient method.

Page 20: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 21: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Richardson-Lucy iteration

Iterative methods use iteration count as regularizationparameter.Consider

J(f) =m∑

i=1

di log[Kf ]i .

Approximation of maximizer by Richardson-Lucy iteration:

f ν+1j =

f νj

kj

m∑i=1

kij

(di∑n

l=1 kil f νl

), where kj =

m∑l=1

klj .

We get a sequence of approximations to the maximizer of J(f)subject to

m∑i=1

[Kf ]i =m∑

i=1

di .

Page 22: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Modified Reduced Newton Steepest Descent

ν := 0;f0 := nonnegative initial guess;g0 := KT (Kf0 − d);γ := (g0, f0. ∗ g0);begin

pν := −fν . ∗ gν ;u := Kpν ;τbndry := min{−[fν ]i/[pν ]i | [pν ]i < 0};fν+1 := fν + τνpν ;

gν+1 := gν + τνKT u;γ := (gν+1, fν+1. ∗ gν+1);ν := ν + 1;

end

Page 23: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 24: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

1D

Page 25: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

2D

Page 26: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Outline

1 Introduction

2 Theory of constrained optimization

3 Numerical variational methods

4 Iterative nonnegative regularization methods

5 Numerical test results

6 Conclusions

Page 27: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Summary

Nonnegitivity constraints

Theory of constrained optimizationVariational methods

Gradient projection methodProjected Newton methodGradient projection-reduced Newton methodGradient projection-CG method

Iterative methodsRichardson-Lucy iterationModified Steepest Descent algorithm

Numerical test results

Page 28: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Conclusions

Optimization with nonnegativity constraints often leads tomore accurate reconstructions with e.g. less unwantedoscillations.

Optimizing Poisson likelihood is more accurate than LeastSquares.

Iterative methods are preferable if no good a priori value ofthe regularization parameter is available.

Variational regularization methods are more flexible,because they allow the use of prior information about thesolution and constraints.

Page 29: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Literature.

C.R. Vogel: Computational methods for inverse problems, SIAM,Philadelphia, 2002, pp. 151-171.

J. Nocedal and S.J. Wright: Numerical optimization,Springer-Verlag, New York, 1999.

S.G. Nash and A. Sofer: Linear and nonlinear programming,McGraw-Hill, New York, 1996.

H.W. Engl, M. Hanke and A. Neubauer: Regularization of inverseproblems, Kluwer Academic Publishers, Dordrecht, 1996.

W.H. Richardson: Bayesian-based iterative methods for imagerestoration, Journal of the Optical Society of America, 62 (1972),pp. 55-59.

B. Lucy: An iterative method for the rectification of observeddistributions, Astronomical Journal, 79 (1974), pp. 745-754.

Page 30: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Literature.

C.R. Vogel: Computational methods for inverse problems, SIAM,Philadelphia, 2002, pp. 151-171.

J. Nocedal and S.J. Wright: Numerical optimization,Springer-Verlag, New York, 1999.

S.G. Nash and A. Sofer: Linear and nonlinear programming,McGraw-Hill, New York, 1996.

H.W. Engl, M. Hanke and A. Neubauer: Regularization of inverseproblems, Kluwer Academic Publishers, Dordrecht, 1996.

W.H. Richardson: Bayesian-based iterative methods for imagerestoration, Journal of the Optical Society of America, 62 (1972),pp. 55-59.

B. Lucy: An iterative method for the rectification of observeddistributions, Astronomical Journal, 79 (1974), pp. 745-754.

Page 31: slides in PDF

/centre for analysis, scientific computing and applications

Introduction Theory of constrained optimization Numerical variational methods Iterative nonnegative regularization methods Numerical test results Conclusions

Questions?