10 - Musteranalyse/Pattern Analysis (früher Mustererkennung 2) (PA) [ID:385]
50 von 603 angezeigt

So we talk about optimization today and if you have the feeling during the lecture that

this is not pattern analysis you have the right feeling.

So we do some math background today that I consider to be important to be a good expert

in pattern analysis.

So basically we do here a refresher course on optimization and on things that you should

know or that I originally expected you to know but I know that nobody knows that so

I'm summarizing that.

Yesterday I started out with a motivation and I told you there is no lecture on pattern

analysis or pattern recognition without the sectional optimization techniques.

And in pattern recognition one or in the winter semester lecture I usually say well we have

a continuous objective function we compute the gradient and the zero crossings and then

we have our optimization scheme.

That's basically the rule and in many cases we will find closed form solutions for the

zero crossings of the gradient so we get nice iteration schemes or a closed form solution

for optimizing.

But in many situations, practical situations we will not be able to find a closed form

solution and the question is what do we do that and so we need some numerical methods

for optimization and the first question we are considering is the constraint, unconstrained

optimization.

So we just have a function that accepts as input a d-dimensional feature vector or a

dimensional variable and returns a real value and returns a real value.

And what we are looking for is we look for the minimum and I assume that the function

is convex that means the epigraph is a convex set.

So we have functions like this one here.

So this is a typical convex function and here we look for the minimum and the optimal point

is denoted by x star.

Whenever I use a variable with a star I mean the optimal point, right?

I mean the optimal value and what we have to solve is minimize the function f with respect

to x and tell me at which point x the function achieves its minimum.

And in this case a necessary and sufficient condition for the minimum is that the gradient

is 0.

It is necessary and sufficient because we have the constraint that we are considering

convex functions at the beginning, okay?

Usually if you have a function like this here you might also reduce the problem to a convex

optimization problem if you are able to restrict the optimization for instance to this interval.

So let's first consider convex functions and we look for a unique minimum, a unique optimal

point of these functions.

So a necessary but sufficient condition is that the gradient is 0.

An example for instance if you have a function like that we have ax minus b and we want to

minimize the Euclidean length of the residual.

This is a function that depends on x1 to let's say xm vectors so it has m times d dimensions

and you want to compute here the minimum of that.

Usually we take the square and this is obviously a convex function that we have to minimize.

We compute the derivative and the zero crossing and basically we find out that we have to

compute the single inverse of a.

So we get some normal equations to solve this least square estimate.

So that's something that we apply quite often and I'm not talking about these things anymore.

Basically I'm saying okay we have to compute the gradient, the zero crossings and then

I do all the computations.

Usually if we cannot find a closed form solution as was the case with the residual minimization

Zugänglich über

Offener Zugang

Dauer

00:45:51 Min

Aufnahmedatum

2009-05-26

Hochgeladen am

2017-07-05 12:49:37

Sprache

en-US

Tags

Analyse PA Optimization Descent methods Backtracking Line Search Steepest Methods
Einbetten
Wordpress FAU Plugin
iFrame
Teilen