9 - Musteranalyse/Pattern Analysis (früher Mustererkennung 2) (PA) [ID:384]
50 von 1284 angezeigt

Welcome everybody to the Monday morning session. As usually Monday morning we will briefly sketch

the big picture such that you do not get lost with a storyline and then we will dig into one

learning algorithm that is round about 40 years old but a fundamental principle in pattern analysis

machine learning the Rosenblatt's perceptron and it's quite technical to a certain extent so be

prepared that we will see a proof that the algorithm converges brief sketch of the proof

because that's important for the general understanding. So what is pattern analysis

the summer semester about? This is all about modeling or estimating the a posteriori probability

for the class Y given a feature vector X right and what we have considered was first logistic

regression as a method to directly model the a posteriori probability using the so-called

sigmoid function just for you to remember what is the shape of the sigmoid function it looks like

a step function. That was a direct representation of the a posteriori probability and we also have

seen that we can right away read out from the sigmoid function the decision boundary. Then we

looked into the modeling of the a posteriori probability using the following factorization

P of Y times P of X given Y so we have the prior and here we have the class conditional

density function and we have modeled the class conditional density function in different ways

for instance we have seen that we can use a Gaussian for that and implement a Gaussian

classifier based on this assumption and while looking at the Gaussian classifier we found out

that we get linear decision boundaries we also talked a little bit about PCA data normalization

how to get spherical data and all that stuff so basic algorithms that are partially known

from pattern recognition in winter semester at least to a few of you. So very important section

where we combine PCA LDA we also have reconsidered basic concepts from linear algebra like the SVD

type of normalization and last week we had a very technical lecture on norms and norm approximations

and it was important to give a brief overview of different norms vector norms matrix norms

because as soon as we set up minimization problems to estimate parameters we have to deal with a

similarity measure and the similarity measure can be based on different norms we have seen the L0

type norm that is more or less the the hemming distance you know from from information theory

we have seen the L1 norm that's the sum of the absolute values we have seen the standard L2 norm

everybody who is not knowing anything about norms it's usually using the L2 norm to set up

his optimization tasks we have seen the L infinity norm we also have seen the quadratic LP norm so we

have seen a whole bunch of norms and we have looked at unit balls and the related problems

we also looked into constrained optimization just remember the rich regression that we have

considered and the rich regression and the lasso tip shirani's lasso we have considered where we

use the L1 norm combined with the L2 norm that bounds the length of the vector so there are

various degrees of freedom and I also pointed out that in modern pattern recognition and modern

pattern analysis research the usage of the norm is a hot topic and many different algorithms are

or many new algorithms are born out of the idea replace the Euclidean norm by any other norm and

solve the corresponding optimization problem and the fact that these norms are more and more

successful in our community is related to the fact that optimization theory has done a huge

progress in the recent ten years or so so convex optimization convex constraint optimization problems

can be solved today very easily using the interior point method some of you might have heard of if

not I will give a brief introduction in one of the following lectures so that's the big picture and

basically we have considered linear decision boundary for two class classification problems

so we are still more in well more academic type of questions in in pattern recognition because in

many applications you have far more than two classes but to understand the basic concepts I

think the two class problem is sufficient and we also have seen that for a given set of training

samples you will find an infinite number of linear decision boundaries if they are linearly

separable and we have seen different ways to compute these linear decision boundaries and

today we will look into Rosenblatt's perceptron this is a very old neural network type of approach

that computes linear decision boundaries but it's a very smart way to compute that and I think it's

still worth presenting this method in the lecture here so I have a new presentation software let's

Zugänglich über

Offener Zugang

Dauer

00:00:00 Min

Aufnahmedatum

2009-05-25

Hochgeladen am

2025-09-30 08:52:01

Sprache

en-US

Tags

Analyse PA Motivation Rosenblatts Perceptron Optimization
Einbetten
Wordpress FAU Plugin
iFrame
Teilen