2 - Mathematics - Deep Learning [ID:12477]
50 von 864 angezeigt

Well, actually, most things I will tell are already 30 years old, I guess.

So it's not so much change.

The problem is every time I talk about these things is I feel really old because I have

survived this first wave of neural networks in the 1990s.

So I still have a few papers on that even from that time in the late 90s.

And then for about 10 years you were in a scientific conference not even allowed to

mention the word neural network.

And in the last years it somehow came back.

And still, again, the second wave is as hysterical as the first one was.

So I think one has to be a bit careful.

And when it's, particularly I think more mathematicians to do something and to get some understanding

and translate it into mathematics.

And this is, well, the idea kind of for this talk.

So for the non-Germans, Langenscheid is always the standard book for translating to other

languages.

So either to other languages or to other complicated things.

So what I want to do is kind of this translation between math and deep learning.

And what scares a lot of mathematicians are things like this.

When people talk about deep learning, they actually do some mathematics.

But what you see is always strange pictures like this.

So a lot of nodes, arrows, or even in fancy colors.

And you actually don't really know what is going on.

So what I try to do is to take the basic things in deep learning or neural networks and relate

them to things that we do for many, many years so that it becomes a bit more apparent what

is going on.

And at some points there are some special things that I will mention.

I have this anonymous quote from a friend of mine, which I think scares.

It again goes on with this caring of mathematicians because when you hear all these nice stories

in deep learning and other areas, it cannot apparently relate it to things that you're

used to.

But actually, in many cases, these are just different names for things that you learn

in numerical analysis or in other classes, actually.

So I hope to convince you there is a lot of relation.

There are things that are actually not so complicated when you go to that.

So first, let me start on explaining a bit what are these crazy pictures there.

So from a mathematical point of view, what these neural network things do, and actually

they do since 50 years or so when first mathematical models of the human brain came out.

They tried to mimic what is going on in a network of neurons.

That's obviously also where the name comes from.

And roughly speaking, what you think about this thing as a neuron and the other things

as the axons going from one to the other.

And what is going on is you propagate signals from one neuron to others.

You kind of add them here.

And if the sum of the incoming signals is strong enough, the neuron will fire and will

propagate a signal to the other neurons it's connected to.

And this is what's going on.

So mathematically, this means you kind of have a system in different layers.

So this is the first index I will use.

You have some input.

Call it x.

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

01:35:25 Min

Aufnahmedatum

2019-12-05

Hochgeladen am

2019-12-05 23:19:02

Sprache

de-DE

Tags

minimization framework approximation distance deep basic weights models layer problem propagation learning training data network parameters regularization sampling solution theory generalization risk theoretical bregman stability estimating empirical estimate optimality
Einbetten
Wordpress FAU Plugin
iFrame
Teilen