9 - Machine Learning for Physicists [ID:11647]
50 von 780 angezeigt

Okay, hello, good evening.

We last time started to, we finished actually the reinforcement learning and then we will

start to discuss the connections to physics and to science.

So let me just give you an introduction how you might draw the connection between neural

networks, artificial neural networks and physics.

So as you can see here, at least initially when you think of an artificial neuron, you

think of it maybe as something that can be off or on, especially if you have the sigmoid,

a nonlinear activation function.

And likewise a real neuron could be silent or it could produce electrical impulses.

And so you can easily make a connection between this and a bit, which can be zero or one.

And if you want to look for a bit in physics, you would take a spin, a spin one half that

can be up or down.

And so neural networks have a strong connection to the physics of spin systems.

And indeed if you let them have stochastic transitions and you write down cost functions

in terms of energy functionals, then you can make this connection mathematically precise.

And there have been many different ways of making this connection.

One particular I mentioned here on the slide was started by Mr. Hopfield in the beginning

of the 80s.

But this is not what we will discuss.

We will not discuss so-called Hopfield models.

Instead what we will discuss is called a Boltzmann machine.

And as I go along, I will explain to you what a Boltzmann machine really is.

But first I want to explain to you what is the goal of a Boltzmann machine.

And that's a rather general goal in machine learning, namely to model probability distributions.

Now, of course, if I give you a probability distribution like a Gaussian distribution,

you could try to say construct a neural network that outputs in its output neurons the shape

of this distribution, namely a Gaussian for example.

But that's not what we are talking about.

We're talking about a situation where you don't want to represent the distribution function,

but you want to sample, you want to produce samples according to a distribution function.

And you don't want to produce samples according to a distribution function that is known to

you like analytically.

But maybe you want to produce samples according to a distribution function that you only have

observed empirically.

So what could this mean?

In practice, it could mean that I give you many different pictures and you want to be

able eventually by having looked at all these pictures to produce pictures that are sort

of similar to the pictures that you have seen before.

So that will be the goal of what we say.

So in one sentence, to use a neural network of some type to generate previously unseen

examples, but according to a probability distribution that you have sampled before in terms of training

data.

We already mentioned one version of this when we had a look at recurrent neural networks,

LSTM networks.

And we had this case where you would train it to predict the probabilities of the next

character in a large text.

And then you can use these probabilities to actually generate the next character in the

text and in this way you can generate whole text.

So in a sense, then you are generating a text according to a probability distribution, namely

of characters, that you have trained before.

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

01:27:58 Min

Aufnahmedatum

2019-06-19

Hochgeladen am

2019-06-20 02:39:02

Sprache

en-US

This is a course introducing modern techniques of machine learning, especially deep neural networks, to an audience of physicists. Neural networks can be trained to perform diverse challenging tasks, including image recognition and natural language processing, just by training them on many examples. Neural networks have recently achieved spectacular successes, with their performance often surpassing humans. They are now also being considered more and more for applications in physics, ranging from predictions of material properties to analyzing phase transitions. We will cover the basics of neural networks, convolutional networks, autoencoders, restricted Boltzmann machines, and recurrent neural networks, as well as the recently emerging applications in physics.

Tags

transition units system machine training distribution transitions
Einbetten
Wordpress FAU Plugin
iFrame
Teilen