31 - Deep Learning - Recurrent Neural Networks Part 1 [ID:16552]
50 von 103 angezeigt

Welcome everybody to a new session of deep learning. Today we want to look into sequential

learning and in particular recurrent neural networks.

Recurrent neural networks? You can write them down in five lines of pseudocode.

So far we only had simple feed-forward networks where we had essentially a fixed size input

and would then generate a classification result like cat, dog or hamster.

But if we have sequences like audio, speech, language or videos that have a temporal context,

the techniques that we've seen so far are not that very well suited.

So we're interested now into looking into methods that will be applicable to arbitrary

long input sequences.

Recurrent neural networks are exactly one method to actually do so.

So after a first review of the motivation, we'll go ahead and look into simple recurrent

neural networks. Then we'll introduce the famous long short-term memory units followed

by gated recurrent units. Then we will compare these different techniques and discuss a bit

the pros and cons. And finally we will talk about sampling strategies for RNNs. Of course

this is way too much for a single video, so we will talk about the different topics in

individual short videos.

Okay, so let's look at the motivation. Well, we had one input for one single image, but

this is not so great for sequential or time-dependent signals such as speech and music, video or

other sensor data where you could even talk about very simple sensors that measure energy

consumption.

So these snapshots with a fixed length are often not that informative. So if you look

at a single word, you probably have trouble into getting the right translation because

the context methods. And temporal context is really important and it needs to be modeled

appropriately.

So the question is now how can we integrate this context into the network? The simple

approach would be to feed the whole sequence to a big network. And this is potentially

a bad idea because we have inefficient memory usage, it's difficult to train or even impossible

to train and we would never figure out the difference between spatial and temporal dimensions.

We would just handle all the same. Actually, maybe it's not such a bad idea for rather

simple tasks as you can see in the reference down on the slide because they actually investigated

this and found quite surprising results with CNNs.

Well, one problem that you have of course is it won't be real time because you need

the entire sequence for the processing. So the approach that we're suggesting in this

and the next couple of videos is to model sequential behavior within the architecture

and that gives rise to recurrent neural networks.

So let's have a look at the simple recurrent neural networks. And the main idea is that

you introduce a hidden state HT that is carried on over the time. So this can be changed but

it is essentially connecting back to the original cell A. So A is our recurrent cell and it

has this hidden state that is somehow allowing us to encode what the current temporal information

has brought to us.

Now we have some input XT and this will then generate some output YT. And by the way, the

first models were from the 1970s and the early 1980s like Hopfield networks. Here we will

stick with the simple recurrent neural network or Elman network as introduced in reference

number five.

Now feed-forward networks only feed information forward. So with recurrent networks in contrast

we can now model loops, we can model memory and experience and we can learn sequential

relationships. So we can provide continuous predictions as the data comes in and this

enables us to process everything in real time.

Now this is again our basic recurrent neural network where we have some input X that is

multiplied with some weight. Then we have the additional input, the hidden state from

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

00:11:08 Min

Aufnahmedatum

2020-05-25

Hochgeladen am

2020-05-26 01:46:24

Sprache

en-US

Deep Learning - Recurrent Neural Networks Part 1

This video introduces the topic of recurrent neural networks and the Elman Cell.

Video References:
Lex Fridman's Channel

Further Reading:
A gentle Introduction to Deep Learning

Tags

introduction architectures artificial intelligence deep learning machine learning pattern recognition Recurrent Neural Networks
Einbetten
Wordpress FAU Plugin
iFrame
Teilen