61 - Recap Clip 10.4: Relevance-Based Learning [ID:30464]
50 von 123 angezeigt

The next thing really works with this second example with the Brazilian, where in learning

you really want to bias yourself to certain, in a way, learnable predicates, which you

want to prefer, A, because they generalize better, and B, because it restricts your hypothesis

space and actually lets you learn with far fewer examples.

That's what we're doing now.

And of course that can be combined with other learning techniques, as we'll see in the end,

or as we did see yesterday in the end by combining these relevance judgments with decision tree

learning.

And the idea here is that we often have pairs of predicates that determine each other and

as such give a very strong relevance statement.

In essence here we have that the nationality fully determines the language, and you can

write that down in logic, which makes for a relatively complicated formula.

What you're really interested in is if we have this determination, nationality over

language, then we only need to and should learn and only need to concentrate on nationality.

So in a way, if I have that determination, then I have a functional dependency on the

two, and I don't have any new information in language.

So we shouldn't actually learn that.

Now the question of course is how do we find out about determinations?

And the answer of course is we're going to learn them.

If we have them, they actually give us great speed ups in learning because they cut down

on the hypothesis tree.

The question is how do you learn them?

And the answer is by looking at the data.

If I have a bunch of data, then what we're interested in is actually the minimal, by

some version of minimal determinations, and this here has actually one, which is the material

and the temperature together determine the conductance.

What's happening here is we're actually using predicates, and the and here, which is the

customary thing to write, you should really read as an intersection.

And is something that acts on true and false like things.

And material and temperature aren't true and false like things.

They are actually predicates or functions or whatever.

So we have to have, so we're using, I just wanted to make you aware that we're using

a slightly interesting operator here.

And it's essentially intersection.

You can figure out exactly what this is.

And once you realize this, a relatively simple generate and test algorithm works.

You look at, and that's what we have here.

We have material and temperature.

We can cut any subset of predicates into a new compound predicate.

And so we essentially have to learn or take into account or under consideration actually

any combination of the other predicates, and that gives us an exponential algorithm.

But it's very simple.

You start with single, you start with single predicate determinations.

Then you go for two subset determinations, and you work your way up all the way.

So you look at those, and then you just basically see whether any determination is, proposed

determination is consistent with the data you have.

And for that, you essentially check things that look like the nationality things.

If two things have the same value in the antecedent, and you have a third one here, you can check

whether the values are equal as well.

And you can implement that with, as I've done it here, with a very simple hash table where

Teil eines Kapitels:
Recaps

Zugänglich über

Offener Zugang

Dauer

00:13:33 Min

Aufnahmedatum

2021-03-30

Hochgeladen am

2021-03-31 11:36:34

Sprache

en-US

Recap: Relevance-Based Learning

Main video on the topic in chapter 10 clip 4.

Einbetten
Wordpress FAU Plugin
iFrame
Teilen