7 - Artificial Intelligence II [ID:9077]
50 von 993 angezeigt

Last week we learned about conditional independence as a way of modeling probabilistic facts about

the world and move towards the kind of main reasoning tool, namely Bayesian networks that

allow us to actually go the step from math to modeling. So the basic idea was that we build

up these networks, networks where the nodes are random variables and the arrows are dependencies.

More importantly, non-arrows mean conditional independence. This is kind of the prototypical

kind of topology that we're interested in. We have two conditionally independent things given

their parents. So that's the idea and if that's the case then we can start the mill grinding

doing normalization and marginalization. We're actually interested in situations. We have a

certain evidence the agent has seen or perceived something and we want to compute certain probability

distributions given that evidence. So we do the normalization and marginalization trick

to get at these kind of probabilities and then these we compute with a chain rule and here is where

the Bayesian networks kicks in. Basically in these conditional probabilities we can drop

lots of stuff, namely everything except the parents of a node. That's going to be where the

good stuff happens, namely things get less complex. That's where we exploit conditional independence.

Conditional independence. Here we've talked about applications and kind of entered the math of Bayesian

networks and so you should think of Bayesian networks being a couple of things at the same time.

It's an inference procedure, something that allows us to program with probabilities and computing

these probability distributions given some evidence. It's a very well engineered substitute for working

with the full joint probability distribution which is unwieldy and gets too big and it's also a

graphical tool where you can just basically look at the model of the world and see whether it makes sense.

So we extensively looked at this example which is up here. We have a Bayesian networks. You remember

we have some causes. We have an alarm that can go off if there's a burglary that's what it's for

and there's an earthquake that's what it's not for and we kind of have this remote

supervision system of the neighbours which call if they hear the alarm and now the question is

John or Mary call you at work and now the question is do you call the police or not?

We have this very simple Bayesian networks that have the intuition that burglary and earthquake

cause the alarm and the alarm causes the calls and we have certain probabilities involved with this.

In the simplest case on the roots we have the prior probabilities on the single

inflow nodes we have a very simple probability transformation. If the alarm is true then

John calls with 90% and if it's false he still calls with 0.5% and if it's false he still calls

with 0.5% and here we have this kind of conditional probability table that tells me the probabilities

of whatever the four possibilities of what burglary and earthquake is.

It's a very succinct and simple representation of the world and

it tells me we have this conditional independence situation here that we want to exploit.

We looked at the quote unquote meaning of the Bayesian network and the idea

of the Bayesian network is that it's a very simple representation of the world.

The idea is that one of the possible meanings is that it's a representation of conditional

independencies and the kind of catchphrase here is that any node or random variable x

is conditionally independent of its non-descendants. Descendants are this,

the downward cones and the non-descendants are everything else.

It's conditionally independent of its non-descendants given its parents so x and zlj

is actually conditionally independent given its parents.

How would this scholastic independency look like?

A normal independence. You just don't have any parents.

In our example burglary and earthquake are independent given their parents which is nothing.

This is a full independence where we only have a conditional independent because we have parents there.

Another way of expressing the meaning and if this was more an inferential meaning,

what can I do with it? What does that help me? Another way of thinking of the meaning,

a more mathematical way of the meaning is that the Bayesian network is a way of representing

the full joint probability distribution of all these variables.

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

01:22:07 Min

Aufnahmedatum

2018-05-02

Hochgeladen am

2018-05-03 08:09:24

Sprache

en-US

Dieser Kurs beschäftigt sich mit den Grundlagen der Künstlichen Intelligenz (KI), insbesondere mit Techniken des Schliessens unter Unsicherheit, des maschinellen Lernens und dem Sprachverstehen. 
Der Kurs baut auf der Vorlesung Künstliche Intelligenz I vom Wintersemester auf und führt diese weiter. 

Lernziele und Kompetenzen
Fach- Lern- bzw. Methodenkompetenz

  • Wissen: Die Studierenden lernen grundlegende Repräsentationsformalismen und Algorithmen der Künstlichen Intelligenz kennen.

  • Anwenden: Die Konzepte werden an Beispielen aus der realen Welt angewandt (bungsaufgaben).

  • Analyse: Die Studierenden lernen über die Modellierung in der Maschine menschliche Intelligenzleistungen besser einzuschätzen.

Einbetten
Wordpress FAU Plugin
iFrame
Teilen