3 - Recap Clip 3.3: Agent Architectures based on Belief States [ID:30405]
43 von 43 angezeigt

Now, we want to extend that and the upshot of everything we did last time is we need

to extend this model to not a model of the world state, but since we don't know the

world state and cannot know the world state, to what we believe about the world state.

And that really needs two things, right, we need to have a belief state, what do we believe

the world is like, and we need to extend our model to allow multiple possible worlds, because

some things we don't know, so we have to kind of plan for both contingencies, and to have

a transition model that actually updates our belief state with new information coming in.

For instance, if I'm unsure whether I'm on, I don't know, I can see outside, so, if I'm

unsure whether I'm on the bottom floor of this building or there's something beneath,

right, I need to have a plan for both contingencies, but I can also just basically go down, see

whether there are some stairs and look what's beneath here, and then I can actually restrict

the number of possible models, making my belief state more accurate.

But of course I have to plan for my actions being unreliable, and when I thought I was

going down, I was actually going up, and get false information.

And of course the environment tells me which of those things I really have to do, if I

have full observability, then my belief state can become a world state, and if I have reliable

sensors and actions, then the transition model can actually be deterministic.

So essentially we're looking at a generalization of the agents we were talking about naively

last semester.

So here we have basically, we've reviewed our kind of agents we talked about last semester,

isn't it nice that we can kind of condense a whole semester onto one slide.

So we look at what our environments are like, they're all fully observable, and what our

world states are essentially, and how we do the updates.

We do it either without any inference or with constraint propagation, or with full logic

based inference, and the planning was different from the first couple, in that we had a dynamic

environment which could actually change, which is the whole point of planning.

We're planning for the changes, so we have to have all these delete and add lists and

all of those kind of things.

So the inference is basically state or plan based search, and we have a transition model

which is essentially strips, meaning I can describe the world by a couple of assertions,

and some I have to leave out after an action, and some I have to add after an action.

Right and what we're going to look at now is essentially the generalization, which we

call probabilistic agents, where we have a belief model, which is essentially a set of

possible worlds, together with their likelihoods of being true, we're going to use probability

theory for expressing that, and we're going to come up with a inferentially efficient

representation here, which means we can actually do inference on this difficult problem, and

the inference of course is going to be probabilistic inference.

Which means we're going to learn about probability theory first.

And then we are going to extend those to what we call the decision theoretic agents, which

are a special kind of utility based agents that actually try to predict how good the

actions are and tries to optimize the positive effects of their actions.

We've essentially just set the stage, so we know what kind of the setting is in which

we are actually now going to extend our repertoire of theoretical methods.

Teil eines Kapitels:
Recaps

Zugänglich über

Offener Zugang

Dauer

00:05:38 Min

Aufnahmedatum

2021-03-30

Hochgeladen am

2021-03-31 10:18:04

Sprache

en-US

Recap: Agent Architectures based on Belief States

Main video on the topic in chapter 3 clip 3.

Einbetten
Wordpress FAU Plugin
iFrame
Teilen