Upload
malina
View
92
Download
0
Tags:
Embed Size (px)
DESCRIPTION
Introduction to particle filter. Presented by: Amir Monemian Advisor : Dr Shiry. Why should you be interested in particle filters?. great way to track the state of a dynamic system Bayesian model you can use particle filters to track your belief state. - PowerPoint PPT Presentation
Citation preview
Presented by:Amir Monemian
Advisor :Dr Shiry
great way to track the state of a dynamic system
Bayesian model you can use particle filters to track your
belief state
the main reason is that for a lot of large or high-dimensional problems, particle filters are tractable whereas Kalman filters are not.
particle filters, which let you use the full, complex model, but just find an approximate solution instead.
Bayes Filtering used to discuss the
method of using a predict/update cycle to estimate the state of a dynamical system from sensor measurements
p(xt | do…t)p(zt | xt),p(xt | xt-1, ut-1)
X is the state variable
Xt is the state variable at time t
U is the inputs to your system
z is the observations made by the sensors
d just refers to inputs andobservations together
probability of xt given all the data we’ve seen so far
perceptual model,
action model
What we are given is the inputs, the observations, the perceptual model, which is the probability that you’ll see a particular observation given that you’re in some state at time t, and the action model, which is the probability that you’ll end up in state xt at time t, assuming that you started in state xt-1 at time t-1, and input ut-1 to your system
The basic idea of particle filters is that any pdf can be represented as a set of samples (particles).
the density of your samples in one area of the state space represents the probability of that region.
This method can represent any arbitrary distribution, making it good for non-Gaussian, multi-modal pdfs.
the key idea is that you find an approximate representation of a complex model
(any arbitrary pdf) rather than an exact representation of a simplified mode (Gaussians).
how do you sample from your posterior? you had some belief from the last time
step that you know how to update with your motion model. (prior belief q(x) )
sample from q(x), and then for each sample that you made, update it using what we will call an ‘importance weight’, based on the observations made.
To start the algorithm, we need the initial belief state, p(x0). This is just our initial guess of the pdf. For robot localization, if we have no idea, we can just scatter particles all over the map.
then loop with three phases: prediction, update, and resample
calculating this equation from right to left
In the prediction step, we take each particle and add a random sample from the motion model.
the update step , weight that is equal to the probability of observing the sensor measurements from that particle’s state.
in the resample step, a new set of particles is chosen so that each particle survives in proportion to its weight.