gusl: (Default)
[personal profile] gusl
I am wondering if, instead of running MCMC and hoping that it has "mixed" ("achieved stationarity"), there are approaches based on computing (or approximating) the principal left eigenvector of the transition matrix.

Of course, in continuous spaces, this "matrix" has as many entries as S^2, where S is the space our parameters live in... so our "principal eigenvector" becomes the "principal eigenfunction". Functional analysts, how do you compute this?

If it helps, we might want to choose a sparse proposal (such as the one corresponding to Gibbs sampling, in which all transitions changing more than one parameter have probability density zero)

(no subject)

Date: 2010-02-01 02:06 pm (UTC)
From: [identity profile] en-ki.livejournal.com
To all appearances, solutions are one-off. My Brownian motion book cites Kac for the case of one-dimensional Brownian motion; I don't seem to have access to Kac's papers from here.

(no subject)

Date: 2010-02-01 07:33 pm (UTC)
From: [identity profile] gustavolacerda.livejournal.com
by "one-off" do you mean: improvised? non-generalizable? few and far between?

(no subject)

Date: 2010-02-01 07:40 pm (UTC)
From: [identity profile] en-ki.livejournal.com
All of the above, though "non-generalizable" is the usual meaning of one-off.

It appears that the problem more or less reduces to solving differential equations in the space of interest and, as with solving DEs, you may be able to get a general solution for a tiny but useful class of problems (like linear DEs or one-dimensional Brownian motion), but in general you have to either discover rare completely ungeneral tricks or do numerical approximations.

(no subject)

Date: 2010-02-01 07:42 pm (UTC)
From: [identity profile] gustavolacerda.livejournal.com
Sounds like a good lead, thanks. But I thought that Brownian motion means you get full diffusion.

I'm thinking I should ask a statistical physicist about this.

(no subject)

Date: 2010-02-01 09:44 pm (UTC)
From: [identity profile] en-ki.livejournal.com
Yeah, I have the book sitting on my shelf, but I haven't read it and am about 3 orders of magnitude away from being an authority.

(no subject)

Date: 2010-02-03 10:24 am (UTC)
From: [identity profile] gustavolacerda.livejournal.com
Btw, I imagine that the eigenfunction problem is what variational approaches to inference are solving.

Our goal is to minimize the distance between the estimate and the posterior, i.e.:
Let f be our estimate. We want to minimize a functional like:

D(f) = \Integral |f(x) - (tf) (x)|^2 dx

where tf is the result of applying transition function t to f.

Transition function t is defined in terms of the proposal g:

t(i,j) = min(1,P(j)/P(i)) g(i,j) , as per Metropolis-Hastings.

g(i,j) is defined as the probability of being in state j at time t+1 given that we were in state i at time t.

Minimizing a function is a typical form of a variational problem.

This confirms my suspicion. (though they use reverse KL rather than L2 distance)

February 2020

S M T W T F S
      1
2345678
9101112131415
16171819202122
23242526272829

Most Popular Tags

Style Credit

Expand Cut Tags

No cut tags