![[personal profile]](https://www.dreamwidth.org/img/silk/identity/user.png)
Lately, I've been faced with problems of the following type:
* there are many ways to represent an object
* a canonical representation is either hard to define, or unnatural to work with
(1) optimizing over the space of k-dim linear subspaces (a.k.a. k-flats). They can be represented as bases (more canonically, orthogonal bases), but that's not enough to identify them uniquely. We can define k-flats to be points in a Grassmanian, which is the appropriate quotient space (in which bases of the same space form an equivalence class). But optimizing here involves doing calculus on surfaces, which requires knowledge of differential geometry... or easier, one could use Lagrange multipliers, but I don't have enough constraints (I don't know how to constrain the rotation of the orthogonal basis).
(2) suppose you have an unknown probability distribution over clusterings of nodes. You can easily assign a label to each node such that two nodes are in the same cluster if they have the same label. Of course, these labels are just names, and permuting them results in an equivalent structure, which necessarily has the same probability. Now, suppose you get several samples from this distribution (say, by MCMC), and you want to estimate the distribution. Now, for each sample, you could simply add the other n!-1 relabelings ("mirror images"), but can you do better?
This is known as the "label-switching problem", and in some cases it can addressed by "artificial identifiability constraints" (e.g. in the form of: WOLOG, node 1 is always labeled with "a"), but this too can become unmanageable and unnatural.
Tangentially, is there a nice way to represent graphs modulo isomorphism? I'm aware of spectral techniques (e.g. graph Laplacian), but these are usually lossy, i.e. querying whether two graphs are the "same" (i.e. isomorphic) will result in some false positives.
* there are many ways to represent an object
* a canonical representation is either hard to define, or unnatural to work with
(1) optimizing over the space of k-dim linear subspaces (a.k.a. k-flats). They can be represented as bases (more canonically, orthogonal bases), but that's not enough to identify them uniquely. We can define k-flats to be points in a Grassmanian, which is the appropriate quotient space (in which bases of the same space form an equivalence class). But optimizing here involves doing calculus on surfaces, which requires knowledge of differential geometry... or easier, one could use Lagrange multipliers, but I don't have enough constraints (I don't know how to constrain the rotation of the orthogonal basis).
(2) suppose you have an unknown probability distribution over clusterings of nodes. You can easily assign a label to each node such that two nodes are in the same cluster if they have the same label. Of course, these labels are just names, and permuting them results in an equivalent structure, which necessarily has the same probability. Now, suppose you get several samples from this distribution (say, by MCMC), and you want to estimate the distribution. Now, for each sample, you could simply add the other n!-1 relabelings ("mirror images"), but can you do better?
This is known as the "label-switching problem", and in some cases it can addressed by "artificial identifiability constraints" (e.g. in the form of: WOLOG, node 1 is always labeled with "a"), but this too can become unmanageable and unnatural.
Tangentially, is there a nice way to represent graphs modulo isomorphism? I'm aware of spectral techniques (e.g. graph Laplacian), but these are usually lossy, i.e. querying whether two graphs are the "same" (i.e. isomorphic) will result in some false positives.
(no subject)
Date: 2009-10-14 04:23 am (UTC)(no subject)
Date: 2009-10-14 04:32 am (UTC)(no subject)
Date: 2009-10-14 04:33 am (UTC)(no subject)
Date: 2009-10-14 04:40 am (UTC)See http://en.wikipedia.org/wiki/Smoothed_analysis
(no subject)
Date: 2009-10-14 05:05 am (UTC)(no subject)
Date: 2009-10-14 05:00 am (UTC)(no subject)
Date: 2009-10-14 05:05 am (UTC)(no subject)
Date: 2009-10-14 05:08 am (UTC)*: it takes a lot for me to get excited about widgets anymore. Maybe if the still-living Michael Jackson had presented the result I might have been interested. Maybe.
Graphs
Date: 2009-10-14 08:19 am (UTC)And the graph isomorphism problem currently has no known polynomial-time solution.
Re: Graphs
Date: 2009-10-14 08:20 am (UTC)Re: Graphs
Date: 2009-10-14 08:29 am (UTC)btw, who is this?
Re: Graphs
Date: 2009-10-14 08:32 am (UTC)Re: Graphs
Date: 2009-10-14 08:33 am (UTC)Re: Graphs
Date: 2009-10-14 08:40 am (UTC)Re: Graphs
Date: 2009-10-14 08:50 am (UTC)Are you Chris H?
Re: Graphs
Date: 2009-10-14 08:53 am (UTC)Re: Graphs
Date: 2009-10-14 08:56 am (UTC)Re: Graphs
Date: 2009-10-14 08:59 am (UTC)Re: Graphs
Date: 2009-10-15 08:58 pm (UTC)(no subject)
Date: 2009-10-14 11:13 am (UTC)(no subject)
Date: 2009-10-14 02:49 pm (UTC)(no subject)
Date: 2009-10-14 02:59 pm (UTC)(no subject)
Date: 2009-10-15 09:41 pm (UTC)(no subject)
Date: 2009-10-15 08:01 am (UTC)my point is, exploiting the structure of your problem may lead to other more useful characterizations of the solution-set which wouldn't require you to actually try to find a representation for the set of k-flats... there are many different methods to compute eigenvalues...
a propósito, sou eu, Ives...
(no subject)
Date: 2009-10-15 05:59 pm (UTC)Both of these are actual problems motivated by research ideas. #1 is like a regression problem: given points on a unit hypersphere centered at the origin, which k-dim "great circle" minimizes the residual sum of squares? The great circles correspond to the set of k-flats (aka "k-planes") through the origin (the circle is just the intersection of the flat and the sphere).
I don't see how eigenvalues might help here. Do you? I don't have a square matrix anywhere.
(no subject)
Date: 2009-10-15 08:55 pm (UTC)Ives
(no subject)
Date: 2009-10-15 08:57 pm (UTC)yup. You could use either geodesic distance (ideal) or orthogonal distance. It probably doesn't matter which.
(no subject)
Date: 2009-10-16 01:47 am (UTC)just to fix notation, the solution (k-flat) corresponds to the vector space spanned by the eigenvectors associated to the k largest eigenvalues of the matrix \sum_i{x_i x_i^T}, x_i being the points on that sphere...
if you really want the proof, I guess I could take some time later to write it down for you or just meet me at UBC some day...
nice exercise, tough... I guess the geodesic distance would be way more complicated...
cheers,
Ives
(no subject)
Date: 2009-10-16 02:03 am (UTC)(no subject)
Date: 2009-10-16 02:15 am (UTC)Ives
(no subject)
Date: 2009-10-16 02:20 am (UTC)(no subject)
Date: 2009-10-16 02:32 am (UTC)maybe we are adopting the same name for different things... let me make it clear...
for me, the sphere is the set of x such that ||x||=1, and the closed ball is the set of x such that ||x|| <= 1
as a subset of the euclidean vector space (inheriting the usual addition and scalar multiplication), the sphere is not convex, but the closed ball is...
are you ok with that?
cheers,
Ives
(no subject)
Date: 2009-10-16 04:08 am (UTC)(no subject)
Date: 2009-10-16 04:07 am (UTC)I think the former. When you take all the eigenvectors, you end up with the original hypersphere.
I now understand why PCA is the solution here.