paper a day : power laws for white/gray matter ratios

I swear I’ll post more often starting soon — I just need to get back into the swing of things. In the meantime, here’s a short but fun paper.

A universal scaling law between gray matter and white matter of cerebral cortex
K. Zhang and T. Sejnowski
PNAS v.97 no. 10 (May 9, 2000)

This paper looks at the brain structure of mammals, and in particular the volumes of gray matter (cell bodies, dendrites, local connections) and white mattern (longer-range inter-area fibers). A plot of white matter vs. gray matter volumes showing different mammals, from a pygmy shrew to an elephant, show a really close linear fit on a log-log scale, with the best line having a slope of log(W)/log(G) = 1.23. This paper suggests that the exponent can be explained mathematically using two axioms. The first is that a piece of cortical area sends and receives ths same cross-sectional area of long-range fibers. The second more important axiom is that the geometry of the cortex is designed to minimize the average length of the long-distance fibers.

By using these heuristics, they argue that an exponent of 4/3 is “optimal” with respect to the second criterion. The difference of 0.10 can be explained by the fact that cortical thickness increases with the size of the animal, so they regressed cortical thickness vs. log(G) to get a thickness scaling of 0.10. It’s a pretty cute analysis, I thought, although it can’t really claim that minimum wiring is a principle in the brain so much as the way brains are is consistent with minimal wiring. Of course, I don’t even know how you would go about trying to prove the former statement — maybe this is why I feel more at home in mathematical engineering than I do in science…

paper a day (month?) : isomap

A Global Geometric Framework for Nonlinear Dimensionality Reduction
J. B. Tenenbaum, V. de Silva, J. C. Langford
Science, v290, p2319 – 2323, 22 December 2000.

I have to present this paper for a computational neuroscience reading group that I can (finally) attend. The basic problem is something called manifold learning. Imagine you have a very large data set in a huge number of dimensions — for example, 64 x 64 pixel images of faces, which live in a 4096-dimensional space. Furthermore, suppose that all of the pictures are of the same person, with only two parameters changing — the angle of rotation of the face, and the illumination. The data has only two degrees of freedom, so you would think it would live on a 2-dimensional subspace.

Unfortunately, the data you have doesn’t occupy a linear subspace in the observed variables. Instead, they live on a manifold, which is like a surface in your high dimensional space that may be strangely curved or twisted, and so may be very poorly approximated by a linear subspace. However, the manifold does have its own coordinate system, and you can
calculate distances between points on the manifold. The shortest path between two points is called a geodesic.

Another way to visualize this is a ribbon that is curled into a spiral. A point A on the ribbon might look close to a point B that is on an outer ring, but if you unwrapped the ribbon they would be far apart. So one thing you might think to do is somehow figure out what the real distance (on the surface of the ribbon) between A and B is. You might be able to do that by somehow hopping from data-point to data-point in short hops that would hopefully follow the contour of the ribbon.

That is exactly what the Isomap algorithm, described in this paper, does to perform its dimensionality reduction. Given a set of points {xk : k = 1, 2,…, K} in n-dimensional space X, they first make a graph G with vertex set {xk} by putting an edge between two points if the distance between them in X is less than some threshold. The edge has a weight given by the distance. Then they find a K x K matrix D whose (i,j)-th entry is the minimum-weight (distance) path in the graph G between xi and xj. Assuming the threshold is not too large and there are a lot of data points, these lengths should closely approximate the true (geodesic) distance on the manifold.

Armed with this matrix of distances between points, we can try to embed the manifold into a d-dimensional Euclinean space without distorting the distances too much. This is an easy problem to visualize — just imagine taking a globe, fixing a few cities, and trying to make a flat map in which the distances between those cities are preserved. There is an algorithm called multidimensional scaling (MDS) that can do this. You can trade off the embedding distortion with the number of dimensions.

This paper comes with its own homepage, which has some data sets and MATLAB code. If only all practitioners were so generous — too often the algorithm implementation is kept under wraps, which makes me wonder if there are some dirty secrets hiding behind the pretty plots.

One thing about reading this paper that annoyed me is that all of the technical details (which I care about) are hidden in tiny-print footnotes. Furthermore, all the citations do not include the paper titles, so you can’t tell cited papers are actually about. I know that page space is precious, but it’s just plain stupid. Shame on you, Science. I expected better.

As as amusing postscript, the commentary on this paper and the locally linear embedding paper (Roweis and Saul) written by Seung and Lee has pictures of Bush and Gore in the print edition but due to copyright issues the online version had to be changed.

medical exploitation of India

Via Krish, a story in Wired about how India is now the big site for clinical trials and drug development. Costs there are low, and as the editor of the American Journal of Bioethics noted:

Individuals who participate in Indian clinical trials usually won’t be educated. Offering $100 may be undue enticement; they may not even realize that they are being coerced.

I heard a radio program on this a few months back and tried to get my mother riled up about it, but it’s really just another strand in the rich and varied tapestry of India’s exploitation by the West/North/what-have-you.

As with most issues surrounding technology development, it boils down to an issue of pragmatics versus ethics. Pharmaceutical companies in Europe and Asia can’t find people willing to do clinical trials of their drugs in the US, even with some generous incentives. After all, who wants a placebo? On the other hand, you can get lots of volunteers for just $100 a pop in India plus paying the doctor to administer the trial, and the FDA will approve your trial. You get your drug approved, patent it, and prevent anyone in India from actually being able to afford it.

It’s not a problem specific to India either — patients in Russia are exploited in similar ways. When access to quality healthcare is limited, desperation is the primary motivating factor. Is it ethical to give a placebo in these situations? Should there be restrictions on how these studies are marketed to the public? Bioethics is going out the window in our rush for progress and refusal to shoulder the risks ourselves.

dasher

I went to a talk by David MacKay today on distributed phase codes for associative memories. One of his demos used a program called Dasher, which is a text entry system with a novel interface that could be used by disabled people. It’s baffling at first, but I imagine it becomes quite intuitive after a while. There are some demos on the website — it’s definitely worth checking out.

As far as the talk went, I have to admit I was a little lost, since everything I know about Hebbian learning and associative memories could fit on a 4×6 index card, probably. The main point of the talk was that utilizing inter-neuron spike times (or phases) and coincidence detectors that look for spatiotemporal spike pattens (with given delays) can produce structures that learn several patterns within the same neurons and can recall multiple patterns simultaneously. It’s an interesting idea, but the presentation was math-poor, so I ended up with very little idea about the “pattern capacity” of these memories, the effects of noise (and how it was modeled). These prosaic engineering questions weren’t really the focus of the talk, however, but maybe I’ll do the back-of-the-envelope calculations later.

evolution of horses

Over at Pharyngula, there is a summary of a paper on the phylogeny of American horse species. The amusing thing to me about it is the chart which has a scatter plot of metatarsal dimensions and an oval with the label New World stilt-legged and Asian asses. I feel inspired to write something with that phrase it in now…

grammatology and math?

I got this seminar announcement:

We will introduce a family of partition-valued Markov processes called exchangeable coalescent processes, and we will discuss four applications. We will explain how these processes describe ancestral processes in a discrete population model, how they describe the genealogy of continuous-state branching processes, how they can be used to model the effect of beneficial mutations on a population, and how one example called the Bolthausen- Sznitman coalescent is related to Derrida’s Generalized Random Energy Models.

Now, I wonder how many people who do probability know Derrida the critical theorist also know Derrida the statistical physicist. And vice versa, of course. Perhaps someone (Sokal?) should try applying generalized random energy models to texts.

get in my belly!

Say recently discovered dinosaur-eating mammals:

At 1 metre long, R. giganticus was big enough to hunt small dinosaurs, and a newly discovered fossil of its smaller cousin, R. robustus, died with its belly full of young dinosaur.

I guess those dino-burgers from the Flinstones are more plausible…

In other news, go Cobb County for sticking it to the creationists.