Archive for the The Universe and Stuff Category

Arrows and Demons

Posted in The Universe and Stuff with tags , , , , , on April 12, 2009 by telescoper

My recent post about randomness and non-randomness spawned a lot of comments over on cosmic variance about the nature of entropy. I thought I’d add a bit about that topic here, mainly because I don’t really agree with most of what is written in textbooks on this subject.

The connection between thermodynamics (which deals with macroscopic quantities) and statistical mechanics (which explains these in terms of microscopic behaviour) is a fascinating but troublesome area.  James Clerk Maxwell (right) did much to establish the microscopic meaning of the first law of thermodynamics he never tried develop the second law from the same standpoint. Those that did were faced with a conundrum.  

 

The behaviour of a system of interacting particles, such as the particles of a gas, can be expressed in terms of a Hamiltonian H which is constructed from the positions and momenta of its constituent particles. The resulting equations of motion are quite complicated because every particle, in principle, interacts with all the others. They do, however, possess an simple yet important property. Everything is reversible, in the sense that the equations of motion remain the same if one changes the direction of time and changes the direction of motion for all the particles. Consequently, one cannot tell whether a movie of atomic motions is being played forwards or backwards.

This means that the Gibbs entropy is actually a constant of the motion: it neither increases nor decreases during Hamiltonian evolution.

But what about the second law of thermodynamics? This tells us that the entropy of a system tends to increase. Our everyday experience tells us this too: we know that physical systems tend to evolve towards states of increased disorder. Heat never passes from a cold body to a hot one. Pour milk into coffee and everything rapidly mixes. How can this directionality in thermodynamics be reconciled with the completely reversible character of microscopic physics?

The answer to this puzzle is surprisingly simple, as long as you use a sensible interpretation of entropy that arises from the idea that its probabilistic nature represents not randomness (whatever that means) but incompleteness of information. I’m talking, of course, about the Bayesian view of probability.

 First you need to recognize that experimental measurements do not involve describing every individual atomic property (the “microstates” of the system), but large-scale average things like pressure and temperature (these are the “macrostates”). Appropriate macroscopic quantities are chosen by us as useful things to use because they allow us to describe the results of experiments and measurements in a  robust and repeatable way. By definition, however, they involve a substantial coarse-graining of our description of the system.

Suppose we perform an idealized experiment that starts from some initial macrostate. In general this will generally be consistent with a number – probably a very large number – of initial microstates. As the experiment continues the system evolves along a Hamiltonian path so that the initial microstate will evolve into a definite final microstate. This is perfectly symmetrical and reversible. But the point is that we can never have enough information to predict exactly where in the final phase space the system will end up because we haven’t specified all the details of which initial microstate we were in.  Determinism does not in itself allow predictability; you need information too.

If we choose macro-variables so that our experiments are reproducible it is inevitable that the set of microstates consistent with the final macrostate will usually be larger than the set of microstates consistent with the initial macrostate, at least  in any realistic system. Our lack of knowledge means that the probability distribution of the final state is smeared out over a larger phase space volume at the end than at the start. The entropy thus increases, not because of anything happening at the microscopic level but because our definition of macrovariables requires it.

ham

This is illustrated in the Figure. Each individual microstate in the initial collection evolves into one state in the final collection: the narrow arrows represent Hamiltonian evolution.

 

However, given only a finite amount of information about the initial state these trajectories can’t be as well defined as this. This requires the set of final microstates has to acquire a  sort of “buffer zone” around the strictly Hamiltonian core;  this is the only way to ensure that measurements on such systems will be reproducible.

The “theoretical” Gibbs entropy remains exactly constant during this kind of evolution, and it is precisely this property that requires the experimental entropy to increase. There is no microscopic explanation of the second law. It arises from our attempt to shoe-horn microscopic behaviour into framework furnished by macroscopic experiments.

Another, perhaps even more compelling demonstration of the so-called subjective nature of probability (and hence entropy) is furnished by Maxwell’s demon. This little imp first made its appearance in 1867 or thereabouts and subsequently led a very colourful and influential life. The idea is extremely simple: imagine we have a box divided into two partitions, A and B. The wall dividing the two sections contains a tiny door which can be opened and closed by a “demon” – a microscopic being “whose faculties are so sharpened that he can follow every molecule in its course”. The demon wishes to play havoc with the second law of thermodynamics so he looks out for particularly fast moving molecules in partition A and opens the door to allow them (and only them) to pass into partition B. He does the opposite thing with partition B, looking out for particularly sluggish molecules and opening the door to let them into partition A when they approach.

The net result of the demon’s work is that the fast-moving particles from A are preferentially moved into B and the slower particles from B are gradually moved into A. The net result is that the average kinetic energy of A molecules steadily decreases while that of B molecules increases. In effect, heat is transferred from a cold body to a hot body, something that is forbidden by the second law.

All this talk of demons probably makes this sound rather frivolous, but it is a serious paradox that puzzled many great minds. Until it was resolved in 1929 by Leo Szilard. He showed that the second law of thermodynamics would not actually be violated if entropy of the entire system (i.e. box + demon) increased by an amount every time the demon measured the speed of a molecule so he could decide whether to let it out from one side of the box into the other. This amount of entropy is precisely enough to balance the apparent decrease in entropy caused by the gradual migration of fast molecules from A into B. This illustrates very clearly that there is a real connection between the demon’s state of knowledge and the physical entropy of the system.

By now it should be clear why there is some sense of the word subjective that does apply to entropy. It is not subjective in the sense that anyone can choose entropy to mean whatever they like, but it is subjective in the sense that it is something to do with the way we manage our knowledge about nature rather than about nature itself. I know from experience, however, that many physicists feel very uncomfortable about the idea that entropy might be subjective even in this sense.

On the other hand, I feel completely comfortable about the notion:. I even think it’s obvious. To see why, consider the example I gave above about pouring milk into coffee. We are all used to the idea that the nice swirly pattern you get when you first pour the milk in is a state of relatively low entropy. The parts of the phase space of the coffee + milk system that contain such nice separations of black and white are few and far between. It’s much more likely that the system will end up as a “mixed” state. But then how well mixed the coffee is depends on your ability to resolve the size of the milk droplets. An observer with good eyesight would see less mixing than one with poor eyesight. And an observer who couldn’t perceive the difference between milk and coffee would see perfect mixing. In this case entropy, like beauty, is definitely in the eye of the beholder.

The refusal of many physicists to accept the subjective nature of entropy arises, as do so many misconceptions in physics, from the wrong view of probability.

Full Blast

Posted in Science Politics, The Universe and Stuff with tags , , , , , on April 9, 2009 by telescoper

Yesterday, Paolo Calisse and I were paid a visit by a reporter (Martin Shipton) and a photographer from Welsh newspaper The Western Mail who wanted to cover the sad story of Clover.

Paolo is heavily involved with Clover, but I was a bit hesitant about doing this because I’m not really part of the Clover team. Paolo suggested it might be an advantage that I wasn’t so directly involved as I might be able to give a more balanced view of the importance of the experiment than him. Anyway, the story came out today in the newspaper and is available online too.

DrThis is the picture they took of me and Paolo in the Clover lab, fiddling with the cryostat. I’ve already had my leg pulled enough about pretending to be an instrumentalist for the photograph so no jokes please…

 

 

 

 

In the same issue of the paper there is another feature about Cardiff’s astronomy research, concerning BLAST (Balloon-borne Large Aperture Submillimetre Telescope). This is a much happier story, as it marks the release of results from a highly successful science run from 2006. In the print version of the Western Mail the two stories were run on the same page, one above the other, making very effectively the point that cutting the funding of the Astronomy Instrumentation Group jeopardizes a great deal of world-leading research besides Clover itself. And when I say “world-leading” I mean it, whatever the RAE panel might have thought.

A deluge of articles about BLAST appeared on the arXiv today, one of which is now published in Nature. I thought I’d put up the abstracts here in order to draw attention to these results. The author lists contain many Cardiff authors and, as you’ll see, the results are both fascinating and wide-ranging. I’ve put links to the arXiv after each abstract:

Title: BLAST: Correlations in the Cosmic Far-Infrared Background at 250, 350, and 500 microns Reveal Clustering of Star-Forming Galaxies

Authors: Marco P. Viero, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Carrie J. MacTavish, Gaelen Marsden, Peter G. Martin, Philip Mauskopf, Lorenzo Moncelsi, Mattia Negrello, Calvin B. Netterfield, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Donald V. Wiebe

We detect correlations in the cosmic far-infrared background due to the clustering of star-forming galaxies, in observations made with the Balloon-borne Large Aperture Submillimeter Telescope (BLAST), at 250, 350, and 500 microns. Since the star-forming galaxies which make up the far-infrared background are expected to trace the underlying dark matter in a biased way, measuring clustering in the far infrared background provides a way to relate star formation directly to structure formation. We test the plausibility of the result by fitting a simple halo model to the data. We derive an effective bias b_eff = 2.2 +/- 0.2, effective mass log(M_eff/M_sun) = 13.2 (+0.3/-0.8), and minimum mass log(M_min/M_sun) = 9.9 (+1.5/-1.7). This is the first robust clustering measurement at submillimeter wavelengths.

http://arxiv.org/abs/0904.1200

Title: Over half of the far-infrared background light comes from galaxies at z >= 1.2

Authors: Mark J. Devlin, Peter A. R. Ade, Itziar Aretxaga, James J. Bock, Edward L. Chapin, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Peter G. Martin, Philip Mauskopf, Lorenzo Moncelsi, Calvin B. Netterfield, Henry Ngo, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

Journal-ref: Nature, vol. 458, 737-739 (2009) DOI: 10.1038/nature07918

Submillimetre surveys during the past decade have discovered a population of luminous, high-redshift, dusty starburst galaxies. In the redshift range 1 <= z <= 4, these massive submillimetre galaxies go through a phase characterized by optically obscured star formation at rates several hundred times that in the local Universe. Half of the starlight from this highly energetic process is absorbed and thermally re-radiated by clouds of dust at temperatures near 30 K with spectral energy distributions peaking at 100 microns in the rest frame. At 1 <= z <= 4, the peak is redshifted to wavelengths between 200 and 500 microns. The cumulative effect of these galaxies is to yield extragalactic optical and far-infrared backgrounds with approximately equal energy densities. Since the initial detection of the far-infrared background (FIRB), higher-resolution experiments have sought to decompose this integrated radiation into the contributions from individual galaxies. Here we report the results of an extragalactic survey at 250, 350 and 500 microns. Combining our results at 500 microns with those at 24 microns, we determine that all of the FIRB comes from individual galaxies, with galaxies at z >= 1.2 accounting for 70 per cent of it. As expected, at the longest wavelengths the signal is dominated by ultraluminous galaxies at z > 1.

http://arxiv.org/abs/0904.1201

Title: The Balloon-borne Large Aperture Submillimeter Telescope (BLAST) 2006:
Calibration and Flight Performance

Authors: Matthew D. P. Truch, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, Simon R. Dicker, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Peter G. Martin, Philip Mauskopf, Lorenzo Moncelsi, Calvin B. Netterfield, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas E. Thomas, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

The Balloon-borne Large Aperture Submillimeter Telescope (BLAST) operated successfully during a 250-hour flight over Antarctica in December 2006 (BLAST06). As part of the calibration and pointing procedures, the red hypergiant star VY CMa was observed and used as the primary calibrator. Details of the overall BLAST06 calibration procedure are discussed. The 1-sigma absolute calibration is accurate to 10, 12, and 13% at the 250, 350, and 500 micron bands, respectively. The errors are highly correlated between bands
resulting in much lower error for the derived shape of the 250-500 micron continuum. The overall pointing error is <5″ rms for the 36, 42, and 60″ beams. The performance of the optics and pointing systems is discussed.

http://arxiv.org/abs/0904.1202

Title: A Bright Submillimeter Source in the Bullet Cluster (1E0657–56) Field Detected with BLAST

Authors: Marie Rex, Peter A. R. Ade, Itziar Aretxaga, James J. Bock, Edward L. Chapin, Mark J. Devlin, Simon R. Dicker, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Peter G. Martin, Philip Mauskopf, Calvin B. Netterfield, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

We present the 250, 350, and 500 micron detection of bright submillimeter emission in the direction of the Bullet Cluster measured by the Balloon-borne Large-Aperture Submillimeter Telescope (BLAST). The 500 micron centroid is coincident with an AzTEC 1.1 millimeter detection at a position close to the peak lensing magnification produced by the cluster. However, the 250 micron and 350 micron emission is resolved and elongated, with centroid positions shifted toward the south of the AzTEC source and a differential shift between bands that cannot be explained by pointing uncertainties. We therefore conclude that the BLAST detection is contaminated by emission from foreground galaxies associated with the Bullet Cluster. The submillimeter redshift estimate based on 250-1100 micron photometry at the position of the AzTEC source is z_phot = 2.9 (+0.6/-0.3), consistent with the infrared color redshift estimation of the most likely Spitzer IRAC counterpart. These flux densities indicate an apparent far-infrared luminosity of L_FIR = 2E13 L_sun. When the amplification due to the gravitational lensing of the cluster is removed, the intrinsic far-infrared luminosity of the source is found to be L_FIR <= 1E12 L_sun, consistent with typical luminous infrared galaxies.

http://arxiv.org/abs/0904.1203

Title: Radio and mid-infrared identification of BLAST source counterparts in the Chandra Deep Field South

Authors: Simon Dye, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, James S. Dunlop, Stephen A. Eales, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Philip Mauskopf, Lorenzo Moncelsi, Calvin B. Netterfield, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

We have identified radio and/or mid-infrared counterparts to 198 out of 351 sources detected at >= 5 sigma over ~ 9 sq. degrees centered on the Chandra Deep Field South (CDFS) by the Balloon-borne Large Aperture Submillimeter Telescope (BLAST) at 250, 350, and 500 microns. We have matched 92 of these counterparts to optical sources with previously derived photometric redshifts and fitted SEDs to the BLAST fluxes and fluxes at 70 and 160 microns acquired with the Spitzer Space Telescope. In this way, we have constrained dust temperatures, total far-infrared/submillimeter luminosities and star formation rates for each source. Our findings show that the BLAST sources lie at significantly lower redshifts and have significantly lower rest-frame dust temperatures compared to submm sources detected in surveys conducted at 850 microns. We demonstrate that an apparent increase in dust temperature with redshift in our sample arises as a result of selection effects. This paper
constitutes the public release of the multi-wavelength catalog of >= 5 sigma BLAST sources contained within the full ~ 9 sq. degree survey area.

http://arxiv.org/abs/0904.1204

Title: BLAST: Resolving the Cosmic Submillimeter Background

Authors: Gaelen Marsden, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, Simon R. Dicker, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Philip Mauskopf, Benjamin Magnelli, Lorenzo Moncelsi, Calvin B. Netterfield, Henry Ngo, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

The Balloon-borne Large Aperture Submillimeter Telescope (BLAST) has made one square-degree, deep, confusion-limited maps at three different bands, centered on the Great Observatories Origins Deep Survey South field. By calculating the covariance of these maps with catalogs of 24 micron sources from the Far-Infrared Deep Extragalactic Legacy Survey (FIDEL), we have determined that the total submillimeter intensities are 8.60 +/- 0.59, 4.93 +/- 0.34, and 2.27 +/- 0.20 nW m^-2 sr^-1 at 250, 350, and 500 microns, respectively. These numbers are more precise than previous estimates of the cosmic infrared background (CIB) and are consistent with 24 micron-selected galaxies generating the full intensity of the CIB. We find that more than half of the CIB originates from sources at z >= 1.2. At all BLAST wavelengths, the relative intensity of high-z sources is higher for 24 micron-faint sources than it is for 24 micron-bright sources. Galaxies identified very broadly as AGN by their Spitzer Infrared Array Camera (IRAC) colors contribute 32-48% of the CIB, although X-ray-selected AGN contribute only 7%. BzK-selected galaxies are found to be brighter than typical 24 micron-selected galaxies in the BLAST bands, and contribute 32-42% of the CIB. These data provide high-precision constraints for models of the evolution of the number density and intensity of star-forming galaxies at high redshift.

http://arxiv.org/abs/0904.1205

Title: BLAST: A Far-Infrared Measurement of the History of Star Formation

Authors: Enzo Pascale, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, Simon Dye, Steve A. Eales, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Philip Mauskopf, Lorenzo Moncelsi, Calvin B. Netterfield, Luca Olmi, Guillaume Patanchon, Marie Rex, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

We use measurements from the Balloon-borne Large Aperture Sub-millimeter Telescope (BLAST) at wavelengths spanning 250 to 500 microns, combined with data from the Spitzer Infrared telescope and ground-based optical surveys in GOODS-S, to determine the average star formation rate of the galaxies that comprise the cosmic infrared background (CIB) radiation from 70 to 500 microns, at redshifts 0 < z < 3. We find that different redshifts are preferentially probed at different wavelengths within this range, with most of the 70 micron background generated at z < ~1 and the 500 micron background generated at z >~1. The spectral coverage of BLAST and Spitzer in the region of the peak of the background at ~200 microns allows us to directly estimate the mean physical properties (temperature, bolometric luminosity and mass) of the dust in the galaxies responsible for contributing more than 80% of the CIB. By utilizing available redshift information we directly measure the evolution of the far infrared luminosity density and therefore the optically obscured star formation history up to redshift z ~3.

http://arxiv.org/abs/0904.1206

Title: BLAST: The Mass Function, Lifetimes, and Properties of Intermediate Mass Cores from a 50 Square Degree Submillimeter Galactic Survey in Vela (l = ~265)

Authors: Calvin. B. Netterfield, Peter A. R. Ade, James J. Bock, Edward L. Chapin, Mark J. Devlin, Matthew Griffin, Joshua O. Gundersen, Mark Halpern, Peter C. Hargrave, David H. Hughes, Jeff Klein, Gaelen Marsden, Peter G. Martin, Phillip Mauskopf, Luca Olmi, Enzo Pascale, Guillaume Patanchon, Marie Rex, Arabindo Roy, Douglas Scott, Christopher Semisch, Nicholas Thomas, Matthew D. P. Truch, Carole Tucker, Gregory S. Tucker, Marco P. Viero, Donald V. Wiebe

We present first results from an unbiased, 50 square degree submillimeter Galactic survey at 250, 350, and 500 microns from the 2006 flight of the Balloon-borne Large Aperture Submillimeter Telescope (BLAST). The map has resolution ranging from 36″ to 60″ in the three submillimeter bands spanning the thermal emission peak of cold starless cores. We determine the temperature, luminosity, and mass of more than a thousand compact sources in a range of evolutionary stages and an unbiased statistical characterization of the population. From comparison with C^18 O data, we find the dust opacity per gas mass, kappa/R = 0.16 cm^2/g at 250 microns, for cold clumps. We find that 2% of the mass of the molecular gas over this diverse region is in cores colder than 14 K, and that the mass function for these cold cores is consistent with a power law with index alpha = -3.22 +/- 0.14 over the mass range 14 M_sun < M < 80 M_sun, steeper than the Salpeter alpha = -2.35 initial massfunction for stars. Additionally, we infer a mass dependent cold core lifetime of tau(M) = 4E6 (M/20 M_sun)^-0.9 years — longer than what has been found in previous surveys of either low or high mass cores, and significantly longer than free fall or turbulent decay time scales. This implies some form of non-thermal support for cold cores during this early stage of star formation.

http://arxiv.org/abs/0904.1207

You can find a lot more detailed information on the dedicated BLAST website.

Random Thoughts: Points and Poisson (d’Avril)

Posted in The Universe and Stuff with tags , , , on April 4, 2009 by telescoper

I’ve got a thing about randomness. For a start I don’t like the word, because it covers such a multitude of sins. People talk about there being randomness in nature when what they really mean is that they don’t know how to predict outcomes perfectly. That’s not quite the same thing as things being inherently unpredictable; statements about the nature of reality are ontological, whereas I think randomness is only a useful concept in an epistemological sense. It describes our lack of knowledge: just because we don’t know how to predict doesn’t mean that it can’t be predicted.

Nevertheless there are useful mathematical definitions of randomness and it is also (somtimes) useful to make mathematical models that display random behaviour in a well-defined sense, especially in situations where one has to take into account the effects of noise.

I thought it would be fun to illustrate one such model. In a point process, the random element is a “dot” that occurs at some location in time or space. Such processes occur in wide range of contexts: arrivals of buses at a bus stop, photons in a detector, darts on a dartboard, and so on.

Let us suppose that we think of such a process happening in time, although what follows can straightforwardly be generalised to things happening over an area (such a dartboard) or within some higher-dimensional region. It is also possible to invest the points with some other attributes; processes like this are sometimes called marked point processes, but I won’t discuss them here.

The “most” random way of constructing a simple point process is to assume that each event happens independently of every other event, and that there is a constant probability per unit time of an event happening. This type of process is called a Poisson process, after the French mathematician Siméon-Denis Poisson, who was born in 1781. He was one of the most creative and original physicists of all time: besides fundamental work on electrostatics and the theory of magnetism for which he is famous, he also built greatly upon Laplace’s work in probability theory. His principal result was to derive a formula giving the number of random events if the probability of each one is very low. The Poisson distribution, as it is now known and which I will come to shortly, is related to this original calculation; it was subsequently shown that this distribution amounts to a limiting of the binomial distribution. Just to add to the connections between probability theory and astronomy, it is worth mentioning that in 1833 Poisson wrote an important paper on the motion of the Moon.

In a finite interval of duration T the mean (or expected) number of events for a Poisson process will obviously just be proportional to the product of the rate per unit time and T itself; call this product l.
 
The full distribution is then

This gives the probability that a finite interval contains exactly x events. It can be neatly derived from the binomial distribution by dividing the interval into a very large number of very tiny pieces, each one of which becomes a Bernoulli trial. The probability of success (i.e. of an event occurring) in each trial is extremely small, but the number of trials becomes extremely large in such a way that the mean number of successes is l. In this limit the binomial distribution takes the form of the above expression. The variance of this distribution is interesting: it is alsol.  This means that the typical fluctuations within the interval are of order the square root of l on a mean level of l, so the fractional variation is of the famous “one over root n” form that is a useful estimate of the expected variation in point processes.  Indeed, it’s a useful rule-of-thumb for estimating likely fluctuation levels in a host of statistical situations.

If football were a Poisson process with a mean number of goals per game of, say, 2 then would expect must games to have 2 plus or minus 1.4 (the square root of 2)  goals, i.e. between about 0.6 and 3.4. That is actually not far from what is observed and the distribution of goals per game in football matches is actually quite close to a Poisson distribution.

This idea can be straightforwardly extended to higher dimensional processes. If points are scattered over an area with a constant probability per unit area then the mean number in a finite area will also be some number l and the same formula applies.

As a matter of fact I first learned about the Poisson distribution when I was at school, doing A-level mathematics (which in those days actually included some mathematics). The example used by the teacher to illustrate this particular bit of probability theory was a two-dimensional one from biology. The skin of a fish was divided into little squares of equal area, and the number of parasites found in each square was counted. A histogram of these numbers accurately follows the Poisson form. For years I laboured under the delusion that it was given this name because it was something to do with fish, but then I never was very quick on the uptake.

This is all very well, but point processes are not always of this Poisson form. Points can be clustered, so that having one point at a given position increases the conditional probability of having others nearby. For example, galaxies like those shown in the nice picture are distributed throughout space in a clustered pattern that is very far from the Poisson form. But it’s very difficult to tell from just looking at the picture. What is needed is a rigorous statistical analysis.

The statistical description of clustered point patterns is a fascinating subject, because it makes contact with the way in which our eyes and brain perceive pattern. I’ve spent a large part of my research career trying to figure out efficient ways of quantifying pattern in an objective way and I can tell you it’s not easy, especially when the data are prone to systematic errors and glitches. I can only touch on the subject here, but to see what I am talking about look at the two patterns below:

 

pointbpointa

You will have to take my word for it that one of these is a realization of a two-dimensional Poisson point process and the other contains correlations between the points. One therefore has a real pattern to it, and one is a realization of a completely unstructured random process.

I show this example in popular talks and get the audience to vote on which one is the random one. The vast majority usually think that the top  is the one that is random and the bottom one is the one with structure to it. It is not hard to see why. The top pattern is very smooth (what one would naively expect for a constant probability of finding a point at any position in the two-dimensional space) , whereas the bottom one seems to offer a profusion of linear, filamentary features and densely concentrated clusters.

In fact, it’s the bottom  picture that was generated by a Poisson process using a  Monte Carlo random number generator. All the structure that is visually apparent is imposed by our own sensory apparatus, which has evolved to be so good at discerning patterns that it finds them when they’re not even there!

The top  process is also generated by a Monte Carlo technique, but the algorithm is more complicated. In this case the presence of a point at some location suppresses the probability of having other points in the vicinity. Each event has a zone of avoidance around it; the points are therefore anticorrelated. The result of this is that the pattern is much smoother than a truly random process should be. In fact, this simulation has nothing to do with galaxy clustering really. The algorithm used to generate it was meant to mimic the behaviour of glow-worms which tend to eat each other if they get  too close. That’s why they spread themselves out in space more uniformly than in the random pattern.

Incidentally, I got both pictures from Stephen Jay Gould’s collection of essays Bully for Brontosaurus and used them, with appropriate credit and copyright permission, in my own book From Cosmos to Chaos. I forgot to say this in earlier versions of this post.

The tendency to find things that are not there is quite well known to astronomers. The constellations which we all recognize so easily are not physical associations of stars, but are just chance alignments on the sky of things at vastly different distances in space. That is not to say that they are random, but the pattern they form is not caused by direct correlations between the stars. Galaxies form real three-dimensional physical associations through their direct gravitational effect on one another.

People are actually pretty hopeless at understanding what “really” random processes look like, probably because the word random is used so often in very imprecise ways and they don’t know what it means in a specific context like this.  The point about random processes, even simpler ones like repeated tossing of a coin, is that coincidences happen much more frequently than one might suppose.

I suppose there is an evolutionary reason why our brains like to impose order on things in a general way. More specifically scientists often use perceived patterns in order to construct hypotheses. However these hypotheses must be tested objectively and often the initial impressions turn out to be figments of the imagination, like the canals on Mars.

Now, I think I’ll complain to wordpress about the widget that links pages to a “random blog post”.

I’m sure it’s not really random….

Talking Planck

Posted in The Universe and Stuff with tags , , on April 3, 2009 by telescoper

Since the Planck mission is due to be launched very soon, I thought it would be nice to put this lecture by George Efstathiou here in order to give some background. It’s from a page of science talks about Planck.

George is the Professor of Astrophysics (1909) at the University of Cambridge. The 1909 isn’t when he was born, but when the Chair he holds was set up. I have a hundred-year-old Chair in my house too.
He is also the Director of the impressive Kavli Institute for Cosmology.
He’s a leading member of the Planck science team and is coordinating the UK effort that will be applied to analysing the data. He’s an FRS, citation millionaire, and general all-round clever clogs. He would cut an even more impressive figure were it not for the fact that he supports Arsenal.

Clover Story

Posted in Science Politics, The Universe and Stuff with tags , , on April 2, 2009 by telescoper

Just a quick note for those interested in the story of Clover, Physics World have run a news item on their website.

You may also like to read the article by Alan Heavens over on the e-astronomer.

Note added on Monday 6th April: the Nature slant on the story is now published online, complete with quote from yours truly…

Another update (9th April). Welsh Newspaper The Western Mail has now run a story on the clover cancellation and there was a short item on the BBC Radio Wales News this evening.

Another update (14th April). A statement from Walter Gear, Principal Investigator of the Clover project, about the current status of Clover has been placed on the Cardiff University School of Physics & Astronomy web pages.

Update: 22nd April 2009. Here is the text of a piece I wrote for today’s Research Fortnight:

An undeserved end

Science projects don’t get much purer than CLOVER, an experiment designed to search for evidence of the existence of primordial gravitational waves by making ultra-sensitive measurements of the polarisation of the cosmic microwave background.

From its vantage point in the Atacama Desert in Chile, CLOVER was intended to probe the state of the universe when it was less than a billionth of a billionth of a second old, to test our understanding of the Big Bang theory. Unfortunately, the Science and Technology Facilities Council says it is cancelling funding for the experiment.

Gravitational waves have been studied theoretically and are known to be intimately related to the structure of space-time itself, the understanding of which is arguably the fundamental goal of modern science. The first discovery of the presence of gravitational waves will lead to the emergence of a brand new area of physics. In anticipation of this new science, the CLOVER team—entirely British, with members in the universities of Cardiff, Cambridge, Oxford and Manchester—has established a technical capability in the UK that is second to none. Cancellation will prevent the team from making direct experimental observations of the universe that would not only have been of immense scientific importance, but could also have had deep cultural significance.

So if CLOVER is so good, why is it being cancelled?

The answer lies in an unfortunate combination of circumstances. CLOVER was initially funded in 2004, with
£4.8 million from the Particle Physics and Astronomy Research Council, one of the forerunners of the STFC. This budget was not sufficient to complete the experiment, for two main reasons. First, the original grant did not include the costs of setting up a site, which was originally to be provided by overseas collaborators in Antarctica. When this option fell through, the cost of the alternative site in Chile (approximately £0.8m) had
to be found. Second, there were delays due to technical challenges, such as the need to develop some of the world’s most sensitive far-infrared superconducting cameras. So, the CLOVER team was unable to complete the project within the original budget, and went back to the STFC to request extra money. This brought a third factor into play.

Since 2007, the research councils, including the STFC, have changed their method of funding university-based research. In the new full-economic-costs regime, costs are substantially higher than at the time of the original award. These elements combined to leave the CLOVER team with a shortfall of about £2.6m, bringing the overall cost to completion to about £7.5m, although the increase in resources required would be only around 20 per cent if calculated on the pre-FEC basis of the initial funding.

Unfortunately, despite receiving strong support from the scientific community and being rated extremely highly in recent prioritisation exercises, the STFC Council has decided that it does not have the funds and has abruptly cancelled the CLOVER experiment.

The background to this decision is one of dire financial circumstances within the research council. Created in 2007, the STFC was set up with insufficient funding to continue all the programmes that it inherited from its predecessors. The deficit (of around £80m) has led to swingeing cuts in research grants over the past year. The pound has also fallen dramatically against the euro, increasing the cost of subscriptions to the European Space Agency, Cern and the European Southern
Observatory. The balance sheet of the STFC is now in total disarray. CLOVER is the first casualty in what may become a large-scale cull of fundamental science projects.

The STFC’s decision on CLOVER means that an important instrument will be lost, and the millions already spent on it wasted. The technology will be difficult to replace. The many gifted scientists who have been working on CLOVER will have to leave the UK to continue in the field, and are unlikely to return. Their fate is unlikely to tempt younger people into a career in science either.

In cancelling CLOVER, the council has effectively closed the door on UK involvement in cosmic microwave background science in general, an area that has already led to two Nobel prizes for physics. The decision also provides worrying evidence that the STFC seems to be turning away from fundamental science towards technology- driven projects. For example the lunar probe Moonlite has recently won funding for initial development studies without ever passing through the rigorous peer review required of CLOVER. If this really is the way the STFC is going, then we may be witnessing the beginning of the end for British astronomy.

The Waste Land

Posted in Poetry, Science Politics, The Universe and Stuff with tags , , , , on April 1, 2009 by telescoper

APRIL is the cruellest month, sending
Clover into the dead land, ditching
The great for the dire, erring
Dead heads caused spring pain.
Keith Mason fucked it up, smothering
Good science with tons of shit, ending
Our little dream; we’re the losers.

After The Waste Land, Part I: The Burial of the Dead, by T.S. Eliot.

Clover and Out

Posted in Science Politics, The Universe and Stuff with tags , , , , , , , , , on March 31, 2009 by telescoper

One of the most exciting challenges facing the current generation of cosmologists is to locate in the pattern of fluctuations in the cosmic microwave background evidence for the primordial gravitational waves predicted by models of the Universe that involve inflation.

Looking only at the temperature variation across the sky, it is not possible to distinguish between tensor  (gravitational wave) and scalar (density wave) contributions  (both of which are predicted to be excited during the inflationary epoch).  However, scattering of photons off electrons is expected to leave the radiation slightly polarized (at the level of a few percent). This gives us additional information in the form of the  polarization angle at each point on the sky and this extra clue should, in principle, enable us to disentangle the tensor and scalar components.

The polarization signal can be decomposed into two basic types depending on whether the pattern has  odd or even parity, as shown in the nice diagram (from a paper by James Bartlett)

The top row shows the E-mode (which look the same when reflected in a mirror and can be produced by either scalar or tensor modes) and the bottom shows the B-mode (which have a definite handedness that changes when mirror-reflected and which can’t be generated by scalar modes because they can’t have odd parity).

The B-mode is therefore (in principle)  a clean diagnostic of the presence of gravitational waves in the early Universe. Unfortunately, however, the B-mode is predicted to be very small, about 100 times smaller than the E-mode, and foreground contamination is likely to be a very serious issue for any experiment trying to detect it.

An experiment called Clover (involving the Universities of  Cardiff, Oxford, Cambridge and Manchester) was designed to detect the primordial B-mode signal from its vantage point in Chile. You can read more about the way it works at the dedicated webpages here at Cardiff and at Oxford. I won’t describe it in more detail here, for reasons which will become obvious.

The chance to get involved in a high-profile cosmological experiment was one of the reasons I moved to Cardiff a couple of years ago, and I was looking forward to seeing the data arriving for analysis. Although I’m primarily a theorist, I have some experience in advanced statistical methods that might have been useful in analysing the output.  It would have been fun blogging about it too.

Unfortunately, however, none of that is ever going to happen. Because of its budget crisis, and despite the fact that it has spent a large amount (£4.5M) on it already,  STFC has just decided to withdraw the funding needed to complete it (£2.5M)  and cancel the Clover experiment.

Clover wasn’t the only B-mode experiment in the game. Its rivals include QUIET and SPIDER, both based in the States. It wasn’t clear that Clover would have won the race, but now that we know  it’s a non-runner  we can be sure it won’t.

**** Energy

Posted in Poetry, The Universe and Stuff with tags , , , , , on March 30, 2009 by telescoper

The phrase expletive deleted was made popular at the time of Watergate after the release of the expurgated tapes made by Richard Nixon in the Oval Office when he was President of the United States of America. These showed that, as well as been a complete crook, he was practically unable to speak a single sentence without including a swear word.

Nowadays the word expletive is generally taken to mean an oath or exclamation, particularly if it is obscene, but that’s not quite what it really means. Derived from the latin verb explere (“to fill out”) from which the past participle is expletus, the meaning of the word in the context of English grammar is  “something added to a phrase or sentence that isn’t strictly needed for the grammatical sense”.  An expletive is added either to fill a syntactical role or, in a poem, simply to make a line fit some metrical rule.

Examples of the former can be found in constructions like “It takes two to Tango” or “There is lots of crime in Nottingham”; neither  “it” nor “there” should really be needed but English likes to have something before the verb.

The second kind of use is illustrated wonderfully by Alexander Pope in his Essay on Criticism, which is a kind of guide to what to avoid in writing poetry. It’s a tour de force for its perceptiveness and humour. The following excerpt is pricelessly apt

These equal syllables alone require,
Tho’ oft the open vowels tire;
While expletives their feeble aid do join;
And ten low words oft creep in one dull line

Here the expletive is “do”,  and it is cleverly incorporated in the line talking about expletives, adding  the syllable needed to fit with a strict pentameter. Apparently, poets often used this construction before Pope attacked it but it quickly fell from favour afterwards.

His other prosodic targets are the “open vowels” which means initial vowels that produce an ugly glottal sound, such as in “oft” (especially ugly when following “Tho”). The last line is brilliant too, showing how using only monosyllabic “low” words makes for a line that plods along tediously just like it says.

It’s amazing how much Pope managed to fit into this poem, given the restrictions imposed by the closed couplet structure he adopted. Each idea is compressed into a unit of twenty syllables, two lines of ten syllables with a rhyme at the end of each. This is such an impressive exercise in word-play that it reminds me a lot of the skill showed by the best cryptic crossword setters. Needless to say I’m no more successful at writing poetry than I am at setting crossword clues.

After my talk in Dublin last Friday, somebody in the audience asked me what I thought about Dark Energy. There’s some discussion in the comments after my post on that too.

The Dark Energy is an ingredient added to the standard model of cosmology to reconcile  observations of a flat Universe with a matter density that seems too low to account for it.

Other than that it makes the  cosmological metric work out satisfactorily (geddit?), we don’t understand what Dark Energy means and would rather it wasn’t there.  Most people think the resulting model is inelegant or even ugly.

In other words, it’s an expletive…

Dublin Back

Posted in Art, Books, Talks and Reviews, Crosswords, The Universe and Stuff with tags , , , , on March 28, 2009 by telescoper

I’m just back from a flying visit to Dublin, where I gave a talk yesterday at a meeting of the Astronomical Science Group of Ireland (ASGI), an organization which promotes scientific collaborations between individuals and institutions on both sides of the border between Northern Ireland and the Republic of Eire. The venue for the twice-yearly meetings moves around both countries, but this time it was held in the splendid environment of Trinity College, Dublin.

It turned out to be an easy trip from Cardiff to Dublin and my first opportunity to try out Cardiff’s fine little airport. A small airline called Air Arann operates the route to Dublin from there, and it all went to schedule despite the plane having to struggle against a 70 mph head wind across the Irish sea. For our small propeller-driven plane, that made a signficant difference to the flying time.

Arriving in Dublin on Thursday I had time to have a nice dinner before settling in to my hotel in the Temple Bar region of the city. There’s a huge concentration of bars and nightclubs there and it’s a traditional area for Stag and Hen Parties. There was plenty of evidence of drunken debauchery going on into the early hours of the morning, which remind me of the way the Irish rugby fans carried on last weekend in Cardiff.

Anyway, the meeting itself was interesting with a wide range of talks most of which were given by PhD students. I enjoy meetings where the younger scientists are encouraged to speak; too many conferences involve the same people giving the same talk time after time. Solar Physics was particularly  well represented, and I learned quite a bit about about things that are far from my own province. 

There isn’t much actual cosmology done in Ireland (North or South) so my brief as invited speaker was to give an overview of the current state of the field for astronomers who are not  experts in cosmological matters. I therefore gave a summary of the concordance model which I’ve blogged about before and then made some comments about things that might point to a more complete theory of the Universe. I also mentioned some of the anomalies in the cosmic microwave background that I’ve also blogged about on here.

I usually use this piece of Hieronymus Bosch The Last Judgement to illustrate my feelings about the concordance model:

das_letzte_gericht

 

 
The top part represents the concordance cosmology. It clearly features an eminent cosmologist surrounded by postdoctoral researchers. Everything appears to be in heavenly harmony, surrounded by a radiant glow of self-satisfaction. The trumpets represent various forms of exaggerated press coverage.

But if you step back from it, and get the whole thing in a proper perspective, you realise that there’s an awful lot going on underneath that’s not so pleasant or easy to interptet. I don’t know what’s going down below there although the unfortunate figures slaving away in miserable conditions and suffering unimaginable torments are obviously supposed to represent graduate students.

The main point is that the concordance model is based on rather strange foundations: nobody understands what the dark matter and dark energy are, for example. Even more fundamentally, the whole thing is based on a shotgun marriage between general relativity and quantum field theory which is doomed to fail somewhere along the line.

Far from being a final theory of the Universe I think we should treat our standard model as a working hypothesis and actively look for departures from it. I’m not at all against the model. As models go, it’s very successful. It’s a good one, but it’s still just a model.

That reminds me of the school report I got after my first year at the Royal Grammar School. The summary at the bottom described me as a “model student”. I was so thrilled I went and looked up the word model in a dictionary and found it said “a small imitation of the real thing.”

Anyway, the talk went down pretty well (I think) and after a quick glass of Guinness (which definitely went down well) I was back to Dublin airport and home to Cardiff soon after that: Cardiff airport to my house was less than twenty minutes. I greatly enjoyed my short visit and was delighted to be asked to do a couple of seminars back there in the near future.

I was in a  good mood when I got home, which got even better when I found out that I won the latest Crossword competition in the Times Literary Supplement. And the prize isn’t even a dictionary. It’s cash!

Social Physics and Astronomy

Posted in The Universe and Stuff with tags , , , , , on March 23, 2009 by telescoper

When I give popular talks about Cosmology,  I sometimes look for appropriate analogies or metaphors in television programmes about forensic science, such as CSI: Crime Scene Investigation which I used to watch quite regularly (to the disdain of many of my colleagues and friends). Cosmology is methodologically similar to forensic science because it is generally necessary in both these fields to proceed by observation and inference, rather than experiment and deduction: cosmologists have only one Universe;  forensic scientists have only one scene of the crime. They can collect trace evidence, look for fingerprints, establish or falsify alibis, and so on. But they can’t do what a laboratory physicist or chemist would typically try to do: perform a series of similar experimental crimes under slightly different physical conditions. What we have to do in cosmology is the same as what detectives do when pursuing an investigation: make inferences and deductions within the framework of a hypothesis that we continually subject to empirical test. This process carries on until reasonable doubt is exhausted, if that ever happens.

Of course there is much more pressure on detectives to prove guilt than there is on cosmologists to establish the truth about our Cosmos. That’s just as well, because there is still a very great deal we do not know about how the Universe works.I have a feeling that I’ve stretched this analogy to breaking point but at least it provides some kind of excuse for writing about an interesting historical connection between astronomy and forensic science by way of the social sciences.

The gentleman shown in the picture on the left is Lambert Adolphe Jacques Quételet, a Belgian astronomer who lived from 1796 to 1874. His principal research interest was in the field of celestial mechanics. He was also an expert in statistics. In Quételet’s  time it was by no means unusual for astronomers to well-versed in statistics, but he  was exceptionally distinguished in that field. Indeed, Quételet has been called “the father of modern statistics”. and, amongst other things he was responsible for organizing the first ever international conference on statistics in Paris in 1853.

 

His fame as a statistician owed less to its applications to astronomy, however, than the fact that in 1835 he had written a very influential book which, in English, was titled A Treatise on Man but whose somewhat more verbose original French title included the phrase physique sociale (“social physics”).

Apparently the philosopher Auguste Comte was annoyed that Quételet appropriated the phrase “social physics” because he did not approve of the quantitative statistical-based  approach that it had come to represent. For that reason Comte  ditched the term from his own work and invented the subject of  sociology…

Quételet had been struck not only by the regular motions performed by the planets across the sky, but also by the existence of strong patterns in social phenomena, such as suicides and crime. If statistics was essential for understanding the former, should it not be deployed in the study of the latter? Quételet’s first book was an attempt to apply statistical methods to the development of man’s physical and intellectual faculties. His follow-up book Anthropometry, or the Measurement of Different Faculties in Man (1871) carried these ideas further, at the expense of a much clumsier title.

This foray into “social physics” was controversial at the time, for good reason. It also made Quételet extremely famous in his lifetime and his influence became widespread. For example, Francis Galton wrote about the deep impact Quételet had on a certain British lady:

Her statistics were more than a study, they were indeed her religion. For her Quételet was the hero as scientist, and the presentation copy of his “Social Physics” is annotated on every page. Florence Nightingale believed – and in all the actions of her life acted on that belief – that the administrator could only be successful if he were guided by statistical knowledge. The legislator – to say nothing of the politician – too often failed for want of this knowledge. Nay, she went further; she held that the universe – including human communities – was evolving in accordance with a divine plan; that it was man’s business to endeavour to understand this plan and guide his actions in sympathy with it. But to understand God’s thoughts, she held we must study statistics, for these are the measure of His purpose. Thus the study of statistics was for her a religious duty.

The name of the lady in question was Florence Nightingale. Not many people know that she was an adept statistician who was an early advocate of the use of pie charts to represent data graphically; she apparently found them useful when dealing with dim-witted army officers and dimmer-witted politicians.

The type of thinking described in the quote  also spawned a number of highly unsavoury developments in pseudoscience, such as the eugenics movement (in which Galton himself was involved), and some of the vile activities related to it that were carried out in Nazi Germany. But an idea is not responsible for the people who believe in it, and Quételet’s work did lead to many good things, such as the beginnings of forensic science.

A young medical student by the name of Louis-Adolphe Bertillon was excited by the whole idea of “social physics”, to the extent that he found himself imprisoned for his dangerous ideas during the revolution of 1848, along with one of his Professors, Achile Guillard, who later invented the subject of demography, the study of racial groups and regional populations. When they were both released, Bertillon became a close confidante of Guillard and eventually married his daughter Zoé. Their second son, Adolphe Bertillon, turned out to be a prodigy.

Young Adolphe was so inspired by Quételet’s work, which had no doubt been introduced to him by his father, that he hit upon a novel way to solve crimes. He would create a database of measured physical characteristics of convicted criminals. He chose 11 basic measurements, including length and width of head, right ear, forearm, middle and ring fingers, left foot, height, length of trunk, and so on. On their own none of these individual characteristics could be probative, but it ought to be possible to use a large number of different measurements to establish identity with a very high probability. Indeed, after two years’ study, Bertillon reckoned that the chances of two individuals having all 11 measurements in common were about four million to one. He further improved the system by adding photographs, in portrait and from the side, and a note of any special marks, like scars or moles.

Bertillonage, as this system became known, was rather cumbersome but proved highly successful in a number of high-profile criminal cases in Paris. By 1892, Bertillon was exceedingly famous but nowadays the word bertillonage only appears in places like the Observer’s Azed crossword.

The main reason why Bertillon’s fame subsided and his system fell into disuse was the development of an alternative and much simpler method of criminal identification: fingerprints. The first systematic use of fingerprints on a large scale was implemented in India in 1858 in an attempt to stamp out electoral fraud.

The name of the British civil servant who had the idea of using fingerprinting in this way was William Herschel, although I don’t think he was related to the astronomer of the same name.

That would be too much of a coincidence.