Archive for astronomy

The Problem of the Steady State

Posted in Science Politics with tags , , on February 24, 2009 by telescoper

Just as a quick postscript to my recent item about proposed changes to the method of funding PhD students by STFC, let me point out the following simple calculation.

Assume that the number of permanent academic positions in a given field (e.g. astronomy) remains constant over time. If that is the case, each retirement (or other form of departure) from a permanent position will be replaced by one, presumably junior, scientist.

This means that over an academic career, on average, each academic will produce just one PhD who will get a permanent job. This of course doesn’t count students coming in from abroad, or those getting faculty positions abroad but in the case of the UK these are probably relatively small corrections.

Under the present supply of PhD studentships an academic can expect to get a PhD student at least once every three years or so. At a minimum, therefore, over a 30 year career one can expect to have ten PhD students. A great many supervisors have more PhD students than this, but this just makes the odds worse. The expectation is that only one of these will get a permanent job in the UK. The others (nine out of ten, according to my conservative estimate) above must either leave the field or the country to find permanent employment.

The arithmetic of this situation is a simple fact of life, but I’m not sure how many prospective PhD students are aware of it.

Scientiae Doctores

Posted in Science Politics with tags , , , on February 22, 2009 by telescoper

The season for recruiting new research students is well and truly upon us and at the same the Science & Technology Facilities Council (STFC) is consulting about changing the way that it allocates PhD studentships to departments.

Most postgraduate students studying for PhDs in Astronomy are funded by STFC (although some Universities also fund their own internal studentships). The result of this arrangement is that successful applicants to a PhD course can receive a stipend which amounts to about £13K per annum. It’s not a huge amount of money, but it is a stipend rather than a salary so it’s tax-free. Since a PhD student also remains a student and therefore qualifies for various other fringe benefits (Council Tax, student discounts, etc), it’s not actually a bad deal for the student. Anyway, if it were significantly more then it’s possible PhD students would have to start paying back their student loans, which would make things worse. STFC also pays a tuition fee to the University concerned, but this is done directly and the student doesn’t even see that element of the funding.

Since about 1995, PPARC and then STFC has funded research studentships in areas within its remit by means of peer review. Departments have bid for studentships (every two years) and a panel awards an allocation depending on the quality of the bid. Of course, everyone asks for many more studentships than are available so what you get is a fraction of what you ask for. I wrote the application for the first ever quota studentships for the Astronomy group at the University of Nottingham, and did it again a couple of times after that. Each time, despite going into best bullshit mode to write the case, I was frustrated by the relatively small number of studentships we were awarded. Although we succeeded in building up gradually from zero to 2-3 per year, it was a very slow process.

In recent years, the funding mechanism has evolved slightly so that studentship fees and stipends were devolved to the departments concerned in terms of Doctoral Training Grants (DTGs) rather than being administered centrally by PPARC/STFC. In the old days, students used to get their stipend from PPARC/STFC whereas now they are paid by their department from a cash grant.

Anyway, for various reasons (chief among them being no doubt to save administrative costs) STFC has decided to consult on changes to the mechanism for allocating the DTGs to the various departments around the country. The most serious proposed change is to follow the practice at the Engineering and Physical Sciences Research Council (EPSRC) and dispense with peer review. Instead, the proposal is to award studentships based on a formula involving how successful the department is at obtaining postdoctoral research assistant (PDRA) support from STFC.

Here is the proposed formula:

 Specifically, the studentship award per department should be proportional to the product of volume and average quality per academic within the department, that is to:

 

V * Q

 

The Committee has followed guidance in developing measures of V and Q that are non-subjective, repeatable and transparent.  The volume V is defined as the number of academics (including Fellows) eligible to hold STFC research grants. The  quality Q is measured by the number of STFC-funded PDRAs (P) awarded per academic (i.e. P/V), since this measures the success of the academic staff in securing STFC funding for PDRAs through peer-review.  More precisely we define quality per academic as Q = [1 +(P/V)].

 

Although the Committee felt this definition of quality applied primarily to responsive-mode PDRAs, it agreed that PDRAs on project grants should be included, but with a weighting, relative to responsive-mode, of 0.33.

 

Using these definitions, the Committee recommends that the studentship award per department should be proportional to a simple product of volume and average quality per academic within the department, that is to:

 

N(students) µ V * Q

 

where Q = [1 + (P/V)]

 

And so the departmental quota is proportional to: 

 

V[1+(P/V)] = V+P

 

In addition, recognising that very small departments offer more limited training opportunities on their own, a threshold is proposed, such that no studentships are awarded for V < 3. Instead, these very small departments/groups would be able to collaborate with other larger departments in seeking STFC studentship support.

 

Hence

 

         N(students) µ V+P  for V ³ 3

                             = 0        for V < 3

 

The constant of proportionality is chosen such that the total number of studentships equals the number available for allocation.

 

 

I think this is a fairly reasonable proposal, actually. The one thing I don’t really understand relates to the fact that STFC doesn’t just fund PDRAs on its grants, but under the Full Economic Cost regime (FEC), it also pays for fractions of academic staff effort for people working on its projects. On my recent successful STFC grant, for example, I was awarded 25% of my time (i.e. 0.25 FTE, full-time-equivalent) to do the research as well as a PDRA. Since the proposal above will have to cope with the question of what staff are “eligible” then why not make the quantity V proportional to the total FTEs funded, or at least only count those for whom some FEC time is allocated? And why not include staff FTE in the Q-factor too?

My guess is that such a modification wouldn’t make much difference to astronomy departments, but the original proposal has caused cries of anguish from particle physicists. This is because the number of PDRAs in particle physics is much smaller than in astronomy, so many large groups face a big reduction in their PhD quota. Including FEC numbers in the mix might well smooth the transition for them. For your information, the number of PDRAs per active astronomy researcher  is around 0.5 at present.

Anyway, the deadline for consulting on this has passed (on February 20th) so we now wait to see what STFC actually does. Probably the consultation period is a purely cosmetic exercise anyway and what will emerge is exactly what was proposed.

If you ask me (and nobody did), all this is mere tinkering. I think there are serious problems with graduate funding in the UK and these require much more radical remedies. At the risk of (and indeed with the intention of) being provocative, here is my diagnosis and suggested remedies:

  • There are too many PhDs in astronomy. STFC funded 160 studentships in 2006, compared with 88 in 2000. There are nowhere near enough PDRA positions to accommodate this number of PhDs in academic research. And even those who get their first PDRA position have very limited prospects of getting a permanent job. The result is a generation of disaffected students employed as low-paid assistants for 3-4 years and then thrown aside when they have got their PhD.
  • Of course, applicants for PhD places don’t know what research is really like and some will leave academia of their own volition when they find out that it’s not for them. In my experience, though, most graduate applicants simply don’t realise how heavily the odds are stacked against them. Less than one in ten can possibly stay in research in the long term, and the more PhDs are funded the worse the odds against them become.
  • The short duration of a British PhD disadvantages our students with respect to those from the USA or continental europe, who all do a lengthy Masters course before taking their PhD. These take at least 5 years to complete.  The result is that our home-grown PhDs are seriously disadvantaged in the job market against competitors from abroad. Similar points have been made forcefully by Ian Halliday.
  • My remedy is simple. Reduce the number of studentships but extend each one to five years and require each hosting department to provide a proper graduate school with intensive graduate-level courses to make up for the progressive reduction in content of undergraduate physics courses.
  • Even more unpopularly, I think the UK should scrap 4 years Masters (MPhys) programmes and embrace the structure of the Bologna agreement, i.e. a universal 3+2+3 structure of 3 years Bachelors, 2 years’ Masters and three years PhD.
  • Currently STFC stipends can only be paid to UK nationals and residents. It’s an open secret that most departments would preferentially recruit European physics graduates to their PhD positions if they were allowed to do so, because their undergraduate preparation is much better than that provided in UK universities. I propose that we abandon this protectionism and open up PhD opportunities to European applications, just as we would legally have to do if a PhD were considered to be a job.
  • Finally, I think the UK should consider the introduction of a common graduate entrance examination, perhaps based on the US GRE, to ensure the maintenance of appropriate standards for postgraduate entry and eligibility for STFC funding.

There are of course some advantages to the current British PhD system. For one thing, the PhD is earned very quickly. I was 25 when I got my PhD, and already had several publications. Most of my European collaborators were at least 30 before they got theirs (additional years have to be added for national service in many countries, but we don’t have it in the UK). But I am painfully aware that my technical knowledge outside the immediate area of my PhD is much thinner than most academics in the field. Now, in middle age, I feel like a long-distance runner who had inadequate preparation, went off too fast at the start of the race, and is now struggling along while people overtake him with monotonous regularity.

The nature of research in astronomy and cosmology has changed so much in the 20 years since I got my PhD that the old system has to go. Instead of tinkering with funding formula, driven principally by the need to save adminstrative costs within STFC, we need a radical overhaul of the entire graduate education system in the UK, involving all research councils and their political masters.

Unfortunately, though, for the time being at least the politicians have other more pressing matters to worry about, such the collapse of the economy.

What’s all the Noise?

Posted in Science Politics, The Universe and Stuff with tags , , , , on January 18, 2009 by telescoper

Now there’s a funny thing…

I’ve just come across a news item from last week which I followed up by looking at the official NASA press release. I’m very slow to pick up on things these days, but I thought I’d mention it anyway.

The experiment concerned is called ARCADE 2, which is an somewhat contrived acronym derived from Absolute Radiometer for Cosmology, Astrophysics and Diffuse Emission. It is essentially a balloon-borne detector designed to analyse radio waves with frequencies in the range 3 to 90 Ghz. The experiment actually flew in 2006, so it has clearly taken considerable time to analyse the resulting data.

Being on a balloon that flies for a relatively short time (2.5 hours in this case) means that only a part of the sky was mapped, amounting to about 7% of the whole celestial sphere but that is enough to map a sizeable piece of the Galaxy as well as a fairly representative chunk of deep space.

There are four science papers on the arXiv about this mission: one describes the instrument itself; another discusses radio emission from our own galaxy, the Milky Way; the third discusses the overall contribution of extragalactic origin in the frequency range covered by the instrument; the last discusses the implications about extragalactic sources of radio emission.

The thing that jumps out from this collection of very interesting science papers is that there is an unexplained, roughly isotropic, background of radio noise, consistent with a power-law spectrum. Of course to isolate this component requires removing known radio emission from our Galaxy and from identified extragalactic sources, as well as understanding the systematics of the radiometer during its flight. But after a careful analysis of these the authors present strong evidence of excess emission over and above known sources. The spectrum of this radio buzz falls quite steeply with frequency so appears in the two long-wavelength channels at 3 and 8 GHz.

So where does this come from? Well, we just don’t know.

The problem is that no sensible extrapolation of known radio sources to high redshift appears to be able to generate an integrated flux equivalent to that observed. Here is a bit of the discussion from the paper:

It is possible to imagine that an unknown population of discrete sources exist below the flux limit of existing surveys. We argue earlier that these cannot be a simple extension of the source counts of star-forming galaxies. As a toy model, we consider a population of sources distributed with a delta function in flux a factor of 10 fainter than the 8.4 GHz survey limit of Fomalont et al. (2002). At a flux of 0.75 μJy, it would take over 1100 such sources per square arcmin to produce the unexplained emission we see at 3.20 GHz, assuming a frequency index of −2.56. This source density is more than two orders of magnitude higher than expected from extrapolation to the same flux limit of the known source population. It is, however, only modestly greater than the surface density of objects revealed in the faintest optical surveys, e.g., the Hubble Ultra Deep Field (Beckwith et al. 2006).  The unexplained emission might result from an early population of non thermal emission from low-luminosity AGN; such a source would evade the constraint implied by the far-IR measurements.

The point is that ordinary galaxies produce a broad spectrum of radiation and it is difficult to boost the flux at one frequency without violating limits imposed at others. It might be able to invoke Active Galactic Nuclei (AGN) to do the trick, but I’m not sure. I am sure there’ll be a lot work going on trying to see how this might fit in with all the other things we know about galaxy formation and evolution but for the time being it’s a mystery.

I’m equally sure that these results will spawn a plethora of more esoteric theoretical explanations, inevitably including the ridiculous as well as perhaps the sublime. Charged dark matter springs to mind.

Or maybe it’s not even extragalactic. Could it be from an unknown source inside the Milky Way? If so, it might shed some light on the curiosities we find in the cosmic microwave background that I’ve mentioned here and there, but it seems to peak at too low a frequency to account for much of the overall microwave sky temperature.

But it does have a lesson for astronomy funders. ARCADE 2 is a very cheap experiment (by NASA standards). Moreover, the science goals of the experiment did not include “discovering a new cosmic background”. It just goes to show that even in these times of big, expensive and narrowly targetted missions there is still space for serendipity.

The Physics Overview

Posted in Science Politics with tags , , , , , , , , on January 17, 2009 by telescoper

I found out by accident the other day that the Panels conducting the 2008 Research Assessment Exercise have now published their subject overviews, in which they comment trends within each discipline.

Heading straight for the overview produced by the panel for Physics (which is available together with two other panels here),I found some interesting points, some of which relate to comments posted on my previous items about the RAE results (here and here) until I terminated the discussion.

One issue that concerns many physicists is how the research profiles produced by the RAE panel will translate into funding. I’ve taken the liberty of extracting a couple of paragraphs from the report to show what they think. (For those of you not up with the jargon, UoA19 is the Unit of Assessment 19, which is Physics).

The sub-panel is pleased with how much of the research fell into the 4* category and that this excellence is widely spread so that many smaller departments have their share of work assessed at the highest grade. Every submitted department to UoA19 had at least 70% of their overall quality profile at 2* or above, i.e. internationally recognised or above.

Sub-panel 19 takes the view that the research agenda of any group, or of any individual for that matter, is interspersed with fallow periods during which the next phase of the research is planned and during which outputs may be relatively incremental, even if of high scientific quality. In the normal course of events successful departments with a long term view will have a number of outputs at the 3* and 2* level indicating that the groundwork is being laid for the next set of 4* work. This is most obviously true for those teams involved with very major experiments in the big sciences, but also applies to some degree in small science. Thus the quality profile is a dynamic entity and even among groups of very high international standing there is likely to be cyclic variation in the relative amounts of 3* and 4* work according to the rhythm of their research programmes. Most departments have what we would consider a healthy balance between the perceived quality levels. The subpanel strongly believes that the entire overall profile should be considered when measuring the quality of a department, rather than focussing on the 4* component only.

I think this is very sensible, but for more reasons than are stated. For a start the judgement of what is 4* or 3* must be to some extent subjective and it would be crazy to allocate funding entirely according to the fraction of 4* work. I’ve heard informally that the error in any of the percentages for any assessment is plus or minus 10%, which also argues for a conservative formula. However one might argue about the outcome, the panels clearly spent a lot of time and effort determining the profiles so it would seem to make sense to use all the information they provide rather than just a part.

Curiously, though, the panel made no comment about why it is that physics came out so much worse than chemistry in the 2008 exercise (about one-third of the chemistry departments in the country had a profile-weighted quality mark higher than or equal to the highest-rated physics department). Perhaps they just think UK chemistry is a lot better than UK physics.

Anyway, as I said, the issue most of us are worrying about is how this will translate into cash. I suspect HEFCE hasn’t worked this out at all yet either. The panel clearly thinks that money shouldn’t just follow the 4* research, but the HEFCE managers might differ. If they do wish to follow a drastically selective policy they’ve got a very big problem: most physics departments are rated very close together in score. Any attempt to separate them using the entire profile would be hard to achieve and even harder to justify.

The panel also made a specific comment about Wales and Scotland, which is particularly interesting for me (being here in Cardiff):

Sub-panel 19 regards the Scottish Universities Physics Alliance collaboration between Scottish departments as a highly positive development enhancing the quality of research in Scotland. South of the border other collaborations have also been formed with similar objectives. On the other hand we note with concern the performance of three Welsh departments where strategic management did not seem to have been as effective as elsewhere.

I’m not sure whether the dig about Welsh physics departments is aimed at the Welsh funding agency HEFCW or the individual university groups; SUPA was set up with the strong involvement of SFC and various other physics groupings in England (such as the Midlands Physics Alliance) were actively encouraged by HEFCE. It is true, though, that the 3 active physics departments in Wales (Cardiff, Swansea and Aberystwyth) all did quite poorly in the RAE. In the last RAE, HEFCW did not apply as selective a funding formula as its English counterpart HEFCE with the result that Cardiff didn’t get as much research funding as it would if it had been in England. One might argue that this affected the performance this time around, but I’m not sure about this as it’s not clear how any extra funding coming into Cardiff would have been spent. I doubt if HEFCW will do any different this time either. Welsh politics has a strong North-South issue going on, so HEFCW will probably feel it has to maintain a department in the North. It therefore can’t penalise Aberystwyth too badly for its poor RAE showing. The other two departments are larger and had very similar profiles (Swansea better than Cardiff, in fact) so there’s very little justification for being too selective there either.

The panel remarked on the success of SUPA which received a substantial injection of cash from the Scottish Funding Council (SFC) and which has led to new appointments in strategic areas in several Scottish universities. I’m a little bit skeptical about the long-term benefits of this because the universities themselves will have to pick up the tab for these positions when the initial funding dries up. Although it will have bought them extra points on the RAE score the continuing financial viability of physics departments is far from guaranteed because nobody yet knows whether they will gain as much cash from the outcome as they spent to achieve it. The same goes for other universities, particularly Nottingham, who have massively increased their research activity with cash from various sources and consequently done very well in the RAE. But will they get back as much as they have put in? It remains to be seen.

What I would say about SUPA is that it has definitely given Scottish physics a higher profile, largely from the appointment of Ian Halliday to front it. He is an astute political strategist and respected scientist who performed impressively as Chief Executive of the now-defunct Particle Physics and Astronomy Research Council and is also President of the European Science Foundation. Having such a prominent figurehead gives the alliance more muscle than a group of departmental heads would ever hope to have.

So should there be a Welsh version of SUPA? Perhaps WUPA?

Well, Swansea and Cardiff certainly share some research interests in the area of condensed-matter physics but their largest activities (Astronomy in Cardiff, Particle Physics in Swansea) are pretty independent. It seems to me to be to be well worth thinking of some sort of initiative to pool resources and try to make Welsh physics a bit less parochial, but the question is how to do it. At coffee the other day, I suggested an initiative in the area of astroparticle physics could bring in genuinely high quality researchers as well as establishing synergy between Swansea and Cardiff, which are only an hour apart by train. The idea went down like a lead balloon, but I still think it’s a good one. Whether HEFCW has either the resources or the inclination to do something like it is another matter, even if the departments themselves were to come round.

Anyway, I’m sure there will be quite a lot more discussion about our post-RAE strategy if and when we learn more about the funding implications. I personally think we could do with a radical re-think of the way physics in Wales is organized and could do with a champion who has the clout of Scotland’s SUPA-man.

The mystery as far as I am concerned remains why Cardiff did so badly in the ratings. I think the first quote may offer part of the explanation because we have large groups in Astronomical Instrumentation and Gravitational Physics, both of which have very long lead periods. However, I am surprised and saddened by the fact that the fraction rated at 4* is so very low. We need to find out why. Urgently.

Professor Who?

Posted in Biographical, Music, Television, The Universe and Stuff with tags , , , , , on January 7, 2009 by telescoper

As a Professor of Astrophysics I am often asked “Why on Earth did you take up such a crazy subject?”

I guess many astronomers, physicists and other scientists have to answer this sort of question. For many of them there is probably a romantic reason, such as seeing the rings of Saturn or the majesty of the Milky Way on a dark night. Others will probably have been inspired by TV documentary series such as The Sky at Night, Carl Sagan’s Cosmos or even Horizon which, believe it or not, actually used to be quite good but which is nowadays uniformly dire. Or it could have been something a bit more mundane but no less stimulating such as a very good science teacher at school.

When I’m asked this question I’d love to be able to put my hand on my heart and give an answer of that sort but the truth is really quite a long way from those possibilities. The thing that probably did more than anything else to get me interested in science was a Science Fiction TV series or rather not exactly the series but the opening titles.

The first episode of Doctor Who was broadcast in the year of my birth, so I don’t remember it at all, but I do remember the astonishing effect the credits had on my imagination when I saw later episodes as a small child. Here are some tests for the sequence as it appeared in the very first series featuring William Hartnell as the first Doctor.

To a younger audience it probably all seems quite tame, but I think there’s a haunting, unearthly beauty to the shapes conjured up by Bernard Lodge. Having virtually no budget for graphics, he experimented in a darkened studio with an old-fashioned TV camera and a piece of black card with Doctor Who written on it in white. He created the spooky kaleidoscopic patterns you see by simply pointing the camera so it could see into its own monitor, thus producing a sort of electronic hall of mirrors.

What is so fascinating to me is how a relatively simple underlying concept could produce a rich assortment of patterns, particularly how they seem to take on an almost organic aspect as they merge and transform. I’ve continued to be struck by the idea that complexity could be produced by relatively simple natural laws which is one of the essential features of astrophysics and cosmology. As a practical demonstration of the universality of physics this sequence takes some beating.

As well as these strange and wonderful images, the titles also featured a pioneering piece of electronic music. Officially the composer was Ron Grainer, but he wasn’t very interested in the commission and simply scribbled the theme down and left it to the BBC to turn it into something useable. In stepped the wonderful Delia Derbyshire, unsung heroine of the BBC Radiophonic Workshop who, with only the crudest electronic equipment available, turned it into a little masterpiece. Ethereal yet propulsive, the original theme from Doctor Who is definitely one of my absolute favourite pieces of music and I’m glad to see that Delia Derbyshire is now receiving the acclaim she deserves from serious music critics.

It’s ironic that I’ve now moved to Cardiff where new programmes of Doctor Who and its spin-off, the anagrammatic Torchwood, are made. One of the great things about the early episodes of Doctor Who was that the technology simply didn’t exist to do very good special effects. The scripts were consequently very careful to let the viewers’ imagination do all the work. That’s what made it so good. I’m pleased that the more recent incarnations of this show also don’t go overboard on the visuals. Perhaps thats a conscious attempt to appeal to people who saw the old ones as well as those too young to have done so. It’s just a pity the modern opening title music is so bad…

Anyway, I still love Doctor Who after all these years. It must sound daft to say that it inspired me to take up astrophysics, but it’s truer than any other explanation I can think of. Of course the career path is slightly different from a Timelord, but only slightly.

At any rate I think The Doctor is overdue for promotion. How about Professor Who?

Power isn’t Everything

Posted in The Universe and Stuff with tags , , , , , , , on January 6, 2009 by telescoper

WMapThe picture above shows the latest available all-sky map of fluctuations in the temperature of the cosmic microwave background across the sky as revealed by the Wilkinson Microwave Anisotropy Probe, known to its friends as WMAP.

I’ve spent many long hours fiddling with the data coming from the WMAP experiment, partly because I’ve never quite got over the fact that such wonderful data actually exists. When I started my doctorate in 1985 the whole field of CMB analysis was so much pie in the sky, as no experiments had yet been performed with the sensitivity to reveal the structures we now see. This is because they are very faint and easily buried in noise. The fluctuations in temperature from pixel to pixel across the sky are of order one part in a hundred thousand of the mean temperature (i.e. about 30 microKelvin on a background temperature of about 3 Kelvin). That’s smoother than the surface of a billiard ball. That’s why it took such a long time to make the map shown above, and why it is such a triumphant piece of science.

I blogged a few days ago about the idea that the structure we see in this map was produced by sound waves reverberating around the early Universe. The techniques cosmologists use to analyse this sound are similar to those used in branches of acoustics except that we only see things in projection on the celestial sphere which requires a bit of special consideration.

One of the things that sticks in my brain from my undergraduate years is being told that if a physicist doesn’t know what they are doing they should start by making a Fourier transform. This breaks down the phenomenon being studied into a set of independent plane waves with different wavelengths corresponding to the different tones present in a complicated sound.

It’s often very good advice to do such a decomposition for one-dimensional time series or fluctuation fields in three-dimensional Cartesian space, even you do know what you’re doing, but it doesn’t work with a sphere because plane waves don’t fit properly on a curved surface. Fortunately, however, there is a tried-and-tested alternative involving spherical harmonics rather than plane waves.

Spherical harmonics are quite complicated beasts mathematically but they have pretty similar properties to Fourier harmonics in many respects. In particular they are represented as complex numbers having real and imaginary parts or, equivalently, an amplitude and a phase (usually called an argument by mathematicians). The latter representation is the most useful one for CMB fluctuations because the simplest versions of inflation predict that the phases of each of the spherical harmonic modes should be randomly distributed. What this really means is that there is no information content in their distribution so that the harmonic modes are in a state of maximum statistical disorder or entropy. This property also guarantees that the distribution of fluctuations over the sky should have a Gaussian distribution.

If you accept that the fluctuations are Gaussian then only the amplitudes of the spherical harmonic coefficients are useful. Indeed, their statistical properties can be specified entirely by the variance of these amplitudes as a function of mode frequency. This pre-eminently important function is called the power-spectrum of the fluctuations, and it is shown here for the WMAP data:

080999_powerspectrumm

Although the units on the axes are a bit strange it doesn”t require too much imagination to interpret this in terms of a sound spectrum. There is a characteristic tone (at the position of the peak) plus a couple of overtones. However these features are not sharp so the overall sound is not at all musical.

If the Gaussian assumption is correct then the power-spectrum contains all the useful statistical information to be gleaned from the CMB sky, which is why so much emphasis has been placed on extracting it accurately from the data.

Conversely, though, the power spectrum is completely insenstive to any information in the distribution of spherical harmonic phases. If something beyond the standard model made the Universe non-Gaussian it would affect the phases of the harmonic modes in a way that would make them non-random.

So far, so good. It sounds like it should be a straightforward job to work out whether the WMAP phases are random or not. Unfortunately, though, this task is heavily complicated by the presence of noise and systematics which can be quite easily cleaned from the spectrum but not from more sophisticated descriptors. All we can say so far is that the data seem to be consistent with a Gaussian distribution.

However, I thought I’d end with a bit of fun and show you how important phase information could actually be, if only we could find a good way of exploiting it. Let’s start with a map of the Earth, with the colour representing height of the surface above mean sea level:

sw_world

You can see the major mountain ranges (Andes, Himalayas) quite clearly as red in this picture and note how high Antarctica is…that’s one of the reasons so much astronomy is done there.

Now, using the same colour scale we have the WMAP data again (in Galactic coordinates).

sw_ilc

The virture of this map is that it shows how smooth the microwave sky is compared to the surface of the Earth. Note also that you can see a bit of crud in the plane of the Milky Way that serves as a reminder of the difficulty of cleaning the foregrounds out.

Clearly these two maps have completely different power spectra. The Earth is dominated by large features made from long-wavelength modes whereas the CMB sky has relatively more small-scale fuzz.

Now I’m going to play with these maps in the following rather peculiar way. First, I make a spherical harmonic transform of each of them. This gives me two sets of complex numbers, one for the Earth and one for WMAP. Following the usual fashion, I think of these as two sets of amplitudes and two sets of phases. Note that the spherical harmonic transformation preserves all the information in the sky maps, it’s just a different representation.

Now what I do is swap the amplitudes and phases for the two maps. First, I take the amplitudes of WMAP and put them with the phases for the Earth. That gives me the spherical harmonic representation of a new data set which I can reveal by doing an inverse spherical transform:

sw_worldphases

This map has exactly the same amplitudes for each mode as the WMAP data and therefore possesses an identical power spectrum to that shown above. Clearly, though, this particular CMB sky is not compatible with the standard cosmological model! Notice that all the strongly localised features such as coastlines appear by virtue of information contained in the phases but absent from the power-spectrum.

To understand this think how sharp features appear in a Fourier transform. A sharp spike at a specific location actually produces a broad spectrum of Fourier modes with different frequencies. These modes have to add in coherently at the location of the spike and cancel out everywhere else, so their phases are strongly correlated. A sea of white noise also has a flat power spectrum but has random phases. The key difference between these two configurations is not revealed by their spectra but by their phases.

Fortunately there is nothing quite as wacky as a picture of the Earth in the real data, but it makes the point that there are more things in Heaven and Earth than can be described in terms of the power spectrum!

Finally, perhaps in your mind’s eye you might consider what it might look lie to do the reverse experiment: recombine the phases of WMAP with the amplitudes of the Earth.

sw_ilcphases

If the WMAP data are actually Gaussian, then this map is a sort of random-phase realisation of the Earth’s power spectrum. Alternatively you can see that it is the result of running a kind of weird low-pass filter over the WMAP fluctuations. The only striking things it reveals are (i) a big blue hole associated with foreground contamination, (ii) a suspicious excess of red in the galactic plane owing to the same problem, and (iiI) a strong North-South asymmetry arising from the presence of Antarctica.

There’s no great scientific result here, just a proof that spherical harmonics can be fun.

PS. These pictures were made by a former PhD student of mine, Patrick Dineen, who has since quit astronomy to work in high finance. I hope he is weathering the global financial storm!

Misplaced Confidence

Posted in Bad Statistics, The Universe and Stuff with tags , , , on December 10, 2008 by telescoper

From time to time I’ve been posting items about the improper use of statistics. My colleague Ant Whitworth just showed me an astronomical example drawn from his own field of star formation and found in a recent paper by Matthew Bate from the University of Exeter.

The paper is a lengthy and complicated one involving the use of extensive numerical calculations to figure out the effect of radiative feedback on the process of star formation. The theoretical side of this subject is fiendishly difficult, to the extent that it is difficult to make any progress with pencil-and-paper techinques, and Matthew is one of the leading experts in the use of computational methods to tackle problems in this area.

One of the main issues Matthew was investigating was whether radiative feedback had any effect on the initial mass function of the stars in his calculations. The key results are shown in the picture below (Figure 8 from the paper) in terms of cumulative distributions of the star masses in various different situations.

untitled

The question that arises from such data is whether these empirical distributions differ significantly from each other or whether they are consistent with the variations that would naturally arise in different samples drawn from the same distribution. The most interesting ones are the two distributions to the right of the plot that appear to lie almost on top of each other.

Because the samples are very small (only 13 and 15 objects respectively) one can’t reasonably test for goodness-of-fit using the standard chi-squared test because of discreteness effects and because not much is known about the error distribution. To do the statistics, therefore, Matthew uses a popular non-parametric method called the Kolmogorov-Smirnov test which uses the maximum deviation D between the two distributions as a figure of merit to decide whether they match. If D is very large then it is not probable that it can have arisen from the same distribution. If it is smaller then it might have. As for what happens if it is very small then you’ll have to wait a bit.

This is an example of a standard (frequentist) hypothesis test in which the null hypothesis is that the empirical distributions are calculated from independent samples drawn from the same underlying form. The probability of a value of D arising as large as the measured one can be calculated assuming the null is true and is then the significance level of the test. If there’s only a 1% chance of it being as large as the measured value then the significance level is 1%.

So far, so good.

But then, in describing the results of the K-S test the paper states

A Kolmogorov-Smirnov (K-S) test on the …. distributions gives a 99.97% probability that the two IMFs were drawn from the same underlying distribution (i.e. they are statistically indistinguishable).

Agh! No it doesn’t! What it gives is a probability of 99.97% that the chance deviation between the two distributions is expected to be larger than that actually measured. In other words, the two distributions are surprisingly close to each other. But the significance level merely specifies the probability that you would reject the null-hypothesis if it were correct. It says nothing at all about the probability that the null hypothesis is correct. To make that sort of statement you would need to specify an alternative distribution, calculate the distribution of D based on it, and hence determine the statistical power of the test. Without specifying an alternative hypothesis all you can say is that you have failed to reject the null hypothesis.

Or better still, if you have an alternative hypothesis you can forget about power and significance and instead work out the relative probability of the two hypotheses using a proper Bayesian approach.

You might also reasonably ask why might D be so very small? If you find an improbably low value of chi-squared then it usually means either that somebody has cheated or that the data are not independent (which is assumed for the basis of the test). Qualitatively the same thing happens with a KS test.

In fact these two distributions can’t be thought of as independent samples anyway as they are computed from the same initial conditions but with various knobs turned on or off to include different physics. They are not “samples” drawn from the same population but slightly different versions of the same sample. The probability emerging from the KS machinery is therefore meaningless anyway in this context.

So a correct statement of the result would be that the deviation between the two computed distributions is much smaller than one would expect to arise from two independent samples of the same size drawn from the same population.

That’s a much less dramatic statement than is contained in the paper, but has the advantage of not being bollocks.

Operation Skyphoto

Posted in The Universe and Stuff with tags , on December 9, 2008 by telescoper

Katherine Blundell from Oxford just contacted me with a request that I post the following message. I’m more than happy to oblige.

Dear All,

There is a one-off opportunity to buy vintage prints of the original photographic plates of the Palomar All-Sky Survey. Although no longer useful for science (they fell into disuse two decades ago because of modern data digitization) they make rather handsome objets d’art when suitably mounted and framed.

These prints are for sale to raise money for Alexander Thatte’s treatment for leukemia – Alexander is the 5-year old son of two of our colleagues.

The mounted/framed photographs could make very nice Christmas presents. For a small additional payment we can deliver them to you already tastefully gift-wrapped.

A very limited number of photographs have kindly been signed by Jocelyn Bell Burnell – please email us if you wish to request one of these.

Please see http://www.physics.ox.ac.uk/skyphoto for an order form and further details. Please feel free to forward this email to anyone whom you think might be interested in purchasing a piece of astronomical history, and helping a child in need.

Best wishes,

Katherine & the Astro Grads

I can’t think of a better Christmas gift for an astronomer.

Go on. You know you want to.

If you leave it too late to buy your presents you might end up buying something really naff. Like a paperweight.

Look, I’ve even made it easier for you. Just click the link here.

So now there’s no excuse. Do it. Buy one. Now.

Pluralia Tantum

Posted in Literature, Pedantry with tags , , , on December 5, 2008 by telescoper

Meanwhile, over on the e-astronomer, Andy Lawrence recently posted an item about the lamentable tendency of astronomers to abuse the English language. The focus of his venom was “extincted”, a word used by many astro-types as an adjective to describe the state of affairs when light from a source (e.g. a quasar) has suffered “extinction” by intervening matter. “Extinction” is formed from the verb “extinguish” in the same way that “distinction” is formed from “distinguish”. Nobody would describe a professor as “distincted” (certainly not if it is Andy Lawrence) so, clearly, “extincted” is inappropriate. Actually if you really want to nit-pick you could object to “extinction” being applied to an object such as a  quasar, when it isn’t actually the object that is suffering from it but the light it has emitted.

But as a gripe, this is fair enough I’d say. Andy went on to encourage his legions of adoring readers to contribute their own pet hates, preferably with an astronomical orientation. My contribution was “decimate” which  means “to remove the tenth part” or “to reduce by ten percent”, from the Roman practice of punishing disobedient legions by killing every tenth man, but is often regrettably now used to mean “annihilate” or “obliterate”. You might think this hasn’t got much to do with astronomy but, sadly, it does. Indeed, a press release from STFC discussing the recent ten percent cuts to its grants budget states that consequent reduction in PDRAS

..will not cause the decimation of physics departments as has been speculated in media reports.

I would expect a civil servant to have done a bit better, so presumably this was written by an astronomer too. At any rate, it is precisely wrong.

You might argue that things like this don’t matter.  Language evolves,  and if modern usage deviates from its previous meanings then we should just let it change. I fully accept the dynamic nature of language and do not by any means object to all such changes. Society changes and so must the words we use. But if a change is (a) a result of sloppiness and (b) results in the loss of a very good use to be replaced by a bad one, then I think educated people should stand their ground and fight it. If we don’t do that language doesn’t just change, it decays.

Most of us practising scientists have to spend a lot of our time writing scientific papers, departmental memos, grant applications and even books. I think many astronomers see this activity as a chore, take no pleasure from it, and invest the minimum care on it. I was fortunate to have a really excellent writer, John Barrow, as my thesis supervisor and he convinced me that it was worth making the effort to write the best prose I could whatever the context. Not only does this attitude eliminate the ambiguity which is the bane of scientific writing. Taking pains over style and grammar also allows one to feel the pleasure of craftsmanship for its own sake. With John’s guidance and encouragement, I learned to enjoy writing through the satisfaction experienced by finding neat forms of words or nice turns of phrase. You never really feel good about what you do if you scrape through at the miminum acceptable level. Try to make the effort and you will be more fulfilled and the long hours of slog you spend putting together a complicated paper will at least be enlivened by a genuine sense of delight when things fall neatly into place, and a warm glow of achievement when you read it back and it sounds not just acceptable but actually good.

But I digress.

One of the other contributors to Andy’s list of examples of bad grammar was a chap called Norman Gray who objected to astronomers’ use of the word “data” as a plural noun, as in “the data indicate” rather than “the data indicates”. I was taken aback by this because I was expecting the opposite objection.

He has a lengthy rant about this on his own blog so I won’t repeat his arguments in detail here, merely a synopsis. The word “data” is formed from the latin plural of the word “datum” (itself formed from the past participle of the latin verb “dare”, meaning “to give”) hence meaning “things given” or words to that effect. The usage of “data” that we use now (to refer to measurements or quantitative information) seems not to have been present in roman or mediaeval times so Norman argues that it is a deliberate archaism to treat it as a latin plural now. He also argues that “data” in modern usage is a “mass noun” so should on that grounds also be treated as singular.

For those of you who aren’t up with such things, English nouns can be of two forms: “count” and “non-count” (or “mass”). Count nouns are those that can be enumerated and therefore have both plural and singular forms:  one eye, two eyes, etc. Non-count nouns (which is a better term than “mass nouns”) are those which describe something which is not enumerable, such as “furniture” or “cutlery”. Such things can’t be counted and they don’t have a different singular and plural forms. You can have two chairs (count noun) but can’t have two furnitures (non-count noun).

Count and non-count nouns require different grammatical treatment. You can ask “how much furniture do you have?” but not how many. The answer to a “how much” question usually requires a unit or measure word (e.g. “a vanload of furniture”) but the answer to a “how many” question would be just a number. Next time you are in a supermarket queue where it says “ten items or less” you will appreciate that it the sign is grammatically incorrect. “Item” is most definitely a count noun, so the correct form should be “ten items or fewer”..

Anyway, Norman Gray asserts that (a) “data” is a non-count noun and that (b) it should therefore be singular. Forms such as “the data are..” are out (“a vile anacoluthon”) and “the data is…” is in.

So is he right?

Not really.  Unkind though it may be to dismantle a carefully constructed obsession, I think his arguments have quite a few problems with them.

For a start, it seems clear to me that there are (at least) two distinct uses of the word data. One is clearly of non-count type. This is the use of “data” to describe an undifferentiated unspecified or unlimited quantity of information such as that stored on a computer disk. Of such stuff you might well ask “how much data do you have?” and the answer would be in some units (e.g. Gbytes). This clearly identifies it as a mass noun.

But there is another meaning, which is that ascribed to specified pieces of information either given (as per the original latin) or obtained from a measurement. Such things are precisely defined, enumerable and clearly therefore of count-noun form. Indeed one such entity could reasonably be called a datum and the plural would be data. This usage applies when the context defines the relevant quantum of information so no unit is required. This is the usage that arises in most scientific papers, as opposed to software manuals. “In Figure 1, the data are plotted…” is correct. Although it sounds clumsy you could well ask in such a situation “how many data do you have?” (meaning how many measurements do you have) and the answer would just be a number. Archaism? No. It’s just right.

To labour the point still further,  here are another two sentences that show the different uses:

“If I had less data my disk would have more free space on it.” (Non-count)

“If I had fewer data I would not be able to obtain an astrometric solution.” (Count).

Contrary to Norman’s claims, it is not unusual for the same words (if they’re nouns) to have both count and non-count forms in different contexts. I give the example of “whisky” as in “my glass is full of whisky” (non-count) versus “two whiskies, please, barman”. His objection to this was that in the second case a whisky is an artefact of a metonymic shift which takes the word “whisky” to refer to the glass containing it.

Metonymy involves using a word related to a thing rather than the word for thing itself, as in “I have hungry mouths to feed”; it’s not really the mouths that are fed, but the people the mouths belong to. In fact there’s a bit of this going on when people talk about sources being “extincted” rather than their light.

This invalidates the example because, Norman alleges, the resulting meaning is different. This objection is a bit silly because the whole point is that the two forms should have different meanings, otherwise why have them? In any case the  example  simply involves me asking for two well-defined quantities of whisky. I’m not convinced of the relevance of metonymy here. What I care about is the whisky, not what it comes in, and when I drink the whisky I don’t drink the glass anyway. Metonymy would apply if I talked about drinking a couple of glasses. Consider “I drank two whiskies, one after the other” versus “I drank two glasses one after the other”. In both cases what has actually been drunk?

There are countless other examples (pun intended). “Fire” can be a mass noun “fire is dangerous”) but also a count noun (“the firemen were fighting three fires simultaneously”). Another nice one  is “hair” which is non-count when it is on someone’s head (“my hair is going grey”) but count when  they, in the plural, are being split.

Interestingly, though, the  non-count forms of these nouns are all singular. Indeed, many non-count nouns exist only in the singular: such nouns are called singularia tantum. Examples include “dust” and “wealth”. So,  if we accept that “data” can be a non-count noun, does that mean that it should necessarily be treated as singular when it does take on that role?

An example that might be taken to support this view could be “statistics” (the field thereof) which is a non-count noun. Although it appears to be derived from a plural, you would certainly say “statistics is a hard subject”  rather than “statistics are a hard subject”.  On the other hand “statistics” can refer to a set, each element of which is a statistic (i.e. a number), thus giving another example of a noun that can be of either count or non-count form; you might reasonably say “the statistics are impressive” in the count case.  The non-count form “statistics” is a better  example of metonymy than the example above, as it refers to the study of the (count) statistics rather than to the things themselves.

In fact there are also mass nouns, described as pluralia tantum, which exist only in the plural. A (not entirely accurate) list is given here. Examples include scissors and pants, for which the normal measure  is a “pair”. Although these are technically non-count nouns (in the sense that you can’t have one scissor, etc) they don’t shed much light on the example in front of us. Perhaps more pertinent is the word “clothes” which is of non-count type but which is certainly plural. You can’t have one “clothe” (or any other number for that matter) but you would definitely say “your clothes are dirty”.

A more subtle example with relevance to the latin root of “data” is “media” which can refer to broadcast media (non-count) or plural of medium (count).  “The media are out to get me”  seems a correct construction to me, so the non-count form of this noun is a plurale tantum (singular of pluralia tantum).

So,  just because a word may be a non-count noun, it doesn’t necessarily have to be singular.

To summarise,  my argument is that (a) it is not correct to assert “data” is a mass noun. It may or may not be, depending on the context. If it is acting as a count noun (which I contend is the case in most science writing) then it is definitely plural. Furthermore, even in cases where it is clearly a mass noun, and especially if you reject the alternative meaning as a count noun, then  it is still by no means obvious that it must be treated as singular (because of the existence of the plurale tantum). In fact I would go a bit further and argue that you can only justify the singular non-count form at all if you accept that there is a count alternative. To be honest, though, I think I prefer the singular interpretation in the non-count case, as in “statistics”. It just sounds better.

If anyone has managed to read all the way through this exercise in pedantry I’d be interested to see any comments on my analysis of data.