Wednesday, February 28, 2007



Roll on the robots


This is the pre-edited version of my Materials Witness column for the April issue of Nature Materials.

Spirit, the redoubtable Martian rover, has spent the past year driving on just five of its six wheels. In February the Rover’s handling team said it had perfected the art of manoeuvring with one wheel missing, but the malfunction raises the question of whether there are better ways for robots to get around. Walking robots are becoming more efficient thanks to a better understanding of the ‘passive’ mechanism of human locomotion; but a single tumble might put such a robot out of action permanently in remote or extraterrestrial environments.

So a recent survey of rolling robots provided by Rhodri Armour and Julian Vincent of the University of Bath (J. Bionic Eng. 3, 195-208; 2006) is timely. They point out that spherical robots have several advantages: for example, they’ll never ‘fall over’, the mechanics can all be enclosed in a protective hard shell, the robot can move in any direction and can cope with collisions, uneven and soft surfaces.

But how do you make a sphere roll from the inside? Several answers have been explored in designs for spherical robots. One developed at the Politecnico di Bari in Italy aims to use an ingenious internal driver, basically a sprung rod with wheels at each end. It’s a tricky design to master, and so far only a cylindrical prototype exists. Other designs include spheres with ‘cars’ inside (the treadwheel principle), pairs of hemispherical wheels, moving internal ballast masses – the Roball made at the Université de Sherbrooke in Québec, and the Rotundus of Uppsala University in Sweden – and gyroscopic rollers like Carnegie Mellon’s Gyrover.

But Armour and Vincent suggest that one of the best designs is that in which masses inside the sphere can be moved independently along radial arms to shift the centre of gravity in any direction. The Spherobot under development at Michigan State University, and the August robot designed in Iran use this method, as does the wheel-shaped robot made at Ritsumeikan University in Kyoto, which is a deformable rubber hoop with ‘smart’ spokes that can crawl up a shallow incline and even jump into the air.

Although rolling robots clearly have a lot going for them, it might give us pause for thought that nature seems very rarely to employ rolling. There are a few organisms that make ‘intentional’ use of passive rolling, being able to adopt spherical shapes that are blown by the wind or carried along by gravity: tumbleweed is perhaps the most familiar example, but the Namib golden wheel spider cartwheels down sand dunes to escape wasps, and woodlice, when attacked, curl into balls and roll away. Active rollers are rarer still: Armour and Vincent can identify only the caterpillar of the Mother-of-Pearl moth and a species of shrimp, both of which perform somersaults.

Is this nature’s way of telling us that rolling has limited value for motion? That might be jumping to conclusions; after all, wheels are equally scarce in nature, but they serve engineering splendidly.

Tuesday, February 27, 2007

Science on Stage: two views

Carl Djerassi has struck back at Kirsten Shepherd-Barr’s rather stinging critique of his plays in a review of Kirsten’s book Science on Stage in Physics Today. I think his comments are a little unfair; Carl has his own agenda of using theatre to smuggle some science into culture, which is a defensible aim but doesn’t acknowledge that the first question must be: is this good theatre? Or as Kirsten asks, does it have ‘theatricality’? Here is my own take on her book, published in the July issue of Nature Physics last year.

Science on Stage: From Doctor Faustus to Copenhagen
Kirsten Shepherd-Barr
Princeton University Press, 2006
Cloth $29.95
ISBN 0-691-12150-8
264 pages

Over the past decade or so, science has been on stage as never before. Michael Frayn’s Copenhagen (1998), which dramatized the wartime meeting between Werner Heisenberg and Niels Bohr, is perhaps the most celebrated example; but Tom Stoppard had been exploring scientific themes for some time in Hapgood (1988) and Arcadia (1993), while Margaret Edison’s Wit (1998) and David Auburn’s Proof (2001) were both Pulitzer prize-winning Broadway hits, the latter now also a Hollywood movie. There are plenty of other examples.

While this ‘culturization’ of science has largely been welcomed by scientists – it certainly suggests that theatre has a more sophisticated relationship with science than that typified by the ‘mad scientist’ of cinematic tradition – there has been a curious lack of insightful discussion of the trend. Faced with ‘difficult’ scientific concepts, theatre critics tend to seek recourse in bland clichés about ‘mind-boggling ideas’. Scientists, meanwhile, all too often betray an artistic conservatism by revealing that their idea of theatre is an entertaining night out watching a bunch of actors behind a proscenium arch.

Thank goodness, then, for Kirsten Shepherd-Barr’s book. It represents the first sustained, serious attempt that I have seen to engage with the questions posed by science in theatre. In particular, while there has been plenty of vague talk about pedagogical opportunities, about Snow’s two cultures and about whether the ‘facts are right’, Shepherd-Barr explores what matters most about ‘science plays’: how they work (or not) as theatre.

Despite the book’s subtitle, it does not really try to offer a comprehensive historical account of science in theatre. All the same, one can hardly approach the topic without acknowledging several landmark plays of the past that have had a strong scientific content. It is arguably stretching the point to include Marlowe’s Dr Faustus (c.1594), despite its alchemical content, since this retelling of a popular folk legend is largely a morality tale which can be understood fully only in the context of its times. But while that is equally true of Ben Jonson’s The Alchemist (c.1610), both plays are important in terms of the archetypes they helped establish for the dramatic scientist: as arrogant Promethean man and as wily charlatan. There are echoes of both in the doctors of Ibsen’s plays, for example.

More significant for the modern trend is Bertolt Brecht’s Life of Galileo (1938/45) a far more nuanced look at the moral dilemmas that scientists face. Like Copenhagen, Galileo has drawn criticism from some scientists and science historians over the issue of historical accuracy. Some of these criticisms simply betray an infantile need to sustain Galileo as the heroic champion of rationalism in the face of church dogma. That is bad history too, but then, scientists are notorious (or should be) for their lack of real interest in history, as opposed to anecdote. Here Shepherd-Barr is admirably clear and patient, explaining that Copenhagen “takes history simply as material for creating theatre that does what art in general does: poses questions.”

Yet this is something scientists and historians seem to feel uncomfortable about. Writing about Copenhagen, historian Robert Marc Friedman has said “regardless of the playwright's intentions and even extreme care in creating his characters, audiences may leave the theatre with a wide range of impressions. In the case of the London production of Copenhagen on the evening that I attended, members of the audience with whom I spoke came away believing Bohr to be no better morally than Heisenberg; perhaps even less sympathetic. I am not sure, however, that this was the playwright's intention… I felt uncomfortable.” There is something chillingly Stalinist about this view of theatre and art. Should we also worry whether we have correctly divined the playwright’s “intentions” in Hamlet or King Lear?

Shepherd-Barr negotiates admirably around these lacunae between the worlds of science and art. Perhaps her key insight is that the most successful science plays are those that don’t just talk about their themes but embody them, as when the action of Arcadia reveals the thermodynamic unidirectionality of time. But most importantly, she reminds us that theatre is primarily not about words or ideas, but performance. That’s why theatre is so much stronger and more exciting a vehicle for dealing with scientific themes than film (which almost always does it miserably) or even literature. Good theatre, whatever its topic, doesn’t just engage but involves its audience: it is an experiment in which the presence of the observer is critical. Brecht pointed that out; but it is perhaps in theatre’s experimental forms, such as those pioneered by Jacques Lecoq and Peter Brook (who staged Oliver Sack’s The Man Who Mistook His Wife for a Hat in 1991) and exemplified in John Barrow and Luca Ronconi’s Infinities and Theatre de Complicite’s Mnemonic, that we see how much richer it can be than the remote, ponderous literalness of film. What could be more scientific-spirited than this experimental approach? When science has given us such extraordinary new perspectives on the world, surely theatre should be able to do more than simply show us people talking about it.
Don’t censor the state climatologists

Aware that I will no doubt be dismissed as the yes-man of the ‘climate-change consensus’ for my critique of climate sceptics in Prospect (see below), I want to say that I am dismayed at the news that two US state climatologists are being given some heat for disagreeing with the idea that global warming is predominantly anthropogenic. First, it seems that state climatologists have many concerns, of which global climate change is just one (and a relatively minor one at that). But more importantly, it is absurd to expect any scientist to determine their position by fiat so that it is aligned with state policy or any other political position. The matter is quite simple: if the feeling is that a scientist’s position on an issue undermines their credentials as a scientist, they should not be given this kind of status in the first place. If it is true that, as Mike Hopkins says in his Nature story (and Mike gets things right) “Oregon governor Ted Kulongoski said that he wants to strip Oregon's climatologist George Taylor of his title for not agreeing that global warming is predominantly caused by humans”, then Kulongoski is wrong. The only reason Taylor ought to be stripped of his title is that he has been found to be a demonstrably bad climatologist. The same with Pat Michaels at Virginia. As it happens, my impression of Michaels is that he is no longer able to be very objective on the issue of climate change – in other words, he doesn’t seem to be very trustworthy as a scientist on that score. But I’m prepared to believe that he says what he does in good faith, and of course should be allowed to argue his case. Trying to force these two guys to fall in line with the state position is simply going to fan the conspiracy theorists’ flames (I’m awaiting Benny Peiser’s inevitable take on this). But even if these paranoid sceptics did not exist, the demands would be wrong.
The more voices, the better the result in Wiki world
Here's the pre-edited version of my latest article for news@nature…

The secret to the quality of Wikipedia entries is lots of edits by lots of people

Why is Wikipedia so good? While the debate about just how good it is has been heated, the free online encyclopaedia offers a better standard of information than we might have any right to expect from a resource that absolutely anyone can write and edit.

Three groups of researchers claim now to have untangled the process by which many Wikipedia entries achieve an impressive accuracy [1-3]. They say that the best articles are those that are highly edited by many different contributors.

Listening to lots of voices rather than a few doesn't always guarantee the success that Wikipedia enjoys – just think of all those rotten movies written by committee. Collaborative product design in commerce and industry also often generates indifferent results. So why does Wiki work where others have failed?

Wikipedia was created by Jimmy Wales in January 2001, since when it has grown exponentially both in terms of the number of users and the information content. In 2005, a study of its content by Nature [4] concluded that the entries were of a comparable standing to those generated by experts for the Encyclopaedia Britannica (a claim that the EB quickly challenged).

The idea behind Wikipedia is encapsulated in writer James Surowiecki's influential book is The Wisdom of Crowds[5]: the aggregate knowledge of a wide enough group of people will always be superior to that of any single expert. In this sense, Wikipedia challenges the traditional notion that an elite of experts knows best. This democratic, open-access philosophy has been widely imitated, particularly in online resources.

At face value, it might seem obvious that the wider the community you consult, the better your information will be – that simply increases your chances of finding a real expert on Mozart or mud wrestling. But how do you know that the real experts will be motivated to contribute, and that their voices will not be drowned out or edited over by other less-informed ones?

The crucial question, say Dennis Wilkinson and Bernardo Huberman of Hewlett Packard's research laboratories in Palo Alto, California, is: how do the really good articles get to be that way? The idea behind Wikipedia is that entries are iterated to near-perfection by a succession of edits. But do edits by a (largely) unregulated crowd really make an entry better?

Right now there are around 6.4 million articles on Wikipedia, generated by over 250 million edits from 5.77 million contributors. Wilkinson and Huberman is have studied the editing statistics, and say that they don't simply follow the statistical pattern expected from a random process in which each edit is made independently of the others [1].

Instead, there are an abnormally high number of very highly edited entries. The researchers say this is just what is expected if the number of new edits to an article is proportional to the number of previous edits. In other words, edits attract more edits. The disproportionately highly edited articles, the researchers say, are those that deal with very topical issues.

And does this increased attention make them better? Yes, it does. Although the quality of an entry is not easy to assess automatically, Wilkinson and Huberman assume that those articles selected as the 'best' by the Wikipedia user community are indeed in some sense superior. These, they say, are more highly edited, and by a greater number of users, than less visible entries.

Who is making these edits, though? Some have claimed that Wikipedia articles don't truly draw on the collective wisdom of its users, but are put together mostly by a small, select elite, including the system's administrators. Wales himself has admitted that he spends "a lot of time listening to four or five hundred" top users.

Aniket Kittur of the University of California at Los Angeles and coworkers have set out to discover who really does the editing [2]. They have looked at 4.7 million pages from the English-language Wikipedia, subjected to a total of about 58 million revisions, to see who was making the changes, and how.

The results were striking. In effect, the Wiki community has mutated since 2001 from an oligarchy to a democracy. The percentage of edits made by the Wikipedia 'elite' of administrators increased steadily up to 2004, when it reached around 50 per cent. But since then it has steadily declined, and is now just 10 per cent (and falling).

Even though the edits made by this elite are generally more substantial than those made by the 'masses', their overall influence has clearly waned. Wikipedia is now dominated by users who are much more numerous than the elite but individually less active. Kittur and colleagues compare this to the rise of a powerful bourgeoisie within an oligarchic society.

This diversification of contributors is beneficial, Ofer Arazy and coworkers at the University of Alberta in Canada have found [3]. They say that, of the 42 Wikipedia entries assessed in the 2005 Nature study, the number of errors decreased as the number of different editors increased.

The main lesson for tapping effectively into the 'wisdom of the crowd', then, is that the crowd should be diverse: represented by many different views and interests. In fact, in 2004 Lu Hong and Scott Page of the University of Michigan showed that a problem-solving team selected at random from a diverse collection of individuals will usually perform better than a team made up of those who individually perform best – because the latter tend to be too similar, and so draw on too narrow a range of options [6]. For crowds, wisdom depends on variety.

Reference
1. Wilkinson, D. M. & Huberman, B. A. preprint http://xxx.arxiv.org/abs/cs.DL/0702140 (2007).
2. Kittur, A. et al. preprint (2007).
3. Arazy, O. et al. Paper presented at 16th Workshop on Information Technologies and Systems, Milwaukee, 9-10 December 2006.
4. Giles, J. Nature 438, 900-901 (2005).
5. Surowiecki, J. The Wisdom of Crowds (Random House, 2004).
6. Hong, L. & Page, S. E. Proc. Natl Acad. Sci. USA 101, 16385-16389 (2004).

Friday, February 23, 2007


The secret of Islamic patterns
This is the pre-edited version of my latest piece for news@nature. The online version acquired some small errors that may or may not be put right. But what a great paper!

Muslim artists may have used a sophisticated tiling scheme to design their geometric decorations

The complex geometrical designs used for decoration by Islamic artists in the Middle Ages, as seen in buildings such as the Alhambra palace in southern Spain, were planned using a sophisticated tiling system that enabled them to make patterns not known in the West until 20 years ago, two physicists have claimed.

By studying many Islamic designs, Peter Lu of Harvard University in Cambridge, Massachusetts, and Paul Steinhardt of Princeton University in New Jersey have decided they were put together not using a compass and ruler, as previously assumed, but by tessellating a small number of different tiles with complex shapes.

The researchers think that this technique was developed around the start of the thirteenth century, and that by the fifteenth century it had become advanced enough to generate complex patterns now known as quasiperiodic. These were 'discovered' in the 1970s by the British mathematical physicist Roger Penrose, and were later found to account for puzzling materials called quasicrystals. Discovered in 1984 in metal alloys, quasicrystals initially foxed scientists because they seemed to break the geometric rules that govern regular (crystalline) packing of atoms.

The findings provide a further illustration of how advanced Islamic mathematics was in comparison with the medieval West. From around the eleventh century, much of the understanding of science and maths in the Christian West came from Islamic sources. Arabic and Persian scholars preserved the learning of the ancient Greeks, such as Aristotle, Ptolemy and Euclid, in translations and commentaries.

The Muslim writers also made original contributions to these fields. Western scholars learnt Arabic and travelled to the East to make Latin translations of the Islamic books. Among the mathematical innovations of the Islamic world were the use of algebra, algorithms (both of which are words derived from Arabic) and the use of numerals now known as 'Arabic' (although derived in turn from Indian notation).

The mathematical complexity of Islamic decoration has long been admired. The artists used such motifs because representational art was discouraged by the Koran. “The buildings decorated this way were among the most monumental structures in the society, combining both political and religious functions”, says Lu. “There was a great interest, then, in using these structures to broadcast the power and sophistication of the controlling elite, and therefore to make the ornament and decoration equally monumental.”

Lu and Steinhardt now propose that these designs were created in a previously unsuspected way. They say that the patterns known as girih, consisting of geometric polygon and star shapes interlaced with zigzagging lines, were produced from a set of just a handful of tiling shapes ranging from pentagons and decagons (regular ten-sided polygons) to bow-ties, which can be pieced together in many different ways. The two physicists show how these tiles could themselves be drawn using geometric constructions with compasses that were known by medieval Islamic mathematicians.

Some scrolls written by Islamic artists to explain their design methods show tiles with these shapes explicitly, confirming that they were used as 'conceptual building blocks' in making the design. Lu says that they’ve found no evidence that the tiles were actually made as physical objects. “But we speculate they were”, he adds, “so as to be used as templates in laying out the actual tiling on the side of a building.”

Lu and Steinhardt say that designing this way was simpler and faster than starting with the zigzag lines themselves: packing them together in different regular arrays automatically generates the complex patterns. “Once you have the tiles, you can make complicated patterns, even quasicrystalline ones, by following a few simple rules”, says Lu.

The researchers have shown that many patterns on Islamic buildings can be built up from the girih tiles. The resulting patterns are usually periodic – they repeat again and again, and so can be perfectly superimposed on themselves when shifted by a particular distance – but this regularity can be hard to spot, compared say with that of a hexagonal honeycomb pattern.

The patterns also contain many shapes, such as polygons with 5, 10 and 12 sides, that cannot themselves be packed together periodically without leaving gaps. This property of the polygons means that scientists long believed that it was impossible for crystals to show five- ten- and twelvefold symmetries, such that rotating them by a fifth, tenth or twelfth of a full circle would allow them to be superimposed on themselves.

So when 'crystals' that appeared to have these symmetries were discovered in 1984, they seemed to violate the basic rules of geometry. But it became clear that these quasicrystals aren't perfectly periodic. In the same year, Steinhardt pointed out how patterns with the same geometric properties as quasicrystyals could be constructed from the tiling scheme devised by Penrose.

Steinhardt and Lu say that, while there is no sign that the Islamic artists knew of the Penrose tiling, their girih tiling method provides an alternative way to make the same quasicrystalline patterns. The researchers say that a design on the Darb-i-Imam shrine in Isfahan, Iran, made in 1453, is virtually equivalent to a Penrose tiling. One of the mesmerizing features of this pattern is that, like a true quasicrystal, it looks regular but never repeats exactly.

“I’d conjecture that this was quite deliberate”, says Lu. “They wanted to extend the pattern without it repeating. While they were not likely aware of the mathematical properties and consequences of the construction rule they devised, they did end up with something that would lead to what we understand today to be a quasicrystal.”

Reference
Lu, P. J. & Steinhardt, P. J. Science 315, 1106 - 1110 (2007).

Postscript
I have received some comments from Roger Penrose on this work, sadly too late for inclusion in the Nature piece but which provide some valuable perspective on the discovery. This is what he says:
"The patterns are fascinating, and very beautiful, and it is remarkable how much these ancient architects were able to anticipate concerning 5-fold quasi-symmetric organization. But, as Steinhardt (and, in effect, Lu) have confirmed directly with me, the Islamic patterns are not the same as my patterns (on several counts: different basic shapes, no matching rules, no evidence that they used anything like a "Penrose pattern" to guide them, the hierarchical structure indicated by their subdivision of large shapes into smaller ones is not strictly followed, and would not, in any case, enable the patterns to map precisely to a "Penrose tiling"). I do, however, regard this work of Steinhardt and Lu as a most intriguing and significant discovery, and one wonders what more the ancient Islamic designers may have known about such things. I should perhaps add that the great Astronomer Johannes Kepler, in his Harmonice Mundi (vol.2), published in 1619, had independently produced a regular pentagon tiling that is much closer to my own tilings than anything that I have seen so far in this admittedly wonderful Islamic work."

Peter Lu, incidentally, has indicated that he agrees with everything that Penrose says here. The relationship between the Darb-i-Imam pattern and a Penrose tiling is subtle - much more so, it seems, than media reports of this work have tended to imply.

Tuesday, February 13, 2007

When research goes PEAR-shaped

I’ve got a muse@nature.com column up today about the closure of the lab at Princeton that was investigating paranormal phenomena. Inevitably these things have to be chopped and changed before they appear, but here’s the pre-edited version. I feel scientists have no need to get too heavy about this kind of thing – if nothing else, it could serve as an interesting discussion point for students learning about how science is, and should be, done. To judge from the descriptions I’ve read of the PEAR lab and its ethos, we could probably do with a bit more of that in the scientific community. But why, oh why, do these people feel the need to come up with a ‘theory’ that is just a tangle of words? It is, in the time-honoured phrase, not even wrong. Sometimes you can’t help feeling that quantum theory has a lot to answer for.

There should be room for a bit of fringe science – but it's liable to suck you in.

It can't do a great deal for your self-esteem when media interest in your research project seems to catch fire only in response to the project's demise. But Robert Jahn and Brenda Dunne of the Princeton Engineering Anomalies Research (PEAR) laboratory probably aren't too bothered by that. For the attention generated by the closure of the PEAR lab – or rather, by the suggestion in the New York Times that this removes a source of ongoing embarrassment to the university – can surely only enhance the profile of Jahn and Dunne's longer-term vision of exploring "consciousness-related anomalies".

What "anomalies", exactly? With meticulous care, Jahn and Dunne avoid describing the phenomena they've studied using the more familiar words: telekinesis and telepathy. They have been studying people's ability to control machines and to transmit images from remote locations using only the power of the human mind. According to your perspective, that choice of language is a way of either promoting the paranormal by stealth or avoiding knee-jerk criticism.

The affair has inevitably ignited debates about the limits of academic freedom and responsibility. The NY Times quotes physicist Robert Park, a noted debunker of pseudo-science, as saying "It’s been an embarrassment to science, and I think an embarrassment for Princeton", while physicist Will Happer at Princeton says "I don’t believe in anything [Jahn] is doing, but I support his right to do it."

The university itself is trying to keep out of the fray. While stressing that the work done at PEAR was, like most other research at the university, privately funded, Princeton spokeswoman Cass Cliatt says that the lab's closure "was not a university decision". She adds that "the work at the lab was always understood by the university to be a personal interest of Professor Jahn's." Jahn, now an emeritus professor, was former dean of the engineering school and is an expert on electric propulsion.

Jahn and Dunne, a developmental psychologist, confirm that the decision was theirs. "We have accomplished what we originally set out to do 28 years ago, namely to determine whether these effects are real and to identify their major correlates", they say. With Jahn about to retire, "it is time for the next generation of scholars to take over." They hope that their work will be continued through the International Consciousness Research Laboratories, a network established in 1996 and now boasting members from 20 countries.

Some will surely share Park's view that this sort of thing gives science a bad name. But they'd be wrong to let the matter rest there, because PEAR's research reveals some interesting things about the practice and sociology of science.

The PEAR project offers a glimpse of what scientists can expect if they decide to dabble in what is conventionally termed the paranormal. Reasonable scientists cannot rule out the possibility of telekinesis, telepathy and other such 'anomalies' of the mind, simply because there are still such huge gaps in our understanding of consciousness and the brain. But most will say, again reasonably enough, that because all previous attempts to study these putative phenomena have failed to establish anything like a consistent, reproducible and unequivocal body of data, the chances of doing any serious science on the subject are minimal. As John Webster said of witchcraft in the seventeenth century, "There is no greater folly than to be very inquisitive and laborious to find out the causes of such a phenomenon as never had any existence."

In short, they regard effects like these as examples of what American chemist Irving Langmuir famously called pathological science. Experience teaches us that these things, from N-rays to cold fusion and homeopathy, are will 'o' the wispshttp://www2.blogger.com/img/gl.italic.gif: too elusive for fruitful research, and probably imaginary if not downright fraudulent.

At least, this is the standard positivist position. But perhaps a stronger reason why scientists usually steer clear of such things is that it would be professional suicide not to. In a paper called 'The PEAR Proposition'1, published in the Journal of Scientific Exploration (a journal produced by the Society for Scientific Exploration, of which Jahn and Dunne are both officers), the PEAR duo describe the hostility they experienced at Princeton when the lab was set up. They found "covert ridicule,… grudging concession of academic freedom, and… uneasiness in public discussion of the subject." Most scientists find this sort of work not outrageous but simply embarrassing.

Predictably, Jahn and Dunne found it virtually impossible to publish their findings. Their papers, many of which reported the effects of subjects' mental and emotional states on a computerized random number generator, were returned with the comment that they treated an "inappropriate topic". One journal editor said that he would consider the text only when the authors were able to transmit it telepathically.

It is no wonder, then, that those from the academic community who swim in these murky waters are older and already established in their mainstream disciplines. The 'leaders emeritus' of the Society for Scientific Exploration are Peter Sturrock and Laurence Fredrick, emeritus professors at Stanford and Virginia respectively, both with secure reputations in space physics. Not only have such people earned themselves a bit of academic slack (as well as the ability to attract funding) but they cannot simply be cold-shouldered in the way that younger researchers would be. For the same reason, Nobel laureate physicist Brian Josephson has been permitted for years to pursue research on 'mind-matter unification' at Cambridge University amid what one senses to be a mixture of unease and resignation from his colleagues.

'The PEAR Proposition' contains many poignant notes. It shows how awkwardly the habits of academia sit with discussion of the everyday world of human interactions – an unavoidable issue in this line of work. The authors' talk of the "superficial jocularities" of their lab celebrations and the "spontaneous repartee therein" evoke a deeply uncool avuncularity, while Jahn and Dunne hardly do justice to their evidently relaxed working relationship by saying that it "constituted a virtual complementarity of strategic judgment that has triangulated our operational implementation in a particularly productive fashion." It's hard to doubt that the PEAR lab, with its artwork on the walls, its parties and its stuffed animals, was a lot more fun than most research labs. That the attempts to capture this atmosphere in the language of academese are so stilted says a lot about how routinely successful this language is in stripping the research literature of its humanity.

But in the end, this fascinating document undermines itself. When Jahn and Dunne talk about "the tendency of the desired effects to hide within the underlying random data-substructures", and the way their volunteers would often produce "better scores" in their first series of tests, they echo the way that other researchers of pathological science, such as cold fusion and the 'memory of water', betrayed their lack of objectivity with talk of "good runs" and "bad runs".

And perhaps that is the real worry in looking for marginal and unreliable phenomena. Jahn and Dunne are commendably honest about the "bemusing" and "capricious" nature of their measurements, but that only adds to the impression that they decided they were engaged in a battle of wits with nature, who did her darnedest to hide the truth of the matter.

It would be a poorer world that castigates and shuns any researcher who dabbles in unorthodox or even positively weird ideas. But the PEAR experience should be sobering reading for anyone thinking of doing that: it suggests that these things suck you in. You start off with random number generators and unimpeachable experimental technique, and before long you are talking about "an ongoing two-way exchange between a primordial Source and an organizing Consciousness." You have been warned.

Reference
1. Jahn, R. G. & Dunne, B. J. J. Sci. Explor. 19, 195 - 245 (2005).

Friday, February 09, 2007

Sceptical of the sceptics

Here’s the pre-edited version of my March Lab Report column for Prospect. In the course of writing it, I found it necessary to look at some of what has been written and said by the well-known climate-change sceptics, such as those named in the article. This has been interesting. No, let me rephrase that. By a monumental effort of will, I have suppressed the fury, frustration, stupefaction and despair that their comments are apt to induce, and found a precarious way to treat them as ‘interesting’. What I mean by that is that these remarks, coming from people who are undoubtedly smart, are so ill-informed, illogical, prejudiced and emotional that it makes little sense to approach them without trying to get some perspective on what the real issues are. The comments here by Melanie Phillips are a case in point – they are so dripping with furious contempt and scorn that there can be little doubt this touches on something rather personal to her. I suspect that in many of these cases, the issue is that warnings of climate change threaten to compromise a libertarian approach to life, because they imply that there are some freedoms we enjoy now that might have to be curtailed in the future. But I’m guessing, and frankly I don’t find it a very appealing prospect to try to analyse these people.

It would be a quixotic task to try to point out all the errors in the climate-sceptic rants – that would take too long, it would achieve little, and it would be rather boring. What is most striking, however, is that very often these errors are so elementary that they show that these people actually have no interest in trying to understand climate science, or science in general, but just want to find flaws and parade them. That is why the climate-sceptic position is rather repetitive, even obsessive: you just know that they are going to reel out the ‘hockey stick’ argument, even though, first, the criticism of Michael Mann’s work is still very contentious, and second, and most significantly, it is a laughable nonsense to imply that the whole notion of global warming rests on Mann’s ‘hockey stick’. Indeed, the sceptics’ arguments always depend on the notion that we assess global warming, and the anthropogenic contribution to it, by looking at global mean surface temperatures. It must be well over ten years ago now that scientists were explaining that the tell-tale sign of human influence is to be found in the fingerprint of regional differences in the warming trend (and the fingerprint is indeed there).

All the same, I cannot resist pointing out just a few of the idiocies in some of the sceptics’ arguments. This from Phillips has nothing to do with climate change, but tells us at once that this is not someone with more than a cartoon knowledge of the history of science. In lambasting scientists who have found higher than expected methane emissions from plants, she says:
“No doubt Galileo had the same problem when all medieval parchments agreed that the sun went round the earth; or Christopher Columbus, when all navigational maps agreed that the earth was flat.”
Yes, and newspapers print this stuff.
“People say ‘the ice caps are melting’. Well, some are; but others are growing.”
Hmm… aside from the north and south polar ice caps, where are these ‘others’?
“People say ‘the seas are rising’. Well, some are, but others are falling; and where they are rising, the cause often lies in the movement of land rather than any effects of climate change.”
Plain wrong, as simple as that.
“The earth’s climate is influenced by a vastly complex series of factors which interact with each other in literally millions of ways. Computer models, which have created global warming theory, simply cannot deal with all these factors. If over-simplified material is fed into the computers, over-simplified conclusions come out at the other end.”
Melanie Phillips has decided that computer models do not do a good job of modeling the climate system? She is an expert on this? She discounts the endless model verification checks that climate modelers run? On what grounds? Will the Daily Mail let her print any statement she likes (apparently it had no qualms in permitting her to say that most of the Earth’s atmosphere is water vapour).

Nigel Lawson is an interesting case, not least because he used to control the UK’s purse strings, and so you’d like to hope this is a man with a clear head for facts. But if his reasoning on the economy was like his reasoning on climate change, that’s a truly scary thought. Here we have a marshalling of the ‘facts’ that is so selective and so distorted that you wonder just what passes for normal debate in Westminster. Oh, and the occasional lie, such as that the Royal Society tried “to prevent the funding of climate scientists who do not share its alarmist view”. (They did nothing of the sort; Bob Ward of the RS asked ExxonMobil when it intended to honour its promise to stop funding lobby groups who promote disinformation about climate change. There was no suggestion of stopping any funds to scientists.) Lawson’s comment that “the new priests are scientists (well rewarded with research grants for their pains) rather than clerics of the established religions” is about as close as I’ve seen a sceptic come to aping the stock phrases of cranks everywhere, but is also revealing in its implication that Lawson seems to find the idea of experts who know more than him offensive – a common affliction of the privileged and well educated non-scientist.

Alright, enough. I’ll start despairing again if I’m not careful. Here’s the column.
*****************************************************************

The latest report by the Intergovernmental Panel on Climate Change has come as near to blaming global warming on human activities as any scientists are likely to, while adding that its extent and consequences may be worse than we thought. The IPCC has previously been so (properly) tentative that even climate-change sceptics will have a hard time casting them as scaremongerers. So where does this leave the sceptics now?

Many politicians and scientists are hoping they will now shut up. But that’s to make the mistake of thinking this is an argument over scientific evidence.

Consider this, for instance: “As most of you have heard many times, the consensus of climate scientists believe sin global warming. Historically, the claim of consensus has been the first refuge of scoundrels; it is a way to avoid debate by claiming that the matter is already settled. Whenever you hear the consensus of scientists agrees on something or other, reach for your wallet, because you're being had. Let's be clear: the work of science has nothing whatever to do with consensus. Consensus is the business of politics. In science, consensus is irrelevant.”

This is from Michael Crichton – for the author of Jurassic Park has been giving high-level speeches about the ‘myth’ of climate change, and has even been summoned as an ‘expert witness’ on the matter by the US Senate. We need only concede that the Earth orbits the Sun and that humans are a product of Darwinian evolution to see that Crichton is not so much indulging in sophistry as merely saying something stupid. But because he is a smart fellow, stupidity can't account for it.

That's really the issue in tackling climate-change sceptics. There is no mystery about the way that some critics of the IPCC’s conclusions are simply protecting vested interests – ExxonMobil’s funding of groups that peddle climate-change disinformation, or the US government's extensive interference in federally funded climate science needs no more complex explanation than that. But this isn't true 'scepticism' – it is merely denial motivated by self-interest.

The real sceptics – strange bedfellows such as David Bellamy, Nigel Lawson, Melanie Phillips, a handful of real scientists, and Crichton – are a different phenomenon. For them there is a personal agenda involved. It’s less obvious what that might be than in, say, the comparable case of the ‘sceptics’ who denied the link between HIV and AIDS in the early 1990s. But what is immediately evident to the trained ear is that the sceptics’ denials carry the classic hallmarks of the crank – a belief that one's own reasoning betters that of professionals (even though the errors are usually elementary), a victim mentality, an instant change of tack when convincingly refuted, and (always a giveaway) a historically naive invocation of Galileo’s persecution. Of course, some of them simply tell outright lies too.

Bjorn Lomborg is a slightly different matter, since his objections focus less on denying climate change and more on denying the need to do anything about it. Nonetheless, although the economic arguments are complex, Lomborg's rhetoric – for example, suggesting that because climate change is less pressing than, say, AIDS, we should ignore it – is simplistic to a degree that again does not equate with his evident intelligence.

Economics is indeed going to be the future battleground. Yes, the argument goes, so climate change is happening, but that doesn’t mean we have to do anything to prevent it. Far better to adapt to it. This line has been pushed by heavier hitters than Lomborg, such as the eminent economists William Nordhaus at Yale and Partha Dasgupta at Cambridge, who reject the economic analysis of Nicholas Stern. The argument has some force in purely economic terms – which is perhaps not the foremost consideration if you live in coastal Bangladesh or on the Marshall islands – but it will take a lot of either faith or foolishness to let economics alone guide us into uncharted waters where we cannot rule out mass famine, decimation of biodiversity and unforeseen positive feedbacks that accelerate the warming. That’s not what economics is for.

Yet economists are right to say that we need informed rather than knee-jerk responses, and that these will surely involve compromises rather than dreaming of arresting the current trends. But by turning now to economics, however, the celebrity sceptics will only betray their agenda. It’s time to seek more reasoned voices of caution.

*****

How long before we witness the rise of the bird flu sceptic? (Matthew Parris has already staked his claim.) They could be right in one sense – according to Albert Osterhaus, chairman of the European Scientific Working group on Influenza (ESWI), “Isolated outbreaks of avian influenza in Europe are a problem in terms of economy, animal welfare and biodiversity, but the threat to public health will probably be manageable.” But they’ll almost certainly be wrong in another. The H5N1 virus is all too often portrayed as a bolt from the blue, like a bit of really rotten luck. In truth it’s illustrative of a fact of life in the viral world, where, to put it bluntly, shit happens. Last November, leading US virologists Robert Webster and Elena Govorkova stated baldly that “there is no question that there will be another influenza pandemic some day.” The ESWI agrees, and warns that Europe is ill prepared for it. Even if H5N1 doesn’t get us (by mutating into a form readily transmitted between humans), another virus will. Flu viruses are legion, and unavoidable. Here, at least, is one threat for which mitigation, not prevention, is the only option. H5N1 seems less transmissible in warmer weather, but one hopes even climate sceptics won’t see that as a point in their favour.

Friday, February 02, 2007

Space wars

I have an editorial piece on news@nature on China’s recent missile destruction of a satellite. The commentary in the scientific press has had much to say about the possible hazards of the space debris this created, but less about the implications and significance of the act for space militarization. This is my take on that.

Published online: 24 January 2007; doi:10.1038/news070122-8

muse@nature.com: A dangerous game in space
Is China's satellite zapping simply old-fashioned sabre-rattling? Or is it a rational step to restrict the use of space weapons?

How do you reconcile China's shooting down of a satellite earlier this month with the subsequent insistence by its foreign ministry spokesman, Liu Jianchao, that China opposes military competition in space?

China has not yet explained its objectives. But the action makes perfect sense in the context of game theory, the conventional framework for analysing conflict and cooperation.

Put simply, if you want to spur nations to collaborate in curbing space militarization, good intentions are not enough. You need to show that you can get tough if the need arises.

A benign interpretation of China's action, then, is that it might accomplish what years of talking have not: force the United States to negotiate an international treaty on space weaponry. Does China have such a specific goal in mind? Or does it merely wish to leave its options open in dealing with rebellious Taiwan?

These are dangerous questions. But it is worth bearing in mind that the Chinese test is at least consistent with a completely rational approach to securing international enforcement of the peaceful use of space.

The classic scenario to explore cooperation between nations using game theory is the Prisoner's Dilemma. Here, two players are each given the choice of cooperating with each other or betraying the other person (defecting), with different rewards or penalties for each potential outcome. Mutual cooperation is more beneficial to both players than is mutual defection. But temptation gets in the way: the player who defects against a cooperator wins the biggest prize of all.

Although the rational strategy in a one-off bout of the Prisoner's Dilemma is to defect, it runs against self-interest in repeated rounds. Then, the most successful way to play is often a 'tit-for-tat' strategy, in which a player will initially cooperate, then respond in kind to the other player's previous choice.

Robert Axelrod, the political scientist at the University of Michigan in Ann Arbor who pioneered the study of Prisoner's Dilemma strategies, points out that in the real world, players who follow the tit-for-tat strategy need to cultivate a reputation for toughness. Other players must know that provocation will be met with retaliation. In the case of China, the message could be that the militarization of space will not be prevented simply by condemning it, but rather by showing that you can and will play the game if necessary.

The real world is, of course, not a computer simulation, in which the agents are rational. Although game theory is studied in defence-policy circles, no one denies that it gives little more than a cartoon picture of international relations.

But in this case the model fits. China and Russia have been calling for years for a treaty to constrain space weapons. Not only have these calls been ignored by the United States, but last year the White House issued perhaps the most aggressive policy statement about space since the chilliest days of the Cold War. It stated baldly that the United States "will oppose the development of new legal regimes or other restrictions that seek to prohibit or limit US access to or use of space."

The document not only asserted the United States' right to pursue its "national interests" (including "foreign policy objectives") by preserving its "freedom of action" in space, but also threatened to deny adversaries the same freedom.

Is China an 'adversary'? Friendly overtures between NASA and the China National Space Administration might suggest otherwise, but NASA is not the Pentagon. The United States is not only still pursuing its national missile-defence programme but is also developing laser-based weapons that can knock out satellites from the ground or aircraft. It is hardly surprising then, that anyone who is serious about stopping such a relentless and defiant pursuit of space weaponry through international agreement will deploy the bullish lessons of game theory.

This is not to say that the Chinese test is defensible. It is understandable that its neighbours, such as Japan and Australia, should be dismayed by it, and that Taiwan should regard it as an act of aggression. And there is every chance that the United States will interpret it as the opening shot of an arms race rather than as a summons to the negotiating table.

China might think that keeping a strong hand relies on not making its intentions too explicit. All the same, there is a difference between developing space weapons at the same time as opposing the militarization of space, and developing weapons while refusing to ban them. Which would you prefer?

Wednesday, January 24, 2007

Nanotech rolls on

There's a nice little paper about to appear online in Nature Nanotechnology (advance online publication, doi:10.1038/nnano.2006.210). Christian Joachim and his colleagues have made a molecule consisting of a two wheels 8 Å wide on an axle. The wheels are in fact three-bladed paddlewheels, which may rotate by a third of a turn when pushed with the tip of a scanning tunnelling microscope. This sounds rather like Jim Tour's nanocar, in which the wheels are fullerenes. That's altogether a more sophisticated contraption: a four-wheeler with a full chassis, shown in the picture above. But Jim could only infer rolling from the nature of the path that the cars took when pulled with the STM. It looked pretty persuasive to me, especially as tripod cars could only go in circles. But Christian's work has more direct proof of rolling. By looking closely at the way the two-wheeled molecule moves, he and his colleagues can distinguish between simple hopping between the grooves of the atomic corrugations on the metal surface, smooth pushing when the corrugation is very shallow, and jumps due to rotation of one wheel by 120 degrees. Only one wheel turns at a time, so in that case the molecule pivots as it is pushed.

Sunday, January 21, 2007

Water in biology

I have started a new blog called Water in Biology, which does what it says on the can. My hope is that it will act as a forum for presenting and discussing new research on the roles of water in the living cell. It seems that barely a week passes without something new and interesting coming out in this area, and I’m sure I’m seeing only a fraction of it. This will hopefully be a way of keeping track of what is out there, and hearing some informed opinions about it. It’s a controversial area, and I’m well aware that there are risks in setting this up: the behaviour of water in biological systems has been something of a magnet for weird and often not so wonderful theories. I hope to keep them at a distance. No doubt it will seem terribly elitist to the likes of the homeopathy enthusiast who expressed such disappointment on amazon.com about my book on Paracelsus, but I’d respectfully suggest that if homeopathy, Jacques Benveniste’s digital biology, or Masaru Emoto are your bag, this site might not be for you. (I’m sure there are lots of others that are.)

Wednesday, January 17, 2007

Dark matter, memories and chimeras
[This is my Lab Report column for the forthcoming (February) issue of Prospect.]

Cosmic bones
The universe has had its first X-ray, and the skeleton looks much as expected. It is made of mysterious dark matter, which accounts for five sixths of the matter in the universe. It has been mapped out in a project called the Cosmic Evolution Survey (Cosmos), which pieced together more than 500 images from the Hubble space telescope to trace out the fine filaments of dark matter threading through deep space. These act like flypaper for ordinary matter, concentrating it into stars and galaxies.

In effect, Cosmos has revealed the hidden framework behind the entire visible world. Among other things, it’s another reason to be thankful that Nasa was persuaded not to leave the ageing Hubble to decay, rather than risk another shuttle-borne service mission.

Dark matter can’t be seen directly—that’s why it is dark—and isn’t made from any known subatomic particles. But we know it’s there because of the gravitational tug it exerts on visible matter. It was first postulated to explain why galaxies don’t fly apart as they rotate—but not everyone was convinced it existed until a study last year revealed it apparently balancing an otherwise lopsided spread of visible matter in a collision between two galaxies.

The Cosmos team mapped the distribution of dark matter over a big chunk of the universe by looking at the way it bends light from very distant galaxies behind it. The densest clumps of dark matter, where filaments intersect, mostly match the positions of visible galaxies and gas, as they should. But there are a few regions of dark matter without visible matter and vice versa—bones without flesh, and flesh without bones, so to speak. That wasn’t predicted and can’t easily be explained. Might there be entire galaxies made only of dark matter? It seems more likely that the discrepancies are just errors in the data, which depends on a tour-de-force of astronomical measurement.

It remains something of a scandal that we know nearly nothing about the stuff these cosmic bones are made from (our ignorance of the “dark energy” that outweighs dark matter by a factor of three is even more profound). But that’s modern cosmology for you: the harder you look, the more puzzling it gets.

Forgetful computers
Another well-hidden scandal lies in the guts of modern computers. Imagine that your piano had to be retuned from scratch every time you wanted to play, or your television set had to be retuned to the broadcast frequencies whenever it was switched on. This is the way computers work: you turn on and wait for minutes as the working memory, or random access memory (RAM), relearns how to work by reading the “manual” held on the hard disk. Actually, it’s even worse. The RAM then forgets everything you tell it within a microsecond—or would do if its memory cells were not refreshed thousands of times a second. That’s why, if the power turns off unexpectedly, all your unsaved data is lost.
It’s also why the RAM sucks all the juice from laptop batteries within hours.

All this could be changed with a technology that is already commercially available but still too expensive to go into your computer. Non-volatile RAM is a memory that, once loaded with data, holds onto it indefinitely without needing any power. With such a memory, a computer wouldn’t need a hard disk at all, and would be ready to use the moment you switched on. Various forms of non-volatile RAM are being developed, the most advanced being ferroelectric RAM (FeRAM), in which the memory elements are a little like magnets with poles oriented by electric rather than magnetic fields.

FeRAM is big in east Asia—the leaders include Samsung, Matsushita and Fujitsu. There’s even a small FeRAM in the Sony Playstation 2. But big, computer-style memories remain too expensive to make this way, and so FeRAM is currently relegated to low-tech applications such as smart cards for Japanese railways. They can be read from six inches away, however, so you don’t even need to get the card out of your bag.

Dr Moreau visions
The UK Human Fertilisation and Embryo Authority recently concluded that “there needs to be a full and proper public debate and consultation” about whether it should licence the creation of animal-human chimeric cells, for example with human DNA housed by nonhuman cells. The HFEA wisely avoided a snap decision in response to the opposition to such cells in a government white paper. The issue, as it says, is “far from black and white.” But the white paper’s recommendations seem to suffer from visions of Dr Moreau. Xenotransplants of animal organs and tissues into human patients have been happening for years, but the popular idea, peddled by the human genome project, that genes represent the sacred core of humanity has apparently created a sense that they are untouchable.

While there are ethical difficulties with chimeric cells, the medical benefits could be huge. Stem cells can be cultivated from embryos, but human eggs are hard to acquire for this. Animal eggs carrying human DNA would be more readily available, and could usher in all kinds of stem-cell therapies. The stem cells wouldn’t be “half-animal” in any meaningful sense, any more than a molecule taken from a human cell is different from an identical one taken from an animal. The HFEA aims to complete its consultation in the autumn. UK legislation in this area has been sensible and permissive so far, but stem-cell scientists are nervous that this time the yuck factor is about to kick in.

Tuesday, January 16, 2007

You couldn't make it up

"School chemistry ban for teen tartan terrorist". There I was all ready to sink my teeth into this great headline from the Scottish tabloid the Daily Record, only to discover that they made it up.

The story is that the "tartan terrorist" 16-year-old Jamie Hoggan, who has been placed in Polmont young offenders' unit for firebombing council offices in Alloa last August, was being denied chemistry and physics lessons while being sent his homework by his school, the Alloa Academy. Young Jamie apparently wants to "drive the English out of Scotland" (I know, but this is a 16-year-old lad who has presumably watched Braveheart thinking it was history). According to the Daily Record, a "source" said that "science was out of the question while he was in Polmont because of what went before." The notion that one can't make bombs without GCSE chemistry, and that one becomes an expert bomb-maker with it, was so silly that this seemed to demand some comment. (And what was the problem with physics? That he'd nuke the council instead?)

But it was all fabrication. A tired-sounding head at the Academy told me the story was "complete rubbish", before referring me to a nice woman at Corporate Affairs who said she only held back from writing to the Record to, er, set the record straight, by the understandable fear that the newspaper would then run the headline "School chemistry lessons for teen tartan terrorist". A victory for common sense, then, and a big boo to silly tabloids. And Jamie apparently hopes to become a forensic scientist – good luck to him!

Friday, December 22, 2006

Looking for ways to fill up your Christmas holiday?

Then why not heed these shameless plugs for my friends? John Whitfield’s book In The Beat of a Heart is a fabulous read – one of the best science books of 2006, though it’s not received the UK push it deserves. If this doesn’t get shortlisted for next year’s Aventis Prize, there’s no justice. It explains the notion of metabolic scaling in biology – the idea that one can develop a unified view of life by considering the way energy requirements depend on size. Like Nick Lane’s Power, Sex, Suicide, it’s a timely reminder that there’s more to life than genes.

And if you’re in the Lincoln area, pay a visit to Lindsay Seers’ exhibition The Truth Was Always There, which explores connections between magic and alchemy, Lincoln cathedral, Robert Grosseteste, Temple Bruer and John Dee. This is part of an ongoing project by Lindsay, comprising five ‘biographical’ films that document a strange past and an obsession with becoming a human camera. There’s a forthcoming book, called (I think) Human Camera, that records these films and narratives. Yes, I admit that I’m a contributor – but that’s because I think it is a wonderful project. The exhibition runs until late January.
Pushing protons around

[This is the pre-edited version of my Crucible column for the February issue of Chemistry World.]

Life is pretty simple, when you come down to it. It’s a matter of shovelling stuff from one side of a wall to the other – the ‘stuff’ being hydrogen ions, and the wall a cell membrane. The biochemistry that follows from this is fearsome, but at root life is driven by piling up hydrogen ions and then letting them flow, like water released from a dam.

This imbalance of protons across a membrane creates a so-called protonmotive force. It is generated by proton pumps: proteins that can actively move protons ‘uphill, against a concentration gradient. They need energy to do that, and in the light-harvesting chloroplasts of plants that comes ultimately from sunlight, which sets an electron jumping between molecules. In our mitochondria the energy is generated by reactions that break down carbohydrates. In either case, the protonmotive force is used to power the enzyme ATP synthase, which rotates like a water wheel as it lets protons flow through, producing energy-rich ATP in the process.

So if that’s life in a nutshell, these proton pumps clearly need to be efficient and smooth-running pieces of molecular machinery. Even so, the ingenuity life displays in conducting and controlling the movement of protons is breathtaking.

That life exists in water is a boon from the outset – because one of the things water does that other liquids cannot is transport protons rapidly. The hydrogen ion travels faster than other small cations in water by hopping along hydrogen-bonded chains of water molecules: rather like a Newton’s cradle, a proton hits one end of the chain and almost at once (figuratively speaking) another proton pops off the other end. This hopping, called the Grotthuss mechanism after the nineteenth-century German scientist who proposed the basic idea, is exploited by biomolecules to shift protons. Some proteins, such as the light-powered bacterial proton pump bacteriorhodopsin and some cytochromes, are threaded by ‘water wires’, strings of water molecules that act as proton-conducting pathways.

A water wire also winds through the membrane protein aquaporin, which transports water across cell walls. But for aquaporin, letting protons through could be disastrous, as it would disrupt the delicate balance of pH and charge across the membrane. So it has to achieve the seemingly impossible feat of transporting water but not hydrogen ions. How it does so is still not fully clear, but one idea is that the water wire contains a defect: hydrogen-bonding to the amino-acid residues within the pore forces two waters in the chain to sit ‘back to back’, so that a proton can’t jump between them.

That would be an extraordinarily delicate feat of molecular manipulation. But it is possibly trumped by the latest revelation about why proton pumping works so well. Magnus Brändén of Stockholm University and his colleagues (Proc. Natl Acad. Sci. USA, doi:10.1073/pnas.0605909103) say that there are, in effect, little proton circuits written onto the surfaces of cell membranes that help guide protons from a transporter – a pump protein – to molecules that exploit the protonmotive force, such as ATP synthase. The image, then, is not that of a pump spouting out protons into the cytoplasm, where some gradually drift over to where they’re needed; instead, the protons pop out of the pump’s mouth and stick to the membrane before proceeding to hop across it. That way, fewer get lost.

In effect, then, the membrane lipids act as proton-collecting antennas – rather as accessory pigments serve as light-harvesting antennas to shunt light energy onto the photosynthetic reaction centre in photosynthesis.

This idea has been mooted for years, but Brändén and colleagues have pinned it down by looking at the protonation of a single fluorescein dye molecule embedded in the wall of liposomes (closed, cell-like assemblies of lipids). Protonation changes the dye’s fluorescence, and so fluctuations in its brightness can be related to the rate of proton exchange with the surroundings. The researchers show that this happens at a faster rate than would be expected if protons were just being exchanged with the water – so long as the lipid head groups can themselves be protonated. The lipids gather protons and pass them around.

It’s a reminder that molecular biology isn’t just about the cleverness of proteins and nucleic acids. Even the molecules often assumed to be just part of the background or the scaffolding, such as lipids and water, may have inventive roles to play.

Thursday, December 21, 2006

Mining the moon for all it’s worth

It's one of the curious characteristics of space exploration that the usual stringent hurdles for science news stories are nowhere to be seen. Whereas normally science reporters, enthusing breathlessly about new insights into, say, the origin of the universe, face an editor's dismissive "why do we care?", with space stories it is enough that someone did something or other up in space – hit a gold ball, say, or flew a shuttle mission without blowing up. True, if the story involves an unmanned spacecraft doing something like landing on a comet, then there might be some sales work to be done. But if there's a human inside, it doesn't much matter what he's up to – just stick it on the page.

That principle even extends to stuff that hasn't actually been done, but "will" be. We're going to send people to Mars! We're going to build a moon base! Yes, it's true that we've not even managed to finish off a useless piece of manned space junk floating inanely in Earth orbit, but look, we've learnt from our mistakes. (What we've learnt, it seems, is simply to make the claims bolder.) When something is going to be done in space, all critical faculties vanish. We're going to grow crystals there, or plants – and no matter whether that's really going to tell us anything worthwhile. And now the latest fad for which the media seems determined to fall hook, line and sinker is mining the moon for helium-3.

Helium-3, you see, is a wonderful clean fuel that will power our planet through nuclear fusion. Just a shuttle-load will power the USA for a year. And it's just waiting up there in the lunar soil for us to go and collect it.

Well actually, it isn't – we'd have to strip-mine vast areas of the moon to get at it. And while discussions of remote oil resources on Earth routinely have to run the gauntlet of hard-headed economical cost-benefit analysis, no one seems to care very much whether there is any justification for thinking that sending trucks to the moon to pick up this 'fuel' is really going to save anyone any money. All you have to do is talk about lunar helium-3 as a "cash crop".

And on top of that, there's the small difficulty that no one has ever produced energy from nuclear fusion in a commercially viable and sustainable way, and that even the most optimistic estimates put that goal 50 years distant. There's a good chance that it might not happen before 2100 (although making technological projections that far ahead is a bit pointless anyway).

The point is, of course, that these bogus utilitarian rationales are routinely trotted out in defence of a programme of human space exploration that is at root ideological. It's fine to believe that there is an intrinsic value in sending humans to other worlds (I happen to disagree, at least in the present state of affairs in space travel); but let's at least be honest about it. The magic word "resources" was invoked the last time I questioned the value of manned spaceflight (see here and here). But as I'd expected, when it came down to it what that meant was stuff like silicon – as though we are currently facing a silicon shortage here on our rocky planet.

So please – no more helium-3 as the justification for a moon base. That’s truly grasping at straws.

Monday, December 18, 2006


Treacherous beauty

[This is a review of a book on the zoological art of Ernst Haeckel, to be published in Nature.]

Visions of Nature: The Art and Science of Ernst Haeckel
Olaf Breidbach
Prestel, Munich, 2006

When Nature’s millennial issue of 1900 listed the most important scientists of that age, there was only one German biologist among them: Ernst Haeckel, professor of zoology at the University of Jena. Reckoned to have been instrumental to the introduction of Darwinism in Germany, and responsible now for inspiring generations of scientists with his stunning drawings of the natural world, Haeckel still retains a claim to such recognition. He is perhaps most widely known now as the author and illustrator of Art Forms in Nature, a series of plates published between 1899 and 1904 that showed the marvellous forms and symmetries of creatures ranging from radiolarians to antelopes.

But few scientists of his time are more complicated. He was the archetypal German Romantic, who toyed with the idea of becoming a landscape painter, venerated Goethe, and was prone to a kind of Hegelian historical determinism that sat uncomfortably with Darwin’s pragmatic rule of contingency. Haeckel’s view of evolution was a search for order, systematization and hierarchy that would reveal far more logic and purpose in life than a mere struggle for survival. His most famous scientific theory, the so-called biogenetic law which argued that organisms retread evolutionary history as they develop from an egg (‘ontogeny recapitulates phylogeny’), was an attempt to extract such a unifying scheme from the natural world.

It can be argued that this kind of visionary mindset, when it creates strong preconceptions about how the world ought to be, does not serve science well. Haeckel supplies a case study in the collision between Romanticism and science, and that tension is played out in his illustrative work. Olaf Breidbach’s text to this lovingly produced volume never really gets to grips with that. It has a curiously nineteenth-century flavour itself, declining to grapple with the difficult aspects of Haeckel’s life and work.

Here, for instance, is a proposition: Ernst Haeckel’s influence on fin-du-siècle German culture was pernicious in its promotion of a ‘scientific’ racist ideology that fed directly into Nazism. That case has been made (by historian Daniel Gasman in particular), and while it can be debated, Breidbach goes no further than to admit that Haeckel became a ‘biological chauvinist’ during the First World War, and that ‘sometimes the tone of his writing was overtly racist.’ Breidbach admits that this is not a biography as such, but an examination of Haeckel’s visual heritage. Yet one could argue that Haeckel’s dark side was as much a natural consequence of his world view as was Art Forms in Nature.

The claim that Haeckel doctored images in order to make them fit with a preconceived notion of how biology works is harder to ignore in this context. Even in his own time he was accused of that (particularly by his rival Wilhelm His), and to my eye the evidence (see Nature 410, 144; 2001 and Science 277, 1435; 1997) looks pretty strong. But Breidbach skates over this issue, alluding to the allegations only to suggest that the illustrations ‘instructed the reader how to interpret the shapes of nature properly’. Well, indeed.

On the whole, though, Breidbach simply explains Haeckel’s reliance on image without assessing it. Haeckel’s extraordinary drawings were not made to support his arguments about evolution and morphogenesis; rather, they were the arguments themselves. He believed that these truths should be apparent not by analysing the images but simply by looking at them. ‘Seeing was understanding’, as Breidbach says. If that’s so, it places an immense burden of responsibility on the veracity of the images.

This is the nub of the matter. Breidbach suggests that Haeckel’s drawings are schematic and that, like any illustrator, Haeckel prepared them to emphasize what we are meant to see. But of course this means ‘what Haeckel has decided we should see’. Quite aside from whether he hid nascent appendages that challenged his biogenetic law, consider what this implies for the plates of Art Forms in Nature. They are some of the most beautiful illustrations ever made in natural history – but it seems clear now that Haeckel idealized, abstracted and arranged the elements in such a way that their symmetry and order was exaggerated. They are pictures of Platonic creatures, the ideal forms that Haeckel intuited as he gazed into his microscope. Their very beauty betrays them. They are, as Breidbach says (but seemingly without critical intent), ‘nature properly organized.’ In this way ‘the labour of the analyst was replaced by the fascination of the image’. Absolutely – for ‘fascinate’ used to mean ‘bewitch’.

It is not as if Haeckel did not have the alternative of photography – for microphotography was used as early as the 1850s. Breidbach recapitulates the arguments against an over-reliance on the veracity of photography (these ideas have been much discussed by cultural critics such as Vilém Flusser), pointing out that what one sees is determined by the technology. That is true, and it is apt to give photography a false authority. But are hand-drawn images really any better – let alone those rendered with such apparent skill and realism that their schematic nature is disguised? Indeed, Haeckel felt compelled in 1913 to publish Nature as an Artist, a series of photographs of his subjects which, he said, demonstrates that ‘there can be no talk of reconstruction, touching up, schematisation or indeed forgery’ in his drawings of the same. It was a remarkable work in its own right, but leaves us wondering why Haeckel did not use photos in the first place.

Another danger of such imagery is that it is prone to reflect the artistic styles of the day. Haeckel’s drawings fed into the florid, nature-inspired designs of the Art Nouveau and Jugendstil schools – but he was more influenced than influence. His medusae look like William Morris prints precisely because they have had that visual aesthetic imprinted on them. Breidbach says that for Haeckel, as for Goethe, ‘aesthetics is the foundation of his view of nature.’ But is that a good thing? As Ernst Gombrich has pointed out, artistic styles create unconscious biases and errors: when Gombrich speaks of the artist who ‘begins not with his visual impression but with his idea or concept’, it might as well be Haeckel he is talking of. And what happens when the cultural aesthetic moves on – does nature have to follow suit? Breidbach points out that, by using the visual language of his age, Haeckel helped to make science accessible to the public. But 20 years later, modernism had rendered his arabesque style old-fashioned.

As director of the Ernst Haeckel Museum at Jena, Breidbach has unequalled access to Haeckel’s notes and sketchbooks, and he makes good use of them. But perhaps for that very reason he felt unable to dig too deeply into the problematic areas his subject raises. (Haeckel is clearly still very much a legend at Jena, where his brain was cast in silver.) So although this is undoubtedly a gorgeous book, and the questions it raises are fascinating, I can’t help feeling that it represents an opportunity missed.

Thursday, December 14, 2006

Tainted by association?
[This is the pre-edited version of my latest muse column for news@nature.]

Richard Doll's links with industry are disconcerting but hardly scandalous. And they don't make him a villain.

Few things will polarize opinion like the defamation of a recently deceased and revered figure. So the tone of the debate (here and here and here and here) that has followed the accusation that Sir Richard Doll, the British epidemiologist credited with identifying the link between smoking and lung cancer in the 1950s, compromised the integrity of his research by receiving consultancy payments from the chemicals industry, should surprise no one.

On the one hand, the disclosure of Doll's contracts with the likes of Monsanto and Dow Chemicals have provoked howls of outrage and accusations that his studies of purported links between the companies' products and cancer were nothing less than a cover-up. Much of this is crude conspiracy-theorizing; but there are also more weighty critics. Andrew Watterson, a professor of occupational and environmental health at the University of Stirling in Scotland, has said that "Doll's work… has limited the capacity of the UK over decades to take action on occupational and environmental carcinogens as quickly as it should have… His lack of transparency on and financial relationship with companies have seriously damaged the credibility of aspects of his research."

On the other hand, voices that will be no doubt dismissed by Doll's attackers as those of the 'establishment' have risen to defend his reputation. An editorial in the Times calls the charges "a cheap shot" made by "grave robbers". This, the newspaper intones gravely, is "a sad act of character assassination by people who should know better." Several leading UK scientists have written to the Times saying that "we feel it is our duty to defend Sir Richard's reputation and to recognise his extraordinary contribution to global health."

The situation is perhaps best exemplified by a leader headline in the Observer newspaper: "Richard Doll was a hero, not a villain." All of which brings to mind the comment of Brecht's Galileo: "unhappy is the land that needs a hero."

Unhappy we are indeed, if we cannot accommodate in our pantheons the complexities of real people. And while the collaborations of academe and industry certainly create tensions and problems, it achieves nothing to pretend that they ought not to exist.

Doll's consultancy work is not immune to criticism even by the standards of his time. But the suggestion that his research is invalidated, and his character besmirched, by such conflicts of interest (as they would now certainly be regarded) is one that smells of piety rather than an evaluation of the facts.

Here, then, is the case for the prosecution. Doll proclaimed that Monsanto's Agent Orange posed no carcinogenic hazard while receiving consultancy fees of $1000-1500 a day from the company for nearly 30 years. He compiled a review on behalf of ICA, Dow Chemicals and the Chemical Manufacturers' Association in their defence against claims of cancer induced in workers by exposure to vinyl chloride, for which he was paid £15,000. (Monsanto was also a big producer of vinyl chloride). And he argued that there was little basis for the idea that asbestos is a major health risk, while pursuing a long-term consultancy relationship with the UK's leading asbestos manufacturer Turner & Newall, which later donated £50,000 to set up Green College in Oxford, a medical school of which Doll was a founder and the first warden.

All of this would indict any researcher today who failed to declare such conflicts of interest. Doll was in fact rather inconsistent about such declarations – he made no secret of some of his links to industry, but the Monsanto connection was not disclosed until a court case over vinyl chloride in 2000. In any event, until the 1980s there was no expectation that academics should make this sort of paid work public, and so no reason to expect Doll to have been systematic about doing so.

That is one of the main lines of defence for Doll's supporters: it is absurd to judge him by today's standards, a notorious way of vilifying historical figures and events. It's a fair point, although we have to remember that we're talking here about the 1980s, not the nineteenth century. It doesn't take a great deal of insight to see that being paid by a company while assessing their products is not ideal.

Yet there is no obvious reason to regard this as venal or cynical on Doll's part. Using contract money to help set up a college does not seem particularly blameworthy. Doll donated other fees to charities such as the Medical Foundation for the Victims of Torture.

He gives every impression of being a man conducting his business in an environment that had not thought very hard about the proprieties of industrial research contracts. If he did not think too hard about it either, that does not make him a villain.

And he seems very much a man who knew his own mind. Overcoming industry's resistance to the link between smoking and cancer is hardly the act of someone in the pocket of corporations. Yet his scepticism about the effects of secondary smoking speaks of a man who was not turned into an ideologue by his conclusions.

On this count, we must remember that no evidence has been presented that Doll's conclusions were biased by his contracts. It's hard to see how that could be established either way; but certainly it is unfair to turn Doll into a yes-man bribed by industry.

Indeed, if anything, the affair has served to remind us how manipulative these industries could be. Peto says that Doll came under pressure (which he resisted) from the asbestos manufacturers not to publish any evidence of the harmful effects of their product. They claimed it would damage the national interest by undermining this important industry, and even threatened legal action. The Asbestosis Research Council, founded by Turner & Newall and others in 1957, has been accused of suppressing evidence of the dangers of asbestos, for example by vetting and censoring research on the topic [1].

A curious aspect of this whole business, not mentioned at all in media reports, is that it is all yesterday's news anyway. Doll's links with industry were reported by British newspapers after being discussed in an article published online on 3 November by the American Journal of Industrial Medicine [2]. But the information in that article on Doll's connections with Monsanto, Dow, Turner & Newall and others had all been documented in 2002 by one of the authors, Martin Walker [3]. In that latter paper, Walker states that Doll "has never made any secret of the fact that he has been funded by industry for specific research projects."

Quite aside from demonstrating the media's ability to generate its own content, this fact is notable because the American Journal of Industrial Medicine paper aims not to denigrate Doll but to call for a tightening of policies governing disclosures of interest today. There's still plenty of work to be done (see here and here) in that respect. We should recognise the shortcomings of the past, and move on.

References
1. Tweedale, G. Am. J. Ind. Med. 38, 723-734 (2000).
2. Hardell, L. et al. Am. J. Ind. Med. advance online publication (2006).
3. Walker, M. J., http://www.dipmat.unipg.it/~mamone/sci-dem/contri/walker.htm.

Monday, December 04, 2006


Looking for Turing’s fingerprints

Here is the pre-edited version of my Crucible column for the January 2007 issue of Chemistry World.There’s considerably more on this issue in the forthcoming reworking of my 1999 book The Self-Made Tapestry. OUP will publish this as a series of three books, probably beginning in late 2007.

How did the leopard get its spots? Recent research supports an idea first suggested by legendary code-breaker Alan Turing.

After another long time, what with standing half in the shade and half out of it, and what with the slippery-slidy shadows of the trees falling on them, the Giraffe grew blotchy, and the Zebra grew stripy, and the Eland and the Koodoo grew darker, with little wavy grey lines on their backs like bark on a tree trunk; and so, though you could hear them and smell them, you could very seldom see them, and then only when you knew precisely where to look.

Kipling’s Just So story of how the animals of Africa obtained their distinctive marking patterns is a fine example of Lamarckism – the inheritance of environmentally acquired characteristics, in this case via what seems to be a kind of tanning process imprinted with the shadows of trees. But the explanation that his contemporary biologists would have offered, invoking Darwinian adaptation (the markings being assumed to serve as camouflage), was arguably little more than a Just So story too. It explained why a marking pattern, once acquired, would spread and persist in a population, but it could say nothing about how such a pattern came to be, either in evolutionary terms or during the embryonic development of a particular zebra, giraffe or koodoo.

It seems clear that this is not merely a genetic painting-by-numbers: the markings on two animals of the same species are recognizably alike, but not identical. How do the melanin pigments of animal pelts get distributed across the embryonic epidermis in these characteristically blotchy ways?

The favoured explanation today invokes a mechanism proposed in 1952 by the British mathematician Alan Turing, two years before his suicide by cyanide. Turing is best known for his work on artificial intelligence and the concept of a programmable computer, and for his wartime code-cracking at Bletchley Park. But his paper ‘The chemical basis of morphogenesis’ was something else entirely. It was an attempt to explain how the development of a body plan kicks off.

The fundamental question is how a spherical ball of cells ends up as a shape in which different cells, tissues and appendages are assigned to different locations. How does the initial spherical symmetry get broken? Turing’s comment reminds one of the old joke about physicists over-simplifying biology: “a system which has spherical symmetry, and whose state is changing because of chemical reactions and diffusion, will remain spherically symmetrical for ever… It certainly cannot result in an organism such as a horse, which is not spherically symmetrical.”

Turing proposed a set of differential equations which explained how molecules determining cell fates, called morphogens, might diffuse through a spherical body and induce (bio)chemical processes. Such a scheme is now called a reaction-diffusion system. The patterning results from competition between an autocatalytic reaction, which amplifies random chemical inhomogeneities, and diffusion, which smoothes them out. Turing’s calculations performed by hand (his notion of a digital computer was then barely realized) showed that patchiness could emerge. But it wasn’t until 1972 that Hans Meinhardt and Alfred Gierer in Germany clarified the essential ingredients of Turing’s model. The chemical patterns arise in the presence of an autocatalytic ‘activator’ molecule, and an inhibitor molecule that suppresses the activator. If the inhibitor diffuses more rapidly than the activator, then the concentration of activator is enhanced over short distances but lowered by the inhibitor over longer distances. This gives rise to islands of activator surrounded by regions where it is suppressed.

Calculations showed that this ‘activator-inhibitor’ mechanism could create orderly patterns of spots and stripes, more or less equally sized and equidistant. That brought to mind the leopard’s spots and the zebra’s stripes. But is such a system anything more than a pretty mathematical fiction? That wasn’t clear until 1990, when a team led by Patrick De Kepper at the University of Bordeaux identified the first chemical Turing pattern, using a reaction that, when well mixed, oscillated between yellow and blue states. This was closely related to the Belousov-Zhabotinsky (BZ) reaction, which was known since the 1960s to generate travelling chemical waves. The BZ reaction is a reaction-diffusion system, but does not itself produce stationary Turing patterns, because the relative diffusion rates of the reactants don’t fit Turing’s model.

Are patterns in the living world really made this way? Theoretical activator-inhibitor systems have now been able to provide very convincing mimics of a wide range of animal markings, from the reticulated mesh of the giraffe’s pelt to the crescent-shaped rosettes of the leopard and jaguar, the spots of the ladybird and the stripes of the zebrafish. All this looks plausible enough, but the clinching proof – the identification of diffusing morphogens responsible for pigmentation – has yet to be obtained.

There does now seem to be good evidence that chemical morphogens of the sort Turing envisaged act in biological development. Several proteins belonging to the class called transforming growth factor b proteins seem to act this way in fly and vertebrate morphogenesis, signalling the developmental fate of cells as they diffuse through the embryo.

But does anything like Turing’s mechanism act to shape embryos beyond the question of marking patterns, as Turing suggested? In general, embryogenesis seems more complex than that, operating under close genetic control. However, Thomas Schlake and colleagues at the Max Planck Institute of Immunobiology in Freiburg have very recently discovered that there is at least one other biological patterning process that apparently uses the activator-inhibitor mechanism. They have found that the follicles of mouse hair are positioned in the epidermis by the protein products of two classes of gene, called WNT and DKK. The former appears to take the role of activator, inducing follicle formation, while several variants of DKK proteins act as inhibitors1.

Hair and feather positioning has long been suspected as an example of a Turing pattern – the equidistant, roughly hexagonally packed patterns are just what would be expected. Schlake and colleagues have made that case by looking at how over-expression of WNT and DKK alters the follicle patterns on mutant mice, showing that these match the predictions based on an activator-inhibitor model. It is probably the best reason yet to think that Turing’s intuition was sound.


Reference

1. Sick, S. et al. Science Express doi:10.1126/science.1130088.