Thursday, November 29, 2007

Why 'Never Let Me Go' isn't really a 'science novel'

I have just finished reading Kazuo Ishiguro’s Never Let Me Go. What a strange book. First, there’s the tone – purposely amateurish writing (there can’t be any doubt, given his earlier books, that this is intentional), which creates an odd sense of flatness. As the Telegraph’s reviewer put it, “There is no aesthetic thrill to be had from the sentences – except that of a writer getting the desired dreary effect exactly right.” It’s a testament to Ishiguro that his control of this voice never slips, and that the story remains compelling in spite of the deliberately clumsy prose. That s probably a far harder trick to pull off than it seems. Second, there are the trademark bits of childlike quasi-surrealism, where he develops an idea that seems utterly implausible yet is presented so deadpan that you start to think “Is he serious about this?” – for instance, Tommy’s theory about the ‘art gallery’. This sort of dreamlike riffing was put to wonderful effect in The Unconsoled, which was a dream world from start to finish. It jarred a little at the end of When We Were Orphans, because it didn’t quite fit with the rest of the book – but was still strangely compelling. Here it seems to be an expression of the enforced naivety of the characters, but is disorientating when it becomes so utterly a part of the world that Kathy H depicts.

But my biggest concern is that the plot just doesn’t seem at all plausible enough to create a strong critique of cloning and related biotechnologies. Is that even the intention? I’m still unsure, as were several reviewers. The situation of the donor children is so unethical and so deeply at odds with any current ethical perspectives on cloning and reproductive technologies that one can’t really imagine how a world could have got this way. After all, in other respects it seems to be a world just like ours. It is not even set in some dystopian future, but has a feeling of being more like the 1980s. The ‘normal’ humans aren’t cold-hearted dysfunctionals – they seem pretty much like ordinary people, except that they seem to accept this donor business largely without question – whereas nothing like this would be tolerated or even contemplated for an instant today. It feels as though Ishiguro just hasn’t worked hard enough to make an alternative reality that can support the terrible scenario he portrays. As a result, whatever broader point he is making loses its force. What we are left with is a well told tale of friendship and tragedy experienced by sympathetic characters put in a situation that couldn’t arise under the social conditions presented. I enjoyed the book, but I can’t see how it can add much to the cloning debate. Perhaps, as one reviewer suggested, this is all just an allegory about mortality – in which case it works rather well, but is somewhat perverse.

I’ve just taken a look at M John Harrison’s review in the Guardian, which puts these same points extremely well:
“Inevitably, it being set in an alternate Britain, in an alternate 1990s, this novel will be described as science fiction. But there's no science here. How are the clones kept alive once they've begun "donating"? Who can afford this kind of medicine, in a society the author depicts as no richer, indeed perhaps less rich, than ours?

Ishiguro's refusal to consider questions such as these forces his story into a pure rhetorical space. You read by pawing constantly at the text, turning it over in your hands, looking for some vital seam or row of rivets. Precisely how naturalistic is it supposed to be? Precisely how parabolic? Receiving no answer, you're thrown back on the obvious explanation: the novel is about its own moral position on cloning. But that position has been visited before (one thinks immediately of Michael Marshall Smith's savage 1996 offering, Spares). There's nothing new here; there's nothing all that startling; and there certainly isn't anything to argue with. Who on earth could be "for" the exploitation of human beings in this way?

Ishiguro's contribution to the cloning debate turns out to be sleight of hand, eye candy, cover for his pathological need to be subtle… This extraordinary and, in the end, rather frighteningly clever novel isn't about cloning, or being a clone, at all. It's about why we don't explode, why we don't just wake up one day and go sobbing and crying down the street, kicking everything to pieces out of the raw, infuriating, completely personal sense of our lives never having been what they could have been.”

Monday, November 26, 2007

Listen out

Let me now be rather less coy about media appearances. This Wednesday night at 9 pm I am presenting Frontiers on BBC Radio 4, looking at digital medicine. This meant that I got to strap a ‘digital plaster’ to my chest which relayed my heartbeat to a remote monitor through a wireless link. I am apparently alive and well.

Salt-free Paxo

No one can reasonably expect Jeremy Paxman to have a fluent knowledge of all the subjects on which he has to ask sometimes remarkably different questions on University Challenge. But if the topic is chemistry, you’d better get it word-perfect, because he’s got no latitude for interpretation. Tonight’s round had a moment that went something like this:
Paxman: “Which hydrated ferrous salt was once known as green vitriol?”
Hapless student: “Iron sulphate.”
Paxman: “No, it’s just sulphate.”
I’ve seen precisely the same thing happen before. How come someone doesn’t pick Paxo up on it? The fact is, contestants are advised that they can press their button to challenge if they think their answer was unfairly dismissed. The offending portion of the filming then gets snipped out. But I suspect no one ever does this – it’s just too intimidating to say to Paxo “I think you’ve got that wrong.”

Friday, November 23, 2007

War is not an exact science
[This is my latest muse column for news@nature.com]

General theories of why we go to war are interesting. But they'll never tell the whole story.

Why are we always fighting wars? That’s the kind of question expected from naïve peaceniks, to which historians will wearily reply “Well, it’s complicated.”

But according to a new paper by an international, interdisciplinary team, it isn’t that complicated. Their answer is: climate change. David Zhang of the University of Hong Kong and his colleagues show that, in a variety of geographical regions – Europe, China and the arid zones of the Northern Hemisphere – the frequency of war has fluctuated in step with major shifts in climate, particularly the Little Ice Age from the mid-fifteenth until the mid-nineteenth century [1].

Cold spells like this, they say, significantly reduced agricultural production, and as a result food prices soared, food became scarce – and nations went to war, whether to seize more land or as a result of famine-induced mass migration.

On the one hand, this claim might seem unexceptional, even trivial: food shortages heighten social tensions. On the other hand, it is outrageous: wars, it says, have little to do with ideology, political ambition or sheer greed, but are driven primarily by the weather.

Take, for example, the seventeenth century, when Europe was torn apart by strife. The Thirty Years War alone, between 1618 and 1648, killed around a third of the population in the German states. Look at the history books and you’ll find this to be either a religious conflict resulting from the Reformation of Martin Luther and Jean Calvin, or a political power struggle between the Habsburg dynasty and their rivals. Well, forget all that, Zhang and his colleagues seem to be saying: it’s all because we were suffering the frigid depths of the Little Ice Age.

I expect historians to respond to this sort of thing with lofty disdain. You can see their point. The analysis stops at 1900, and so says nothing about the two most lethal wars in history – which, as the researchers imply, took place in an age when economic, technological and institutional changes had reduced the impact of agricultural production on world affairs. Can you really claim to have anything like a ‘theory of war’ if it neglects the global conflicts of the twentieth century?

And historians will rightly say that grand synoptic theories of history are of little use to them. Clearly, not all wars are about food. Similarly, not all food shortages lead to war. There is, in historical terms, an equally compelling case to be made that famine leads to social unrest and potential civil war, not to the conflict of nation states. But more generally, the point of history (say most historians) is to explain why particular events happened, not why generic social forces sometimes lead to generic consequences. There is a warranted scepticism of the kind of thinking that draws casual parallels between, say, Napoleon’s imperialism and current US foreign policy.

Yet some of this resistance to grand historical theorizing may be merely a backlash. In particular, it stands in opposition to the Marxist position popular among historians around the middle of the last century, and which has now fallen out of fashion. And the Marxist vision of a ‘scientific’ socio-political theory was itself a product of nineteenth century mechanistic positivism, as prevalent among conservatives like Leo Tolstoy and liberals like John Stuart Mill as it was in the revolutionary socialism of Marx and Engels. It was Tolstoy who, in War and Peace, invoked Newtonian imagery in asking “What is the force that moves nations?”

Much of this can be traced to the famous proposal of Thomas Robert Malthus, outlined in his Essay on the Principles of Population (1826), that population growth cannot continue for ever on an exponential rise because it eventually falls foul of the necessarily slower rise in means of production – basically, the food runs out. That gloomy vision was also an inspiration to Charles Darwin, who saw that in the wild this competition for limited resources must lead to natural selection.

Zhang and colleagues state explicitly that their findings provide a partial vindication of Malthus. They point out that Malthus did not fully account for the economic pressures and sheer ingenuity that could boost agricultural production when population growth demanded it, but they say that such improvements have their limits, which were exceeded when climate cooling lowered crop yields in Europe and China.

For all their apparently impressive correlation indices, however, it is probably fair to say that responses to Zhang et al.’s thesis will be a matter of taste. In the end, an awful lot seems to hinge on the coincidence of minimal agricultural production (and maximum in food prices), low average temperatures, and a peak in the number of wars (and fatalities) during the early to mid-seventeenth century in both Europe and China. The rest of the curves are suggestive, but don’t obviously create a compelling historical narrative. At best, they provoke a challenge: if one cannot now show a clear link between climate/agriculture and, say, the Napoleonic wars from the available historical records themselves, historians might be forgiven for questioning the value of this kind of statistical analysis.

Yet what if the study helps us to understand, even a little bit, what causes war? That itself is an age-old question – Zhang and colleagues identify it, for example, in Thucydides’ History of the Peloponnesian Wars in the 5th century BC. Neither are they by any means the first in modern times to look for an overarching theory of war. The issue motivated the physicist Lewis Fry Richardson between about 1920 and 1950 to plot size against frequency for many recent wars (including the two world wars), and thereby to identify the kind of power-law scaling that has led to the notion that wars are like landslides, where small disturbances can trigger events of any scale [2-4]. Other studies have focused on the cyclic nature of war and peace, as for example in ecologist Peter Turchin’s so-called cliodynamics, which attempts to develop a theory of the expansion and collapse of empires [5,6].

Perhaps most prominent in this arena is an international project called the Correlates of War, which has since 1963 been attempting to understand and quantify the factors that create (and mitigate) international conflict and thus to further the “scientific knowledge about war”. Its data sets have been used, for example, in quantitative studies of how warring nations form alliances [7], and they argue rather forcefully against any notion of collapsing the causative factors onto a single axis such as climate.

What, finally, do Zhang and colleagues have to tell us about future conflict in an anthropogenically warmed world? At face value, the study might seem to say little about that, given that it correlates war with cooling events. There is some reason to think that strong warming could be as detrimental to agriculture as strong cooling, but it’s not clear exactly how that would play out, especially in the face of both a more vigorous hydrological cycle and the possibility of more regional droughts. We already know that water availability will become a serious issue for agricultural production, but also that there’s a lot that can still be done to ameliorate that, for instance by improvements in irrigation efficiency.

We’d be wise to greet the provocative conclusions of Zhang et al. with neither naïve acceptance nor cynical dismissal. They do not amount to a theory of history, or of war, and it seems most unlikely that any such things exist. But their paper is at least a warning against a kind of fatalistic solipsism which assumes that all human conflicts are purely the result of human failings.

References

1. Zhang, D. D. et al. Proc. Natl Acad. Sci. USA doi/10.1073/pnas.0703073104
2. Richardson, L. F. Statistics of Deadly Quarrels, eds Q. Wright and C. C. Lienau (Boxwood Press, Pittsburgh, 1960).
3. Nicholson, M. Brit. J. Polit. Sci. 29, 541-563 (1999).
4. Buchanan, M. Ubiquity (Phoenix, London, 2001).
5. Turchin, P. Historical Dynamics (Princeton University Press, 2003).
6. Turchin, P. War and Peace and War (Pi Press, 2005).
7. Axelrod R. & D. S. Bennett, Brit. J. Polit. Sci. 23, 211-233 (1993).

Thursday, November 22, 2007

Schrödinger’s cat is not dead yet

[This is an article I’ve written for news@nature. One of the things I found most interesting was that Schrödinger didn’t set up his ‘cat’ thought experiment with a gun, but with an elaborate poisoning scheme. Johannes Kofler says “He puts a cat into a steel chamber and calls it "hell machine" (German: Höllenmaschine). Then there is a radioactive substance in such a tiny dose that within one hour one atom might decay but with same likelihood nothing decays. If an atom decays, a Geiger counter reacts. In this case this then triggers a small hammer which breaks a tiny flask with hydrocyanic acid which poisons the cat. Schrödinger is really very detailed in describing the situation.” There’s a translation of Schrödinger’s original paper here, but as Johannes says, the wonderful “hell machine” is simply translated as “device”, which is a bit feeble.]

Theory shows how quantum weirdness may still be going on at the large scale.

Since the particles that make up the world obey the rules of quantum theory, allowing them to do counter-intuitive things such as being in several different places or states at once, why don’t we see this sort of bizarre behaviour in the world around us? The explanation commonly offered in physics textbooks is that quantum effects apply only at very small scales, and get smoothed away at the everyday scales we can perceive.

But that’s not so, say two physicists in Austria. They claim that we’d be experiencing quantum weirdness all the time – balls that don’t follow definite paths, say, or objects ‘tunnelling’ out of sealed containers – if only we had sharper powers of perception.

Johannes Kofler and Caslav Brukner of the University of Vienna and the Institute of Quantum Optics and Quantum Information, also in Vienna, say that the emergence of the ‘classical’ laws of physics, deduced by the likes of Galileo and Newton, from the quantum world is an issue not of size but of measurement [1]. If we could make every measurement with as much precision as we liked, there would be no classical world at all, they say.

Killing the cat

Austrian physicist Erwin Schrödinger famously illustrated the apparent conflict between the quantum and classical descriptions of the world. He imagined a situation where a cat was trapped in a box with a small flask of poison that would be broken if a quantum particle was in one state, and not broken if the particle was in another.

Quantum theory states that such a particle can exist in a superposition of both states until it is observed, at which point the quantum superposition ‘collapses’ into one state or the other. Schrödinger pointed out that this means that the cat is neither dead nor alive until someone opens the box to have a look – a seemingly absurd conclusion.

Physicists generally resolve this paradox through a process called decoherence, which happens when quantum particles interact with their environment. Decoherence destroys the delicately poised quantum state and leads to classical behaviour.

The more quantum particles there are in a system, the harder it is to prevent decoherence. So somewhere in the process of coupling a single quantum particle to a macroscopic object like a flask of poison, decoherence sets in and the superposition is destroyed. This means that Schrödinger’s cat is always unambiguously in a macroscopically ‘realistic’ state, either alive or dead, and not both at once.

But that’s not the whole story, say Kofler and Brukner. They think that although decoherence typically intervenes in practice, it need not do so in principle.

Bring back the cat

The fate of Schrödinger’s cat is an example of what in 1985 physicists Anthony Leggett and Anupam Garg called macrorealism [2]. In a macrorealistic world, they said, objects are always in a single state and we can make measurements on them without altering that state. Our everyday world seems to obey these rules. According to the macrorealistic view, “there are no Schrödinger cats allowed” says Kofler.

But Kofler and Brukner have proved that a quantum state can get as ‘large’ as you like, without conforming to macrorealism.

The two researchers consider a system akin to a magnetic compass needle placed in a magnetic field. In our classical world, the needle rotates around the direction of the field in a process called precession. That movement can be described by classical physics. But in the quantum world, there would be no smooth rotation – the needle could be in a superposition of different alignments, and would just jump instantaneously into a particular alignment once we tried to measure it.

So why don’t we see quantum jumps like this? The researchers show that it depends on the precision of measurement. If the measurements are a bit fuzzy, so that we can’t distinguish one quantum state from several other, similar ones, this smoothes out the quantum oddities into a classical picture. Kofler and Brukner show that, once a degree of fuzziness is introduced into measured values, the quantum equations describing the observed objects turn into classical ones. This happens regardless of whether there is any decoherence caused by interaction with the environment.

Having kittens

Kofler says that we should be able to see this transition between classical and quantum behaviour. The transition would be curious: classical behaviour would be punctuated by occasional quantum jumps, so that, say, the compass needle would mostly rotate smoothly, but sometimes jump instantaneously.

Seeing the transition for macroscopic objects like Schrödinger’s cat would require that we be able to distinguish an impractically large number of quantum states. For a ‘cat’ containing 10**20 quantum particles, say, we would need to be able to tell the difference between 10**10 states – just too many to be feasible.

But our experimental tools should already be good enough to look for this transition in much smaller ‘Schrödinger kittens’ consisting of many but not macroscopic numbers of particles, says Kofler and Brukner.

What, then, becomes of these kittens before the transition, while they are still in the quantum regime? Are they alive or dead? ‘We prefer to say that they are neither dead nor alive,’ say Kofler and Brukner, ‘but in a new state that has no counterpart in classical physics.’

References

1. Kofler, J. & Brukner, C. Phys. Rev. Lett. 99, 180403 (2007).
2. Leggett, A. & Garg, A. Phys. Rev. Lett. 54, 857 (1985).
Not natural?
[Here’s a book review I’ve written for Nature, which I put here because the discussion is not just about the book!]

The Artificial and the Natural: An Evolving Polarity
Ed. Bernadette Bensaude-Vincent and William R. Newman
MIT Press, Cambridge, Ma., 2007

The topic of this book – how boundaries are drawn between natural and synthetic – has received too little serious attention, both in science and in society. Chemists are notoriously (and justifiably) touchy about descriptions of commercial products as ‘chemical-free’; but the usual response, which is to lament media or public ignorance, fails to recognize the complex history and sociology that lies behind preconceptions about chemical artifacts. Roald Hoffmann has written sensitively on this matter in The Same and Not the Same (Columbia University Press, 1995), and he contributes brief concluding remarks to this volume. But the issue is much broader, touching on areas ranging from stem-cell therapy and assisted conception to biomimetic engineering, synthetic biology, machine intelligence and ecosystem management.

It is not, in fact, an issue for the sciences alone. Arguably the distinction between nature and artifice is equally fraught in what we now call the fine arts – where again it tends to be swept under the carpet. While some modern artists, such as Richard Long and Andy Goldsworthy, address the matter head-on with their interventions in nature such as the production of artificial rainbows, much popular art criticism now imposes a contemporary view even on the Old Masters. Through this lens, Renaissance writer Giorgio Vasari’s astonishment that Leonardo’s painted dewdrops “looked more convincing than the real thing” appears a little childish, as though he has missed the point of art – for no one now believes that the artist’s job is to mimic nature as accurately as possible. Perhaps with good reason, but it is left to art historians to point out that there is nothing absolute about this view.

At the heart of the matter is the fact that ‘art’ has not always meant what it does today. Until the late Enlightenment, it simply referred to anything human-made, whether that be a sculpture or an engine. The panoply of mutated creatures described in Francis Bacon’s The New Atlantis (1627) were the products of ‘art’, and so were the metals generated in the alchemist’s laboratory. The equivalent word in ancient Greece was techne, the root of ‘technology’ of course, but in itself a term that embraced subtle shades of meaning, examined here in ancient medicine by Heinrich von Staden and in mechanics by Francis Wolff.

The critical issue was how this ‘art’ was related to ‘nature’, approximately identified with what Aristotle called physis. Can art produce things identical to those in nature, or only superficial imitations of them? (That latter belief left Plato rather dismissive of the visual arts.) Does art operate using the same principles as nature, or does it violate them? Alchemy was commonly deemed to operate simply by speeding up natural processes: metals ripened into gold sooner in the crucible than they did in the ground, while (al)chemical medicines accelerated natural healing. And while some considered ‘artificial’ things to be always inferior to their ‘natural’ equivalents, it was also widely held that art could exceed nature, bringing objects to a greater state of perfection, as Roger Bacon thought of alchemical gold.

The emphasis in The Artificial and the Natural is historical, ranging from Hippocrates to nylon. These motley essays are full of wonders and insights, but are ultimately frustrating too in their microcosmic way. There is no real synthesis on offer, no vision of how attitudes have evolved and fragmented. There are too many conspicuous absences for the book to represent an overview. One can hardly feel satisfied with such a survey in which Leonardo da Vinci is not even mentioned. It would have been nice to see some analysis of changing ideas about experimentation, the adoption of which was surely hindered by Aristotle’s doubts that ‘art’ (and thus laboratory manipulation) was capable of illuminating nature. Prejudices about experiments often went even further: even in the Renaissance one could feel free to disregard what they said if it conflicted with a priori ‘truths’ gleaned from nature, rather as Pythagoras advocated studying music “setting aside the judgement of the ears”. And it would have been fascinating to see how these issues were discussed in other cultures, particularly in technologically precocious China.

But most importantly, the discussion sorely lacks a contemporary perspective, except for Bernadette Bensaude-Vincent’s chapter on plastics and biomimetics. This debate is no historical curiosity, but urgently needs airing today. Legislation on trans-species embryology, reproductive technology, genome engineering and environmental protection is being drawn up based on what sometimes seems like little more than a handful of received wisdoms (some of them scriptural) moderated by conventional risk analysis. There is, with the possible exception of biodiversity discussions, almost no conceptual framework to act as a support and guide. All too often, what is considered ‘natural’ assumes an absurdly idealized view of nature that owes more to the delusions of Rousseau’s romanticism than to any historically informed perspective. By revealing how sophisticated, and yet how transitory, the distinctions have been in the past, this book is an appealingly erudite invitation to begin the conversation.

Sunday, November 18, 2007


Astronomy: the dim view

One Brian Robinson contributes to human understanding on the letters page of this Saturday’s Guardian with the following:
“Providing funding for astronomers does not in any way benefit the taxpayer. Astronomy may be interesting, but the only mouths that will get fed are the children of the astronomers. Astronomy is a hobby, and as such should not be subsidised by the Treasury any more than trainspotting.”
The invitation is to regard this as the sort of Thatcherite anti-intellectualism that is now ingrained in our political system. And indeed, the notion that anything state-funded must ‘benefit the taxpayer’ – specifically, by putting food in mouths – is depressing not only in its contempt for learning but also in its ignorance of how the much-vaunted ‘wealth creation’ in a technological society works.

But then you say, hang on a minute. Why astronomy, of all things? Why not theology, archaeology, philosophy, and all the arts other than the popular forms that are mass-marketable and exportable? And then you twig: ‘astronomy is a hobby’ – like trainspotting. This bloke thinks that professional astronomers are sitting round their telescopes saying ‘Look, I’ve just got a great view of Saturn’s rings!’ They are like the funny men in their sheds looking at Orion, only with much bigger telescopes (and sheds). In other words, Mr Robinson hasn’t the faintest notion of what astronomy is.

Now, I have some gripes with astronomers. It is not just my view, but seems to be objectively the case, that the field is sometimes narrowly incestuous and lacks the fecundity that comes from collaborating with people in other fields, with the result that its literature is often far more barren than it has any right to be, given what’s being studied here. And the astronomical definition of ‘metals’ is so scientifically illiterate that it should be banned without further ado, or else all other scientists should retaliate by calling anything in space that isn’t the Earth a ‘star’. But astronomy is not only one of the oldest and most profound of human intellectual endeavours; it also enriches our broader culture in countless ways.

The presence of Mr Robinson’s letter on the letters page, then, is not a piece of cheeky provocation, but an example of the nearly ubiquitous ignorance of science among letters-page editors. They simply didn’t see what he was driving at, and thus how laughable it is. It is truly amazing what idiocies can get into even the most august of places – the equivalent, often, of a reader writing in to say that, oh I don’t know, that Winston Churchill was obviously a Kremlin spy or that Orwell wrote Cold Comfort Farm. Next we’ll be told that astronomers are obviously fakes because their horoscopes never come true.

Monday, November 12, 2007


Is this what writers’ studies really look like?

Here is another reason to love Russell Hoban (aside from his having written the totally wonderful Riddley Walker, and a lot of other great stuff too). It is a picture of his workplace, revealed in the Guardian’s series of ‘writers’ rooms’ this week. I love it. After endless shots of beautiful mahogany desks surrounded by elegant bookshelves and looking out onto greenery, like something from Home and Garden, here at last is a study that looks as though the writer works in it. It is the first one in the series that looks possibly even worse than mine.

The mystery is what all the other writers do. Sure, there may be little stacks of books being used for their latest project – but what about all the other ‘latest projects’? The papers printed out and unread for months? The bills unpaid (or paid and not filed)? The letters unanswered (or ditto)? The books that aren’t left out for any reason, other than that there is no other place to put them? The screwdrivers and sellotape and tissues and plastic bags and stuff I’d rather not even mention? What do these people do all day? These pictures seem to demand the image of a writer who, at the end of the day, stretches out his/her arms and says “Ah, now for a really good tidy-up”. That is where my powers of imagination fail me.

It all confirms that we simply do not deserve Russell Hoban.

Sunday, November 11, 2007

Minority report

Here’s an interesting factoid culled from the doubtless unimpeachable source of Wllson da Silva, editor of Australian science magazine Cosmos: the proportion of scientist who question that humans are responsible for global warming is about the same as the proportion who questions that HIV is the cause of AIDS. Strange, then, that whenever AIDS is discussed on TV or radio, it is not considered obligatory to include an HIV sceptic for ‘balance’.

Of course, one reason for that is that people are not (yet) dying in their thousands from climate change (although even that, after the recent European heat waves, is debatable). This means it can remain fashionable, among over-educated media types with zero understanding of science, to be a climate sceptic. This, not the little band of scientific deniers, less still the so-called ignorant masses that some scientists lament, is the real problem. The intelligensia still love to parade their ‘independent-mindedness’ on this score.

Here, for example, is Simon Hoggart a couple of weeks ago in the Guardian on ‘man-made global warming’: “I'm not going to plunge into this snakepit, except to say that there are more sceptics about than the Al Gores of this world acknowledge, and they are not all paid by carbon fuel lobbies. Also, if it's true, as Booker and North claim [in their book Scared to Death], that there is evidence of global warming on other planets, might it not be possible that the sun has at least as much effect on our climate as we do? I only ask.”

No, Simon, you do not only ask. If you genuinely wanted enlightenment on this matter, you could go to the wonderful Guardian science correspondents, who would put you straight in seconds. No, you want to show us all what a free thinker you are, and sow another little bit of confusion.

I’m not even going to bother to explain why ‘other planets’ (Venus, I assume) have experienced global warming in their past. It is too depressing. What is most depressing of all is the extent to which well-educated people can merrily display such utter absence of even the basics of scientific reasoning (such as comparing like with like). I’m generally optimistic about people’s ability to learn and reason, if they have the right facts in front of them. But I sometimes wonder if that ability declines the more you know, when that knowledge excludes anything to do with science.

Monday, October 22, 2007

Lucky Jim runs out of luck (at last)

[This is also posted on the Prospect blog .]

Jim Watson seems to be genuinely taken aback by the furore his recent comments on race and IQ have aroused. He looks a little like the teenage delinquent who, after years of being a persistent neighbourhood pest, finds himself suddenly hauled in front of a court and threatened with being sent to a detention centre. Priding himself on being a social irritant, he never imagined anyone would deal with him seriously.

The truth is that there is more than metaphor in this image. Watson has throughout his career combined the intelligence of a first-rate scientist and the influence of a Nobel laureate with the emotional maturity of a spoilt schoolboy. There is nothing particularly remarkable about that – it is not hard to find examples of immaturity among public figures – but the scientific community seems to find it particularly difficult to know how to accommodate such cases. For better or worse, there are plenty of niches for emotionally immature show-offs in politics and the media – the likes of Boris Johnson, Ann Widdecombe, Jeremy Clarkson and Ann Coulter all, in their own ways, manage it with aplomb. (It is not a trait unique to right-wingers, but somehow they seem to do it more memorably.) But although they can sometimes leave po-faced opponents spluttering, the silliness is usually too explicit to be mistaken for anything else.

Science, on the other hand, has tended to be blind to this facet of human variety, so that the likes of Watson come instead to be labelled “maverick” or “controversial”, which of course is precisely what they want. The scientific press tends to handle these figures with kid gloves, pronouncing gravely on the propriety of their “colourful” remarks, as though these are sober individuals who have made a bad error of judgement. Henry Porter is a little closer to the mark in the Observer, where he calls Watson an ‘elderly loon’ – the degree of ridicule is appropriate, except that Watson is no loon, and it has been a widespread mistake to imagine that his comments are a sign of senescence.

The fact is that Watson has always considered it great sport to say foolish things that will offend people. He is of the tiresome tribe that likes to display what they deem to be ‘politically incorrectness’ as a badge of pride, forgetting that they would be ignored as bigoted boors if they did not have power and position. It is abundantly clear that behind the director of the Cold Spring Harbor Laboratory still stands the geeky young man depicted behind a model of DNA in the 1950s, whose (eminently deserved) Nobel has protected him from a need to grow up. “He was given licence to say anything that came into his mind and expect to be taken seriously,” said Harvard biologist E. O. Wilson (himself no stranger to controversy, but an individual who exudes far more wisdom and warmth than Watson ever has).

That’s a pitfall for all Nobel laureates, of course, and many are tripped by it. But few have embraced the licence with as much delight as Watson. For example, there was this little gem over a decade ago: “If you are really stupid, I would call that a disease. The lower 10 per cent who really have difficulty, even in elementary school, what’s the cause of it? A lot of people would like to say, ‘Well, poverty, things like that.’ It probably isn't. So I’d like to get rid of that, to help the lower 10 per cent.” Or this one: “Whenever you interview fat people, you feel bad, because you know you’re not going to hire them.”

Watson has been called “extraordinarily naïve” to have made his remarks about race and intelligence and expect to get away with them. But it is not exactly naivety – he probably just assumed that, since he has said such things in the past without major incident, he could do so again. Indeed, he almost did get away with it, until the Independent decided to make it front-page news.

Watson has apologized “unreservedly” for his remarks, which he says were misunderstood. This is mostly a public-relations exercise – it is not clear that there is a great deal of scope for misunderstanding, and evidently Watson now has a genuine concern that he will be dismissed from his post at Cold Spring Harbor. At least by admitting that there is “no scientific basis” for a belief that Africans are somehow “genetically inferior”, he has provided some ammunition to counter the opportunistic use of his remarks by racist groups. But it is inevitable that those groups will now make him a martyr, forced to recant in the manner of Galileo for speaking an unpalatable truth. (The speed with which support for Watson’s comments has come crawling out of the woodwork even in august forums such as Nature’s web site is disturbing.)

The more measured dismay that some, including Richard Dawkins, have voiced over the suppression of free speech implied by the cancellation of some of Watson’s intended UK talks, is understandable, although it seems not unreasonable (indeed, it seems rather civil) for an institution to decide it does not especially want to host someone who has just expressed casual racist opinions. More to the point, it is not clear what ‘free speech’ is being suppressed here – Watson does not appear to be wanting to, and being prevented from, making a case that black people are less intelligent than other races. (In fact it is no longer clear what Watson wanted to say at all; the most likely interpretation is that he simply let a groundless prejudice slip out in an attempt to boost his ‘bad boy’ reputation, and that he now regrets it.) In a funny sort of way, Watson would be less deserving of scorn if he were now defending his remarks on the basis of the ‘evidence’ he alluded to. In that event, any kind of censorship would indeed be misplaced.

Beneath the sound a fury, however, we should remember that Watson’s immense achievements as a scientist do not oblige us to take him seriously in any other capacity. Those achievements are orthogonal to his bully-boy bigotry, and they put no distance at all between Watson and the pub boor.

The real casualty in all this is genetics research, for Watson’s comments past and present can only seem (and in fact not just seem) to validate claims that this research is in the hands of scientists with questionable judgement and sense of responsibility.

Friday, October 19, 2007

Swiss elections get spooky
[This is my latest column for muse@nature.com.]

High-profile applications of quantum trickery raise the question of what to call these new technologies. One proposal is unlikely to catch on.

The use of quantum cryptography in the forthcoming Swiss general elections on 21 October may be a publicity stunt, but it highlights the fact that the field of quantum information is now becoming an industry.

The invitation here is to regard Swiss democracy as being safeguarded by the fuzzy shroud of quantum physics, which can in principle provide a tamper-proof method of transmitting information. The reality is that just a single state – Geneva – is using commercial quantum-cryptography technology already trialled by banks and financial institutions, and that it is doing so merely to send tallies from a vote-counting centre to the state government’s repository.

The votes themselves are being delivered by paper ballot – which, given the controversies over electronic voting systems, is probably still the most secure way to collect them. In any event, with accusations of overt racism in the campaigning of the right-wing Swiss People’s Party (SVP), hacking of the voting system is perhaps the least of the worries in this election.

But it would be churlish to portray this use of quantum cryptography as worthless. There is no harm in using a high-profile event to advertise the potential benefits of the technology. If nothing else, it will get people asking what quantum cryptography is.

The technique doesn’t actually make transmitted data invulnerable to tampering. Instead, it makes it impossible to interfere with the transmission without leaving a detectable trace. Some quantum cryptographic schemes use the quantum-mechanical property of entanglement, whereby two or more quantum particles are woven together so that they become a single system. Then you can’t do something to one particle without affecting the others with which it is entangled.

Entanglement isn’t essential for quantum encryption – the first such algorithm, devised by physicists Charles Bennett and Gilles Brassard in 1984, instead relies on a property called quantum indeterminacy, denoting our fundamental inability to describe some quantum systems exactly. Entanglement, however, is the key to a popular scheme devised in 1991. Here, the sender and receiver each receive one of a pair of entangled particles, and can decode a message by comparing their measurements of the particles’ quantum states. Any eavesdropping tends to randomize the relationship between these states, and is therefore detectable.

Quantum cryptography is just one branch of the emerging discipline of quantum information technology, in which phenomena peculiar to the quantum world, such as entanglement, are used to manipulate information. Other applications include quantum computing, in which quantum particles are placed in superposition states – mixtures of the classical states that would correspond to the binary 1’s and 0’s of ordinary computers – to vastly boost the power and capacity of computation. Quantum teleportation – the exact replication of quantum particles at locations remote from the originals – also makes use of entanglement.

The roots of these new areas of quantum physics lie in the early days of quantum theory, when its founders were furiously debating what quantum theory implied about the physical world. Albert Einstein, whose Nobel-winning explanation of the photoelectric effect was one of the cornerstones of quantum mechanics, doubted that quantum particles could really have the fuzzy properties ascribed to them by the theory, to which one could do no more than assign probabilities.

In 1935 Einstein and his colleagues Boris Podolsky and Nathan Rosen proposed a thought experiment that they hoped would show quantum theory to be an incomplete account of physical reality. They showed how it seemed to predict what Einstein called ‘spooky action at a distance’ that operated instantaneously between two particles.

But we now know that this action at a distance is real – it is the result of quantum entanglement. What Einstein considered a self-evident absurdity is simply the way the world is. What’s more, entanglement and superpositions are now recognized as being key to the way our deterministic classical world, where events have definite outcomes, emerges from the murky haze of quantum probabilities.

Bennett was one of the pioneers who showed that these quantum effects aren’t just abstract curiosities, but can be exploited in applications. For this, he will surely get a Nobel prize some time soon.

So far, most researchers have been happy to talk about ‘quantum cryptography’, ‘quantum computing’ and so forth, vaguely gathered under the umbrella phrase of quantum information. But is that a good name for a technology? Charles Tahan, a physicist at the University of Cambridge who is working on these technologies, thinks not. In a recent preprint, he proposes to draw inspiration from Einstein and call it all ‘spookytechnology’.

This, says Tahan, would refer to “all functional devices, systems and materials whose utility relies in whole or in part on higher order quantum properties of matter and energy that have no counterpart in the classical world.” By higher-order, Tahan means things like entanglement and superposition. He argues that his definition is broad enough to contain more than quantum information technology, but not so broad as to be meaningless.

In that respect, Tahan points to the shortcomings of ‘nanotechnology’, a field that is not really a field at all but instead a ragbag of many areas of science and technology ranging from electronics to biomedicine.

But Tahan's label will never stick, because it violates one of the most fundamental prohibitions in scientific naming: don’t be cute. No scientist is going to want to tell people that he or she is working in a field that sounds as though it was invented by Caspar the Friendly Ghost. True, the folksy ‘buckyballs’ gained some currency as a term for the fullerene carbon molecules (despite Nature’s best efforts) – but its usage remains a little marginal, and has thankfully never caught on for ‘buckytubes’, which everyone instead calls carbon nanotubes.

Attempts to label nascent fields rarely succeed, for names have a life of their own. ‘Nanotechnology’, when coined in 1974, had nothing like the meaning it has today. ‘Spintronics’, the field of quantum electronics that in some sense lies behind this year’s physics Nobel, is arguably a slightly ugly and brutal amalgam of electronics and the quantum property of electrons called spin - yet somehow it works.

Certainly, names need to be catchy: laboured plunderings of Greek and Latin are never popular. But catchiness is extremely hard to engineer. So somehow I don’t think we’re going to see the Geneva elections become a landmark in spookytechnology.

Thursday, October 18, 2007







How tortoises turn right-side up
[This is a story I’ve just written for Nature’s news site. But the deadline was such that we couldn’t include the researchers’ nice pics of tortoises and turtles doing their stuff. So here are some of them. The first is an ideal monostatic body, and a tortoise that approximates it. The second is a flat turtle righting itself by using its neck as a pivot. The last two are G. elegans shells, which are nearly monostatic.]

Study finds three ways that tortoises avoid getting stuck on their backs.

Flip a tortoise or a turtle over, and it’ll find its feet again. Two researchers have now figured out how they do it — they use a clever combination of shell shape and leg and neck manoeuvres.

As Franz Kafka’s Gregor discovered in Metamorphosis, lying on your back can be bad news if you’re cockroach-shaped. Both cockroaches and tortoises are potentially prone to getting stuck on their rounded backs, their feet flailing in the air.

For tortoises, this is more than an accidental hazard: belligerent males often try to flip opponents over during fights for territorial rights. Gábor Domokos of Budapest University of Technology and Economics and Péter Várkonyi of Princeton University in New Jersey took a mathematical look at real animals to see whether they had evolved sensible shapes to avoid getting stuck [1].

The ideal answer would seem to be to have a shell that can’t get stuck at all — one that will spontaneously roll back under gravity, like the wobbly children's toys that “won’t fall down”. Domokos and Várkonyi have investigated the rolling mechanics of idealized shell shapes, and show that in theory, such self-righting shells do exist. They would be tall domes with a cross-section like a half-circle slightly flattened on one side.

The shells of some tortoises, such as the star tortoise Geochelone elegans, come very close to this shape. They can still get stuck because of small imperfections in the shell shape, but it takes only a little leg-wagging to make the tortoise tip over and right itself. The researchers call tall shells that have a single stable resting orientation (on the tortoise's feet) monostatic, denoted as group S1.

The tall and the squat

So, tall shells are generally good for righting with minimal effort, and confer good protection against the jaws of predators. Could this be the best answer for all turtles and tortoises? No real chelonian has a perfectly monostatic shell, which Várkonyi says is probably because tall shells could have disadvantages too: you could be rolled over by wind, for instance. Also, he says, it takes quite a bit of fine-tuning to achieve a truly monostatic shape.

Flatter shells have other advantages: they can, for example, be better for swimming or for use as spade-like implements for digging. The side-necked turtle and the pancake tortoise are flat like this, with two stable resting positions (S2): right side up and on their back.

For such flat shells, righting requires more than a bit of thrashing around. These animals tend to have long necks, which they extend and use as a pivot while pushing with their legs. The longer the neck, the easier it is for the creature to right itself, in the same way that a long lever can be pushed down with less effort than a short one.

Stuck in the middle

In between these two extremes of tall and flat are shells that are moderately domed, as found in Terrapene box turtles. Surprisingly, these have three stable positions (S3): on the back, on the front or halfway between, where the shell rests on its curved side.

Turtles of the S3 class use a combination of both strategies: bobbing of their head or feet tips the shell from the back-down position to the sideways position, and from there the creature can use its neck and feet to pivot over into the belly-down state.

The work is sure to be of interest to tortoise keepers and kids with turtle pets. But it's unlikely that this tortoise-rolling work is going to suggest new ways to help robots pick themselves up — engineers already have a number of quite simple ways of ensuring that. "You can just put ballast in the bottom," Várkonyi admits.

Reference
1. Domokos, G. & Várkonyi, P. L. Proc. R. Soc. B, doi:10.1098/rspb.2007.1188.

Tuesday, October 16, 2007

We’ll never know how we began

[This is the pre-edited text of my Crucible column for the November issue of Chemistry World.]

Oddly, it is easier to explore the origin of the universe than the origin of life on Earth. ‘Easier’ is a relative term here, because the construction of the Large Hardon Collider at CERN in Geneva makes clear the increasing extravagance needed to push back the curtain ever closer to the singularity of the Big Bang. But we can now reconstruct the origin of our universe from about 10**-30 of a second onwards, and the LHC may take us back into the primordial quark-gluon plasma and the symmetry-breaking transition of the Higgs field that created particle masses.

Yet all this is possible precisely because there is so little room for contingency in the first instants of the Big Bang. The further back we go, the less variation we are likely to find between our universe and another one hypothetically sprung from a cosmic singularity – most of what happened then is constrained by physics. So while the LHC might produce some surprises, it could instead simply confirm what we expected.

The origin of life is totally different. There isn’t really any theory that can tell us about it. It might have happened in many different ways, depending on circumstances of which we know rather little. In this sense, it is a genuinely historical event, immune to first-principles deduction in the same way as are the shapes of the early continents or the events of the Hundred Years War. What we know about the former is largely a matter of extrapolating backwards from the present-day situation, and then searching for geological confirmation. We can do the same for the history of life, constructing phylogenetic trees from comparisons of extant organisms and supplementing that with data from the fossil record. But that approach can tell us little about what life was like before it was really life at all.

For the Hundred Years War there is ample documentary evidence. But for life’s origin around 3.8 billion years ago, the geological ‘documents’ tell us very little indeed. Life left its imprint in the rocks once it was fully fledged, but there is no real data on how it got going.

It is a testament to the tenacity and boldness of scientists that they have set out to explore the question anyway. In 1863 Charles Darwin concluded that there was little point in doing so: “It is mere rubbish”, he wrote, “thinking at present on the origin of life.” But he evidently had a change of heart, since eight years later he could be found musing on his “warm little pond” filled with a broth of prebiotic compounds. By the time Alexander Oparin and J. B. S. Haldane speculated about the formation of organic molecules in primitive atmospheres in the 1920s, experimentalists had already shown that substances such as formaldehyde and the amino acid glycine could be cooked up from carbon oxides, ammonia and water.

There was, then, a long tradition behind the ground-breaking experiment of Harold Urey and Stanley Miller at Chicago in 1953. They, however, were the first to use a reducing mixture, and that is why they found such a rich mélange of organics in their brew. Despite geological evidence suggesting that the early terrestrial atmosphere was mildly oxidizing, Miller remained convinced until his recent death that this was the only plausible way life’s building blocks could have been made – some say his stubbornness on this issue ended up hindering progress in the field.

In some ways, the recent study by Paul von Ragué Schleyer of the University of Georgia and his coworkers of the prebiotic synthesis of the nucleic acid base adenine from hydrogen cyanide (D. Roy et al., Proc. Natl Acad. Sci. USA doi:10.1073 pnas.0708434104) is a far cry from Urey and Miller’s makeshift ‘bake and shake’ experiment. It uses state-of-the-art quantum chemical calculations to deduce the mechanism of this reaction, first reported by John Oró and coworkers in Texas in 1960, which produces one of the building blocks of life from five molecules of a single, simple ingredient.

But in another sense, the work might be read as an indication that the field initiated by Urey and Miller is close to having run its course in its present form. The most one could have asked of their approach – and it has amply fulfilled this demand – is that it alleviate George Wald’s objection in 1954 that “one only has to contemplate the magnitude of this task to concede that the spontaneous generation of a living organism is impossible.” There are now more or less plausibly ‘prebiotic’ ways to make most of the key molecular ingredients of proteins, RNA, DNA, carbohydrates and other complex biomolecules. There are ingenious ways of linking them together, in defiance of the deconstructive hydrolysis that dilute solution seems to threaten, ranging from surface catalysis on minerals to the use of electrochemical gradients at hot springs. There are theories of cascading complexification through autocatalytic cycles, and the whole framework of the RNA World (the answer to the chicken-and-egg problem of DNA’s dependence on proteins) seems increasingly well motivated.

And yet there is no more evidence than there was fifty years ago that this is how it all happened. Time has kicked over the tracks. The chemical origin of life has become a discipline of immense experimental and theoretical refinement, as this new paper testifies – and yet it all remains guesswork, barely constrained by hard evidence from the Hadaean eon of our planet. The true history is obliterated, and we may never glimpse it.

Sunday, October 07, 2007


Time to rethink the Outer Space Treaty
[This article on Nature’s news site formed part of the journal’s “Sputnik package”.]

An agreement forged 40 years ago can’t by itself keep space free of weaponry.

Few anniversaries have been celebrated with such mixed feelings as the launch of Sputnik-1 half a century ago. That beeping little metal orb, innocuously named “fellow traveller of Earth”, signalled the beginning of satellite telecommunications, global environmental monitoring, and space-based astronomy, as well as the dazzling saga of human journeys into the cosmos. But the flight of Sputnik was also a pivotal moment in the Cold War, a harbinger of intercontinental nuclear missiles and space-based surveillance and spying.

That’s why it seems surprising that another anniversary this year has gone relatively unheralded. In 1967, 90 nations signed the Outer Space Treaty (OST), in theory binding themselves to an agreement on the peaceful uses of space that prohibited the deployment there of weapons of mass destruction. Formally, the treaty remains in force; in practice, it is looking increasingly vulnerable as a protection against th militarization of space.

Updating and reinvigorating the commitments of the OST seems to be urgently needed, but this currently stand little chance of being realized. Among negotiators and diplomats there is now a sense of gloom, a feeling that the era of large-scale international cooperation and legislation on security issues (and perhaps more widely) may be waning.

Last year was the tenth anniversary of the Comprehensive Test Ban Treaty (CTBT), and next year the fortieth anniversary of the Nuclear Non-Proliferation Treaty. But the world’s strongest nuclear power, the United States, refuses to ratify the CTBT, while some commentators believe the world is entering a new phase of nuclear proliferation. No nuclear states have disarmed during the time of the NPT’s existence, despite the binding commitment of signatory states “to pursue negotiations in good faith on effective measures relating to nuclear disarmament”.

In this arena, the situation does seem to be in decline. For example, the US appears set on developing a new generation of nuclear weapons and deploying a ballistic missile defence system, and it withdrew from the Anti-Ballistic Missile Treaty in 2002. China and Israel have also failed to ratify the CTBT, while other nuclear powers (India, Pakistan) have not even signed it. North Korea, which withdrew from the NPT in 2003, now claims to have nuclear weapons.

Given how poorly we have done so close to home, what are the prospects for outer space? “For the past four decades”, says Sergei Ordzhonikidze, Director-General of the United Nations Office at Geneva, “the 1967 Outer Space Treaty has been the cornerstone of international space law. The treaty was a great historic achievement, and it still is. The strategic – and at the same time, noble and peaceful – idea behind [it] was to prevent the extension of an arms race into outer space.”

Some might argue that those goals were attained and that there has been no arms race in space. But a conference [1] convened in Geneva last April by the United Nations Institute for Disarmament Research suggested that the situation is increasingly precarious, and indeed that military uses of space are well underway and likely to expand.

Paradoxically, the thawing of the Cold War is one reason why the OST is losing its restraining power. During a confrontration of two nuclear superpowers, it is rather easy to see (and game theory confirms) that cooperation on arms limitation is in the national interest. But as Sergey Batsanov, Director of the Geneva Office of the Pugwash group for peaceful uses of science, pointed out in the UN meeting, “after the end of the Cold War, disarmament and non-proliferation in their traditional forms could no longer be considered as vital instruments for maintaining the over-all status quo.” Batsanov suggests we are now in a transitional phase of geopolitics in which new power structures are emerging and there is in consequence a “crisis in traditional international institutions, and the erosion, or perhaps evolution, of norms of international law (such as the inviolability of borders and non-interference in another state’s internal affairs).”

It’s not hard to see what he is alluding to there. Certainly, it seems clear that the US plans for maintaining “space superiority” – the “freedom to attack as well as the freedom from attack” - does much to harm international efforts on demilitarization of space. The tensions created with Russia by US plans to site missile defence facilities in eastern Europe is just one example of that. James Armor, Director of the US National Security Space Office, indicates that, following the “emergence of space-enabled transitional warfare” using satellite reconnaissance in Operation Desert Storm in Iraq in 1991, military space capabilities have now become “seamlessly integrated into the overall US military structure”.

But it would be unwise and unfair to imply that the United States is a lone ‘rogue agent’. China has exhibited a clear display of military capability in space; as Xu Yansong of the National Space Administration of the People’s Republic of China explained at the UN conference, China’s space activities are aimed not only at “utilizing outer space for peaceful purposes” but “protecting China’s national interests and rights, and comprehensively building up the national strength” – which could be given any number of unsettling interpretations. Yet China, like Russia, has been supportive of international regulation of space activities, and it’s not clear how much of this muscle-flexing is meant to create a bargaining tool.

The real point is that the OST is an agreement forged in a different political climate from that of today. Its military commitments amount to a prohibition of nuclear weapons and other “weapons of mass destruction” in space, and the use of the Moon and other celestial bodies “exclusively for peaceful purposes.” That’s a long way from prohibiting all space weapons. As Kiran Nair of the Indian Air Force argued, “the OST made certain allowances for military uses of outer space [that] were exploited then, and are exploited now and ill continue to be so until a balanced agreement on the military utilization of outer space is arrived at.”

What’s more, there was no explicit framework in the OST for consultations, reviews and other interactions that would sustain the treaty and ensure its continued relevance. And as Batsanov says, now there are more players in the arena, and a wider variety of potential threats.

Both Russia and China have called for a new treaty, and earlier this year President Putin announced the draft of such a document. But we don’t necessarily need to ditch the OST and start anew. Indeed, the treaty has already been the launch pad for various other agreements, for example on liability for damage caused by space objects and on the rescue of astronauts. It makes sense to build on structures already in place.

The key to success, however, is to find a way of engaging all the major players. In that respect, the United States still seems the most recalcitrant: its latest National Space Policy, announced in October 2006, states that the OST is sufficient and that the US “will oppose the development of new legal regimes or other restrictions that seek to prohibit or limit US access to or use of space.” In other words, only nuclear space weaponry is to be considered explicitly out of bounds. Armor made the prevailing Hobbesian attitude clear at the Geneva meeting: “In my view, attempts to create regimes or enforcement norm that do not specifically include and build upon military capabilities are likely to be stillborn, sterile and ultimately frustrating efforts.” Whatever framework he envisages, it’s not going to look much like the European Union.

But it needn’t be a matter of persuading nations to be more friendly and less hawkish. There are strong arguments for why pure self-interest in terms of national security (not to mention national expenditure) would be served by the renunciation of all plans to militarize space – just as was the case in 1967. Rebecca Johnson of the Acronym Institute for Disarmament Diplomacy pointed out that after the experience in Iraq, US strategists are “coming to see that consolidating the security of existing assets is more crucial than pursuing the chimera of multi-tiered invulnerability.” The recent Chinese anti-satellite test, for from being a red flag to a bullish military, might be recognized as an indication that no one stays ahead in this race for long, and the US knows well that arms races are debilitating and expensive.

The danger with the current Sputnik celebrations is that they might cast the events in 1957 as pure history, which has now given us a world of Google Earth and the International Space Station. The fact is that Sputnik and its attendant space technologies reveal a firm link between the last world war, with its rocket factories manned by slaves and its culmination in the instant destruction of two cities, and the world we now inhabit. The OST is not merely a legacy of Sputnik but the only real international framework for the way we use space. Unless it can be given fresh life and relevance, we have no grounds for imagining that the military space race is over.

Reference
1. Celebrating the Space Age: 50 Years of Space Technology, 40 Years of the Outer Space Treaty (United Nations Institute for Disarmament Research, Geneva, 2007).

Wednesday, October 03, 2007

Yet more memory of water

This month’s issue of Chemistry World carries a letter from Martin Chaplin and Peter Fisher in response to my column discussing the special issue of Homeopathy on the ‘memory of water’. Mark Peplow asked if I wanted to respond, but I told him that he should regard publication of my response as strictly optional. In the event, he rightly chose to use the space to include another letter on the topic. So here for the record is Martin and Peter’s letter, and my response. I suppose I could be a little annoyed by the misrepresentation of what I said at the end of their letter, but I’m happy to regard it as miscomprehension.


From Martin Chaplin and Peter Fisher


We put together the ‘Memory of water’ issue of the journal Homeopathy, the subject of Philip Ball’s recent column (Chemistry World, September 2007, p38), to show the current state of play. It contained all the current scientific views representing the different experimental and theoretical approaches to the ‘memory of water’ phenomena. Some may be important and others less so, but now the different areas of the field can be fairly judged. The papers mostly demonstrated the similar theme that water preparations may have unexpected properties, contain unexpected solutes and show unexpected changes with time; all very worthy of investigation. Although not the main purpose of the papers, we show the problems as much as the potential of these changed properties in relation to homeopathy.

Ball skirts over the unexpected experimental findings that he finds ‘puzzling’, so ignoring the very heart of the phenomena we are investigating and misinterpreting the issue. He backs up his argument with statements concerning pure water and silicate solutions that are clearly not relevant to the present discussion. Also, he uses Irving-Langmuir to prop up his argument. This is fitting as Langmuir dismissed the Jones-Ray effect (http://www.lsbu.ac.uk/water/explan5.html#JR), whereby the surface tension of water is now known to be reduced by low concentrations of some ions, as this disagreed with his own theories. Finally Ball finishes with the amazing view that he knows the structure of water in such solutions with great confidence; I wish he would share that knowledge with the rest of us.

M F Chaplin CChem FRSC, London, UK

P Fisher, Editor,Homeopathy, Luton, UK



Response from Philip Ball

I have discussed elsewhere some of the experimental papers to which Chaplin and Fisher refer (see http://www.nature.com/news/2007/070806/full/070806-6.html and www.philipball.blogspot.com). Some of those observations are intriguing, but each raises its own unique set of questions and concerns, and they couldn’t possibly all be discussed in my column. Langmuir’s ideas feature nowhere in my argument; I simply point out that he coined the term ‘pathological science.’ If the issues I raise about silicate self-organization are not relevant to the discussion, why do Anick and Ives mention them in their paper? And I never stated that I or anyone else knows the structure of water or aqueous solutions with great confidence; I merely said that there are some things we do know with confidence about water’s molecular-scale structure (such as the timescale of hydrogen-bond making and breaking in the pure liquid), and they should not be ignored.

Monday, October 01, 2007

What’s God got to do with it

There’s a curious article in the September issue of the New Humanist by Yves Gingras, a historian and sociologist of science at the University of Quebec. Gingras is unhappy that scientists are using references to God to sell their science (or rather, their books), thereby “wrap[ping] modern scientific discoveries in an illusory shroud that insinuates a link between cutting-edge science and solutions to the mysteries of life, the origins of the universe and spirituality.” But who are these unscrupulous bounders? Well… Paul Davies, and… and Paul Davies, and… ah, and Frank Tipler. Well yes, Tipler. My colleagues and I decided recently that we should introduce the notion of the Tipler Point, being the point beyond which scientists lose the plot and start rambling about the soul/immortality/parallels between physics and Buddhism. A Nobel prize is apt to take you several notches closer to the Tipler Point, though clearly it’s not essential. And such mention of Buddhism brings us to Fritjof Capra, and if we’re going to admit him to the ranks of ‘scientists’ who flirt with mysticism then the game is over and we might as well bring in Carl Jung and Rudolf Steiner.

Gingras suggests that the anthropic principle is “bizarre and clearly unscientific”, and that it has affinities with intelligent design. Now, I’m no fan of the anthropic principle (see here), but I will concede that it is actually an attempt to do the very opposite of what intelligent design proposes – to obviate the need to interpret the incredible fine-tuning of the physical universe as evidence of design. The fact is that this fine-tuning is one of the most puzzling issues in modern physics, and if I were a Christian of the sort who believes in a Creator (not all have that materialist outlook), I’d seize on this as a pretty strong indication that my beliefs are on to something. The Templeton Foundation, another of Gingras’s targets, has hosted some thoughtful meetings on the theme of fine-tuning, and while I’m agnostic about the value and/or motives of the Templeton Foundation, I don’t obviously see a need to knock them for raising the question.

Paul Davies has indeed hit a lucrative theme in exploring theological angles of modern cosmology, but he does so in a measured and interesting way in which I don’t at all recognize Gingras’s description of “X-files science” or an “oscillation between science and the paranormal.” Frankly, I’m not sure Gingras is on top of his subject – when, as I expected resignedly, he fishes out Stephen Hawking’s famous “mind of God” allusion, he seems to see it as a serious suggestion, and not simply as an attempt by an excellent scientist but indifferent writer to inject a bit of pizzazz into his text. Hawking’s reference is obviously theologically naïve, and gains supposed gravitas only because of the oracular status that Hawking has, for rather disturbing reasons, been accorded.

Still, I suppose I will also be deemed guilty of peddling religious pseudo-science for daring to look, in my next book, at the theological origins of science in the twelfth century…

Friday, September 28, 2007


Space experiments should be a cheap shot
[This is the pre-edited version of my latest article for muse@nature.com - with some added comment at the end.]

We rarely learn anything Earth-shaking from space labs, which is why inexpensive missions like Foton-M3 are the way to go.

Space experiments have rarely seemed as much fun as they do on the European Space Agency’s Foton-M3 mission, which blasted off two weeks ago from the Russian launch site at Baikonur in Kazakhstan for a 12-day spell in low-Earth orbit. Among the experiments in the 400-kg payload were an exploration of capsule delivery to Earth using a 30-km space tether, a study of how human balance works and an investigation of whether organic matter in space rocks can withstand the heat of orbital re-entry so that life could be transferred between planets, as posited in the ‘panspermia’ hypothesis, by sticking a chunk of Scottish rock onto the spacecraft’s side.

None of these experiments seems likely be itself to lead to any major new discoveries or technological breakthroughs. And none can be considered terribly urgent – the balance study, which looks at how a balance organ called the otolith grows in larval fish in zero gravity, has been on the shelf for years, the first attempt being one of the minor casualties of the ill-fated Columbia space shuttle mission in early 2003.

But it would be churlish to criticize Foton-M3 for the incremental nature of its science. Most scientific research in general is like that, and the roster of experiments is not only impressively long for such a relatively cheap mission but also appealingly diverse, spanning subjects from microbiology to geoscience to condensed-matter physics.

What’s more, the tether experiment has arisen from a project called the Young Engineers’ Satellite 2 (YES2), involving more than 450 students throughout Europe. The aim is to use a tether to slow down an object falling back into Earth’s gravity from a spacecraft so that it continues falling instead of being captured in orbit. This could offer a cheap way of delivering payloads from space to Earth.

Admittedly the experiment seems not to have quite worked out as planned, because apparently not all the tether unreeled. And the notion of finding a cheap postal method for the indefensibly expensive white elephant known as the International Space Station, which has so far yielded very little worth delivering in the first place, is rather hard to swallow.

But as a way to engage students in serious space research that poses interesting scientific and technological questions and might conceivably find uses in the future, YES2 can’t be faulted.

Foton-M3 does evoke a degree of dèja-vu – how many earlier space experiments have claimed to be “improving our understanding of protein structure by growing protein crystals in weightlessness”, or learning about loss of bone mass in astronauts? But there’s bound to be some duds in over 40 experiments.

What’s curious about some of these is that they threaten to undermine their own justification. If we can design robotic instruments to look at the growth of bone or tissue cells so that we can predict how astronauts might fare on long-term space missions, can we not design robots to replace those very astronauts? Preparing the ground for human space exploration demands such advances in automation that, by the time we’re ready, we’ll have run out of good scientific reasons for it. There may be non-scientific arguments, such as the educational and inspirational value, but a mission like Foton-M3 at least raises doubts about why there is any reason for near-Earth manned spaceflight.

A report by the UK Royal Astronomical Society (RAS) Commission of the Scientific Case for Human Space Exploration, published in 2005, seems to challenge such scepticism. It claimed, for example, that “the capabilities of robotic spacecraft will fall well short of those of human explorers for the foreseeable future.”

But what this turns out to amount to is a statement of the obvious: robots are nowhere near achieving human-like intelligence and decision-making capabilities. There’s no doubt that having humans on site will permit more flexible, faster and more thoughtful responses to unexpected circumstances in lunar or planetary exploration. But since one can probably have ten robotic missions for the price of one manned (and since it might soon take as little as three months to get to Mars), that isn’t obviously a clinching argument, especially when you think about the cost of failure – the success rate for Mars missions is so far not much more than 1 in 4. And robots are, in many ways, considerably more robust than humans.

The RAS report also claimed that “there are benefits for medical science to be gained from studying the human physiological response to low and zero gravity [and] to the effects of radiation.” This claim drew heavily on a letter from the UK Space Biomedicine Group (UKSBG), who one might imagine to be rather disposed to the idea in the first place. They claim that studying bone demineralization in micro- and zero gravity “could dramatically improve the understanding and treatment of osteoporosis.”

That’s why space experiments like those on board Foton-M3 are relevant to the debate. One experiment on the mission looks at precisely this question of bone mass loss using bovine bone; another involves bone-forming and bone-degrading cells cultured in vitro. In other words, one of the key putative health spinoffs of human spaceflight, according to the RAS Commission, is already being studied in cheap unmanned missions. It is conceivable that we would learn something (the UKSBG doesn’t specify what) from live humans that we would not from dead cows, or from live mice or human cell cultures. But should that unknown increment weigh heavily on the scales that the RAS were seeking to balance?

The considerations raised by the RAS report also bear on the question of why it is that such experiments have enjoyed sustained support in the past despite being pretty uninspiring and lacking in real technological payoff. If we think (rightly or wrongly) that it is intrinsically interesting to blast people into space, we’ll tend to feel that way about the stuff they do there too (so that a golf drive in space makes headlines).

Thus, many space experiments, such as the recent demonstration that Salmonella bacteria on the space shuttle Atlantis were more virulent in zero gravity [1], gain interest not because of the results in themselves but because of the very fact of their having been obtained in space. That particular result was already known from microgravity experiments on Earth, and in any event much of the interest centred on whether it means astronauts will suffer more from germs. The glamour that seems to attach to space experiments almost invariably distorts the import of what they find, all the more so because they are used as their own justification: “look at what space experiments can tell us about stuff that happens in space!”

As a result, Foton-M3 provides a nice illustration of proper cost-benefit thinking. The ‘panspermia’ tests, say, operated by a team from the University of Aberdeen, will at best provide a useful addition to a wealth of previous studies on space- and impact-resistance of organic matter and living organisms. A study of temperature and concentration fluctuations in fluids provides a nice verification of a result that was generally expected theoretical grounds – it is the kind of experiment that would be undertaken without hesitation if it could be done in a lab, but which would certainly not warrant its own dedicated space mission.

In other words, when Foton-M3 plummeted back down to Earth near the Russian/Kazakh border on Wednesday [26 September], it should have blown a big hole in starry-eyed visions of space experimentation. This is how it should really be done: modest but intrinsically interesting investigations, realised at a modest cost, and performed by robots.

Reference

1. Wilson, J. W. et al. Proc.Natl Acad. Sci. USA online early edition, doi:10.1073/pnas.0707 (2007).

The more I think about it, the worse the RAS report seems. When it comes to space exploration generally, they do a fair job of taking into consideration the fact that robots can be guided remotely by human intelligence, and don’t need to be autonomous decision-makers. But even this was rather specious in its use of deep-sea engineering as a means of comparison – getting humans to the sea floor, and the hazards they face there, hardly compares with sending them to Mars. When, however, the discussion turned to biomedical spinoffs, the RAS Commission seemed to forget all about doing things robotically – they simply pleaded lack of expertise, which meant they seemingly relied entirely on the testimonies of the UKSBG and human spaceflight advocate Kevin Fong. At no point do they seem to ask whether the biomedical benefits proposed might be obtained equally in unmanned missions. As far as osteoporosis goes, for example, the question is not whether manned spaceflight might tell us something about it but whether:
1. there are critical questions about the condition that can be answered only by micro- or zero-gravity studies; and
2. these questions can only be answered by studying live human subjects and not animals or cell cultures.
The UKSBG point to no such specific questions, and I rather doubt that they could. (Certainly, it is not as though we need to study astronauts in order to monitor human bone mass loss in vivo.) If there are not good answers to these points, the RAS should not be using this line as a reason for human space exploration (as opposed to stuff you might as well do if you’re going up there anyway).
It’s the same story for the work on Salmonella that I mention. There are vague promises of improved understanding of the emergence of virulent strains on Earth, but no indication of why a space experiment will really tell you much more in this regard than a terrestrial simulation of zero G. Much of the interest seems to centre on the question of whether astronauts would face nastier bugs, which of course becomes an issue only if you put them up there in the first place. This is the kind of fuzzy thinking that defenders of human space exploration get away with all the time.

Wednesday, September 26, 2007

Hybrids and helium
[This is the pre-edited version of my Lab Report column for the October issue of Prospect.]

It’s not obvious that, when the Human Fertilisation and Embryology Authority was established in 1991, anyone involved had much inkling of the murky waters it would be required to patrol. The HFEA was envisaged primarily as a body for regulating assisted conception, and so it seemed sensible to give it regulatory powers over human embryo research more generally. Sixteen years later, the HFEA is having to pronounce on issues that have little bearing on fertility and conception, but instead concerns biological research that some say is blurring the boundaries of what it means to be human.

So far, the HFEA has remained commendably aloof from the ill-founded fears that this research attracts. Its latest permissive ruling on the creation of human-animal cells is the outcome of sober and informed consideration of a sort that still threatens to elude the British government. It belies (in the UK, at least) the fashionable belief that Enlightenment ideals are in eclipse.

There are many different ways human and non-human components might be mixed in embryos. Some research requires human genetic material to be put into animal cells – for example, to create human embryonic stem cells without reliance on a very limited supply of human eggs. There are also arguments for putting animal genes into human cells, which could offer new ways to study the early stages of human development, and might even help assess embryo quality for assisted conception.

Certainly, there are dangers. For example, eviscerating an animal cell nucleus (where most DNA is housed) to make way for a human genome does not remove all the host’s genetic material. Such transfers, which produce so-called cytoplasmic hybrid (‘cybrid’) cells might, if used to make stem cells for medical implantation, run the risk of introducing animal diseases into human populations. Recent findings that genomes can be altered by ‘back-transfer’ from non-genetic material adds to the uncertainties.

But no one is intending at this stage to use cybrids for stem-cell treatments; they are strictly a research tool. The HFEA has decided that there is no ‘fundamental reason’ to prohibit them – recognizing, it seems, that protests about human dignity and unnaturalness impose misplaced criteria. It stresses that the ruling is not a universal green light, however, and that licensing will be made on a case-by-case basis – as they surely should be. The first such applications are already being considered, and are likely to be approved.

The ruling says nothing yet about other human-animal fusions, such as embryos with mixtures of human and animal cells (true chimeras) or hybrids made by fertilization of eggs with sperm of another species. These too may be useful in research, but carry a higher yuk factor. On current form, it seems we can count on the HFEA not to succumb to squeamishness, panic or the mendacious rhetoric of the slippery slope.

*****

Was it vanity or bravery that prompted Craig Venter to allow his complete genome to be sequenced and made public? That probably depends on how you feel about Venter, whose company Celera controversially provided the privatized competition to the international Human Genome Project. Both those efforts constructed a composite genome from the DNA of several anonymous donors, and analysed only one of each pair of the 23 human chromosomes.

In contrast, Venter’s team has decoded both of his chromosomes, revealing the different versions of genes acquired from each parent. It is these variants, along with the way each is controlled within the genome and how they interact with the environment, that ultimately determines our physical characteristics. The analysis reveals other sources of difference between chromosomal ‘duplicates’, such as bits of genes that have bits inserted or cut out. This is, you might say, a study of how much we differ from ourselves – and it should help to undermine the simplistic notion that we’re each built from a single instruction manual that is merely read again and again from conception to the grave.

Venter bares all in a paper in the free-access electronic journal PLoS Biology, joining Jim Watson, a co-discoverer of the structure of DNA, as one of the first individuals to have had his personal genome sequenced. Some have complained that this ‘celebrity’ sequencing sends out the message that personalized genomics will be reserved for the rich and privileged. But no one yet really knows whether such knowledge will prove a benefit or a burden – Venter has discovered a possible genetic propensity towards Alzheimer’s and cardiovascular diseases. The legal and ethical aspects of access to the information are a minefield. Venter himself says that his motive is partly to stimulate efforts to make sequencing cheaper. But right now, he has become in one sense the best-known man on the planet.

*****

The moon has always been a source of myth, and now we have some modern ones. Many people will swear blind, without the slightest justification, that the Apollo missions gave us Teflon and the instant fruit drink Tang. New calls for a moon base are routinely supported now with the claim that we can mine the lunar surface for nuclear-fusion fuel in the form of helium-3, a rare commodity on Earth. BBC’s Horizon bought the idea, and it’s been paraded in front of the US House of Representatives. But as physicist Frank Close pointed out recently, there is no sound basis to it. None of the large fusion projects uses helium-3 at all, and the suggestion that it would be a cleaner fuel simply doesn’t work, at least without a total reactor redesign. That’s not even to mention the cost of it all. But no straw is too flimsy that advocates of human spaceflight will fail to grasp it.