Sunday, February 24, 2008

Engineering biology, with Drew Endy



There's a fascinating essay by/interview with Drew Endy on the Edge website, which appears to be the latest in a series to have emerged from an event they organised last August. I've written about Endy in the past, and he features prominently in the final chapters of Genesis Machines; indeed, I wish I'd had such an illuminating transcript available when I wrote the book.

Endy is an Assistant Professor of Biological Engineering at MIT, and one of the leading figures in synthetic biology. In one particular paragraph, he captures the excitement of this emerging new discipline:

"Programming DNA is more cool, it's more appealing, it's more powerful than silicon. You have an actual living, reproducing machine; it's nanotechnology that works. It's not some Drexlarian (Eric Drexler) fantasy. And we get to program it. And it's actually a pretty cheap technology. You don't need a FAB Lab like you need for silicon wafers. You grow some stuff up in sugar water with a little bit of nutrients. My read on the world is that there is tremendous pressure that's just started to be revealed around what heretofore has been extraordinarily limited access to biotechnology."

Friday, February 15, 2008

Insect lab

I've spend all week running simulation experiments for our ongoing work on ant-based computing, so when I came across the Insect Lab it seemed strangely appropriate.

The artist takes real (dead) insects and customizes their bodies with parts taken from watches and other mechanical devices, to create "cybernetic sculptures".

I'd like to see him do an ant, though... Which train of thought lead me circuitously to Bill Bailey performing his wonderful song Insect Nation (if you just want the lyrics, they're here).

Friday, February 08, 2008

Dr Who

A wonderful present arrived in today's post, courtesy of our equally wonderful friend Eventhia; a signed photograph of Tom Baker! He is, of course, best known for playing the fourth Dr Who, but is probably most familiar to a younger generation as the narrator of Little Britain (and even the delightfully barmy Stagecoach adverts).

Most people of sound mind would name Baker as the best ever Dr Who, despite ludicrous polls to the contrary. A case can be made that the choice of favourite depends on which Doctor a person grew up with, and since Baker's tenure extended from 1974-1981, I would certainly agree.

Anyway, he recently did a signing in Norwich, attended by our friends Kris and Eventhia. They very kindly got Tom to sign the photo "For Martyn," (eventually, I think he had it down as "Martin", and you can see where he's corrected it at E's prompting) "Genetically yours, Tom Baker"

Sigh!

Tuesday, February 05, 2008

Biological complexity: from molecules to systems


I'm delighted to have been invited to speak at an event titled "Biological complexity: from molecules to systems", to be held at University College London from 12-13 June this year. The meeting is sponsored by both UCL and the Weizmann Institute of Science in Israel, and will feature speakers from the fields of immunology, computer science, mathematics, biological chemistry, molecular genetics and bioinformatics. I'll try my best to summarize below the research interests of the other invited speakers (but apologies to anyone whose work I misrepresent!)

Stephen Emmott from Microsoft Research in Cambridge will give the keynote address. Stephen is the founder and Director of Microsoft's European Science Programme, and was the driving force behind the influential Towards 2020 Science project and report.

Representing Israeli activity, Nir Friedman works in computational biology, and recently published a paper arguing that gene duplication may drive the "modularisation" of functional genetic networks (that is, genetic networks that are relatively self-contained, and which perform a specific task).

David Harel is a celebrated computer scientist, having carried out important work in logic, software engineering and computability theory. As a student, I often referred to his award-winning book Algorithmics: The Spirit of Computing, and he is currently working on topics that include the modelling and analysis of biological systems (eg. the nematode worm) and the synthesis and communication of smell.

Shmuel Pietrokovski works in bioinformatics, with particular interest in inteins (protein introns); "selfish" DNA elements that are converted into proteins together with their hosts.

Yitzhak Pilpel's lab takes a systems-level approach to how genes are regulated: "By applying genome wide computational approaches, backed-up by in house laboratory experiments, [the lab] devotes itself to both establishing an in-depth understanding of the different processes controlling gene expression, and to understand[ing] how these processes are orchestrated to establish robustness of the regulatory code."

Gideon Schreiber studies the precise nature of protein-protein interactions and the implications these have for complex biological processes.

Eran Segal is a computer scientist (predominantly) working in computational biology, who has recently reported some fascinating work on a "higher level" genetic code, as well as research on predicting expression patterns from their regulatory sequences in fruit flies.

I've already written at some length about Ehud Shapiro (also here); his recent work has centred on the construction of biological computing devices (known as automata) using DNA molecules and enzymes.

Yoav Soen's group is "using embryonic stem cells models to study how different layers of regulation interact to specify morphogenetic decisions, how these decisions are shaped by interactions between emerging precursors and how they are coordinated across a developing embryonic tissue." He has also worked with a colleague of mine, Netta Cohen at Leeds.

Representing activities in the UK, we have Cyrus Chothia from the Laboratory of Molecular Biology at Cambridge, who studies the "nature of the protein repertoires in different organisms and the molecular mechanisms that have produced these differences."

Jasmin Fisher is leading the new Executable Biology Group at Microsoft Research, and is primarily interested in systems/computational biology.

Mike Hoffman and Ewan Birney are at the European Bioinformatics Institute (EBI) in Cambridge, where Birney leads the EBI contribution to Ensembl. There's a transcript of an interview with him here.

Jaroslav Stark is the Director of the Centre for Integrative Systems Biology at Imperial College. He was recently interviewed for a piece on systems biology on BBC Radio 4's The Material World.

Michael Sternberg heads the Structural Bioinformatics Group and the Imperial College Centre for Bioinformatics. He was previously the head of biomolecular modelling at the Imperial Cancer Research Fund now part of Cancer Research UK.

Perdita Stevens is at Edinburgh, where she works on software engineering and theoretical computer science (with a growing interest in modelling viral infection).

The meeting organisers are particularly keen to encourage the participation of young researchers, and the registration fee for this two-day event is a very reasonable 50 pounds (30 for students). To register and for further information, please contact Michelle Jacobs at Weizmann UK at post@weizmann.org.uk or on 020 7424 6860. Attendance will be limited to 180 delegates.

Thursday, January 31, 2008

The human genome, in book form

I've just been sorting through some old files, and came across this picture (click for a larger version), which I took on my last visit to London. I had some free time before a meeting at the BT Tower, so I popped into the Wellcome Collection.

Quite apart from the fact that the Wellcome Trust spends around 400 million pounds a year on biomedical research, I have a personal affinity with the trust, since my shortlisted entry to their Book Prize (which was won that year by Chris McManus' Right Hand, Left Hand) was picked up by Toby Mundy and eventually evolved into Genesis Machines (pictured below in the Wellcome Collection bookshop, in a nice circular turn of fate).

One of the most striking exhibits they have (alongside a sample of droppings from Dolly the sheep) is the human genome, printed out in book form. As I've said before, we're rank amateurs compared to nature in the information storage stakes (of course, reading and writing data quickly is another matter...) Since the size of the human genome is estimated at around 3 billion genetic letters (taken from the set {A, G, C, T}), then (assuming that one byte is used to store each letter), each cell with a nucleus (that is, every one except red blood cells and the like) contains 3 Gigabytes of genetic "memory". Of course, we don't need an entire byte (8 bits) to store a quarternary (base 4) value, so we could compress this figure by three quarters, and cells actually contain two genomic copies, but I don't want to over-complicate things...

The fact of the matter is that our genome is large: in the past, I've compared it, if printed out in full, to 200 copies of the Manhattan telephone book. This analogy was arrived at by some back-of-an-envelope calculations, and I don't think I really understood its significance until I visited the Wellcome Collection.

There, in a corner of one of the galleries, stand a single set of white shelves, almost 5 metres by 2 metres, containing 120 hefty volumes. One of them stands open, and a closer inspection reveals page upon page of genetic data, rows and rows of A's, G's, C's and T's tightly-set in 4.5-point text.

The sheer scale of the artifact is mind-blowing, both as an illustration of nature's nanotechnology, but also as a reminder of how far we have to go in terms of beginning to piece together even a small fraction of the human circuitry.

Tuesday, January 29, 2008

Sex with robots


One of the sure signs of impending middle age, especially in a university town, is when people stop handing you flyers. There was a time not so long ago when I could nip out of the office for a sandwich and come back burdened with glossy adverts for progressive house, 2-for-1 vodka shots and foam. But no longer. Now, the bright young things actively avoid me as my thirty-something, corduroy-clad figure shambles into view. The kinder ones simply pretend not to see me.

So imagine my delight when, walking down Oxford Road this afternoon after picking up some grapes, I was handed a flyer. And one offering sex with robots, to boot! Impressed by the targeted precision of whoever was marketing such an opportunity, I was about to kick my heels when I realised that it was actually advertising a club night in Manchester. Now, as a long-time veteran of nights such as House of God and Voodoo, I might have been interested...ooh, ten years ago, but with a responsible job and a young daughter, "'avin it large" now means having that third shot of espresso in my cappucino.

Which is a really cheap and tenuous way of introducing a new play that I think you should go and see. Involution is by a new author, Rachel Welch, and deals with many urgent contemporary themes, such as genetic engineering, religion and the human self-image. One of the plot threads concerns "cybernetic companionship", so I'll leave it to you to make the link...

Alfie Talman, a member of the production team and cast (and, coincidentally, a fellow Ipswich fan) enjoyed Genesis Machines, and thought I might be interested in the play. It's on from February 21st to March 15th at the Pacific Playhouse in London, and there are more details here (and here).

Monday, January 14, 2008

"I'm burning, I'm burning!..."

Although every side in an argument tends to have its own complement of fools, the idiocy exhibited by fundamentalist Christians, in debates over evolution or the origin of the universe, often takes us into the realm of comedy.

Take this example, lifted from a list of fundie "bloopers":

"Everyone knows scientists insist on using complex terminology to make it harder for True Christians to refute their claims.

Deoxyribonucleic Acid, for example... sounds impressive, right? But have you ever seen what happens if you put something in acid? It dissolves! If we had all this acid in our cells, we'd all dissolve! So much for the Theory of Evolution, Check MATE!"

The full amusing-and-yet-slightly-scary list is here.

Tuesday, January 08, 2008

Genesis Machines in Japan



Readers in Japan may be interested in the forthcoming edition of Genesis Machines, which is now available for preorder. It's been translated by Kyoko Gibbons, and is out on the 17th of this month.

Friday, August 31, 2007

My Edinburgh talk

I had a wonderful time at the Edinburgh Book Festival over the weekend; a full venue and books to sign afterwards makes for a happy author! Here is a lightly edited version of what I had to say.

In 1959, a great personal hero of mine, the Nobel Prize-winning physicist Richard Feynman gave a visionary talk entitled “There's Plenty of Room at the Bottom”. In his speech, Feynman outlined the possibility of individual molecules, even individual atoms making up the component parts of computers in the future. Remember, this was back when computers filled entire rooms, and were tended by teams of lab-coated technicians, so the idea that you could compute with individual molecules was pretty outlandish. I was struck by a quotation in Oliver's book, attributed to the microbiologist A. J. Kluyver, who said, over fifty years ago, that “The most fundamental character of the living state is the occurrence in parts of the cell of a continuous and directed movement of electrons.” At their most basic, level computers work in exactly the same way; by funnelling electrons around silicon circuits, so I think this hints at the linkages between biology and computers that are only now coming to fruition.

Indeed, it wasn't until 1994 that someone demonstrated, for the first time, the feasibility of building computers from molecular-scale bits. Feynman's vision had waited, not only for the technology to catch up, but for a person with the required breadth of understanding and the will to try something slightly bizarre. That person was Len Adleman, who won the computer science equivalent of the Nobel Prize for his role in the development of the encryption scheme that protects our financial details whenever we buy something on the Internet. Len has always had an interest in biology; when one of his students showed him a program that could take over other programs and force them to replicate it, Len said “Hmmm.... that looks very much like how a virus behaves.” The student was Fred Cohen, author of the first ever computer virus, and Len's term stuck. (Update, 2/9/07: Cohen made the first reference to a "computer virus" in an academic article, but did not write the first virus).

One night in the early 90's, Len was lying in bed reading a classic molecular biology textbook. He came across the section describing a particular enzyme inside the cell that reads and copies DNA, and he was struck by its similarity with an abstract device in computer science known as the Turing Machine. By bringing together two seemingly disparate concepts, Adleman knew at once that, in his own words, “Geez, these things could compute.”

He found a lab at the University of Southern California, where he is a professor, and got down to building a molecular computer. He knew that DNA, the molecule of life that contains the instructions needed to build every organism on the planet, from a slug to.... John Redwood can be thought of as a series of characters from the set A, G, C and T, each character being the first letter of the name of a particular chemical. The title of the film Gattaca, which considers a dystopian future in which genetic discrimination defines a society, is simply a string of characters from the alphabet A, G, C and T.

As Oliver highlights in his own book, molecular biology has always been about the transformation of information, usually inside the living cell. This information is coded in the AGCT sequences of genes and in the proteins that these genes represent. Adleman immediately saw how this mechanism could be harnessed, not to represent proteins, but to store digital data, just like a computer encodes a file as a long sequence of zeroes and ones.

Adleman decided to use this fact to solve a small computational problem. Some of you might have heard of the Travelling Salesman Problem, and Adleman's was a variant of that; given a set of cities connected by flights, does there exist a sequence of flights that starts and ends at particular cities, and which visits every other city only once? This problem is easy to describe, but fiendishly difficult to solve for an even relatively small number of cities. This inherent difficulty is what made the problem interesting in Adleman's eyes, “interesting” being, to a mathematician, a synonym for “hard”.

Len decided to build his computer using the simplest possible algorithm; generate all possible answers (right or wrong), and then throw away the wrong ones. He would build a molecular haystack of answers, and then throw away huge swathes of hay encoding bad answers until he was left with the needle encoding the correct solution (of which there may be just a single copy). For Adleman, the key to his approach was that you can make DNA in the laboratory. A machine the size of a microwave oven will sit in a lab connected to four pots, each containing either A, G, C or T. Type in the sequence you require, and the machine gets to work, threading the letters together like molecular beads on a necklace, making trillions of copies of your desired sequence.

Adleman ordered DNA strands representing each city and each flight for his particular problem. Because DNA sticks together to form the double helix in a very well-defined way, he chose his sequences carefully, such that city and flight strands would glue together like Lego blocks to form long chains, each chain encoding a sequence of flights. Because of the sheer numbers involved, he was pretty sure that a chain encoding the single correct answer would self-assemble. The problem then was to get it out. In a way, Len had built a molecular memory, containing a huge file of lines of text. What he then had to do was sort the file, removing lines that were too long or too short, that started or ended with the wrong words, or which contained duplication. He used various standard lab techniques to achieve this, and, after about a week of molecular cutting and sorting, he was left with the correct solution to his problem.

The example that he solved could be figured out in a minute by a bright 10-year-old using a pen and paper. But that wasn't the point. Adleman had realised, for the first time, Feynman's vision of computing using molecules. After he published his paper, there was a flood of interest in the new field of DNA computing, a tide on which I was personally carried. The potential benefits were huge, since we can fit a vast amount of data into a very small volume of DNA. If you consider that every cell with a nucleus in your body contains a copy of your genome - 3 gigabytes of data, corresponding to 200 copies of the Manhattan phone book – you begin to understand just how advanced nature is in terms of information compression. Suddenly my 4 gig iPod nano doesn't look quite so impressive.

After a few years, though, people began to wonder if molecular computing would ever be used for anything important. They were looking for the “killer application”, the thing that people are willing to pay serious money for, like the spreadsheet, that persuaded small businesses to buy their first ever computer. The fundamental issue with Adleman's approach is tied to the difficulty of the problem; as the number of cities grows only slightly, the amount of DNA required to store all possible sequences of flights grows much more quickly; a small increase in the number of cities quickly leads to a requirement for bathtubs full of DNA, which is enough to induce hysterical laughter in even the sanest biologist. Indeed, it was estimated that if Len's algorithm were to be applied to a map with 200 cities in it, the DNA memory required to store all possible routes would weigh more than the Earth.

It would appear that DNA computing has reached the end of the line, if we are to insist on applying it to computational problems in a head-to-head battle against traditional silicon-based computers. Let's be straight, you're never going to be able to go into PC World and buy a DNA-based computer any time soon. When DNA computing first emerged as a discipline, I was dismayed to see a rash of papers making claims that within a few years we'd be cracking military codes using DNA computers and building artificial molecular memories vastly larger than the human brain. I was dismayed because I knew what had happened 30 years previously to the embryonic field of artificial intelligence. Again, hubristic claims were made for their discipline by the young Turks, ranging from personal robot butlers to automated international diplomacy. When the promised benefits failed to materialise, AI suffered a savage backlash in terms of credibility and funding, from which it is only just beginning to recover. I was very keen to avoid the same thing happening to molecular computing, but I, like many others, knew that we needed to look beyond simply using DNA as a tiny memory storage device.

The next key breakthrough was in realising that, far from being simply a very small storage medium that can be manipulated in a test tube, within its natural environment – the cell – DNA carries meaning. As the novelist Richard Powers observes in The Gold Bug Variations, “The punched tape running along the inner seam of the double helix is much more than a repository of enzyme stencils. It packs itself with regulators, suppressors, promoters, case-statements, if-thens.” Computational structures, that is. DNA encodes a program that controls its own execution. DNA, and the cellular machinery that operates on it, pre-dates electronic computers by billions of years. By re-programming the code of life, we may finally be able to take full advantage of the wonderful opportunities offered by biological wetware.

As Oliver observes in his book, “The world is not just a set of places. It is also a set of processes.” This nicely illustrates the shift in thinking that has occurred in the last few years since the human genome has been sequenced. The notion of a human “blueprint” is outdated a useless. A blueprint encodes specific locational information for the various components of whatever it's intended to represent, whether it be a car or a skyscraper. Nowhere in the human genome will you find a section that reads “place two ears, on on either side of head” or “note to self: must fix design for appendix.” Instead, genes talk to one another, turning each other (and often themselves) on and off in a complex molecular dance. The genome is an electrician's worst nightmare, a tangle of wiring and switches, where turning down a dimmer switch in Hull can switch off the Manhattan underground system.

The human genome project (and the many other projects that are sequencing other organisms, from the orang-utan to the onion) is effectively generating a biological “parts catalogue”; a list of well-understood genes, whose behaviour we can predict in particular circumstances. This is the reductionist way of doing science; break things down, in a top-down fashion, into smaller and smaller parts, through a series of levels of description (for example, organism, molecule, atom). The epitome of this approach is the very well-funded physicists smashing together bits of nature in their accelerators in an attempt to discover what some call the God Particle.

Of course, smashing together two cats and seeing what flies off is only going to give you a limited understanding of how cats work, and it'll probably annoy the cats, so the reductionist approach is of limited use to biologists. Systems biology has emerged in recent years to address this, by integrating information from many different levels of complexity. By studying how different biological components interact, rather then just looking at their structure, as before, systems biologists try to understand biological systems from the bottom up.

An even more recent extension of systems biology is synthetic biology. When a chemist discovers a new compound, the first thing they do is break it down into bits, and the next thing they do it try to synthesise it. As Richard Feynman said just before his death, “What I cannot build I cannot understand.” Synthetic biologists play, not with chemicals, but with the genetic components being placed daily in the catalogue. It's where top down meets bottom up – break things down into their genetic parts, and then put them back together in new and interesting ways. By stripping down and rebuilding microbial machines, synthetic biologists hope to better understand their basic biology, as well as getting them to do weird and wonderful things. It's the ultimate scrapheap challenge.

If we told someone in the field of nanotechnology that we had a man-made device that doesn't need batteries, can move around, talk to its friends and even make copies of itself – and all this in a case the size of a bacterium – they would sell their grandmother for a glimpse. Of course, we already have such devices available to us, but we know them better as microbes. Biology is the nanotechnology that works. By modelling and building new genetic circuits, synthetic biologists are ushering in a new era of biological engineering, where microbial devices are built to solve very pressing problems.

As Oliver notes towards the end of his book, the planet is facing a very real energy crisis. One team is therefore trying to build a microbe to produce hydrogen. Another massive problem facing the developing world is that of arsenic contamination in drinking water. A team here in Edinburgh, made up mainly of undergraduates, has built a bacterial sensor that can quickly and easily monitor arsenic concentrations from a well sample, to within safe tolerances. Jay Keasling, a colleague in California has recently been awarded 43 million dollars by the Bill and Melinda Gates Foundation to persuade E. coli to make substances that are alien to them, but which provide the raw ingredients for antimalarial drugs. The drug is found naturally in the wormwood plant, but it's not cheap – providing it to 70 per cent of the malaria victims in Africa would cost $1 billion, and they can be repeatedly infected. It's been estimated that drug companies would need to cover the entire state of Rhode Island in order to grow enough wormwood, so Keasling wants to produce it in vats, eventually at half the cost.

There are, of course, safety issues with synthetic biology, as well as legal and ethical considerations. I worry that people have this idea that the bugs we use are snarling microbes that have to be physically restrained for fear of them erupting from a Petri dish into the face of an unfortunate researcher, like something from the Alien movies. In reality, the bacteria used in synthetic biology experiments are docile creatures, pathetic even, the crack addicts of the microbial world. They have to be nurtured and cossetted, fed a very specific nutrient brew. Like some academics, they wouldn't last two minutes in the real world. Of course, nature has a habit of weeding out the weak and encouraging the fit, so we still have to be very careful and build in as many safeguards as are practical. The potential for using synthetic biology for weaponry is, to my mind, overstated. As one of the leading researchers said to me, “If I were a terrorist looking to commit a bio-based atrocity, there are much cheaper and easier ways to do it than engineering a specific microbe – anthrax, say.” Synthetic biology will not, in the foreseeable future, return many “bangs per buck”.

Many of the legal concerns centre on the patenting of gene sequences. This was going on well before synthetic biology, but it recently hit the headlines when Craig Venter, head of the private corporation that tied with the Human Genome Project, announced that they intended to patent a synthetic organism.

We must remember that Venter is, first and foremost, a businessman, and it is very much in his interests to keep his company in the public eye. The scientific rationale for some of these patents is not immediately clear. But we should also remember that, for every Craig Venter, there are probably ten or more Jay Keaslings, placing their research in the public domain and working in an open and transparent fashion for the greater good.

On that positive note, I'd like to thank you for listening, and I'll stop there.

Friday, August 24, 2007

My contribution to the synthetic biology debate

You may recall that the Royal Society is soliciting opinions on various aspects of the field of synthetic biology. What follows is a lightly edited version of my own submission, which I sent off today.

In what follows, I highlight some concerns and dangers, speaking as someone who has an definite interest in the field flourishing (and would therefore wish to see these concerns addressed).

1. Terminology

The first concern is over the term “synthetic biology” itself. The two main issues are “what does it mean?” and “what does it cover?” As pointed out at the BBSRC workshop, clinicians have used the term for a while to refer to prosthetic devices. In attempting to offer a fixed definition of the term, the community runs the risk of becoming overly exclusive at a premature stage. However, there is also a risk that “synthetic biology” will become a “catch-all” term that is too loosely applied. The emphasis on the term “biology” may also serve to alienate mathematicians, physicists, computer scientists and others, who may (wrongly) feel that they have no expertise to offer a “biological” discipline. As a counter-example, witness the success of the field of bioinformatics, which would appear to fairly represent the disciplinary expertise in the field (in terms of the general composition of the term, rather than the relative lengths of its components). As a very crude experiment, I searched in Google for both “computational biology” and “bioinformatics”; the first term returned around 1,530,000 hits, the second around 14,000,000.

This leads on to the issue of “language barriers”. This is always an issue in any new field that involves the collision of two or more (often very dissimilar) disciplines. Being seen to publically ask “stupid questions” is a daunting prospect to most young scientists, and yet many of the major breakthroughs have occurred through just that. This opens up the wider debate on inter-disciplinarity in 21st century science, and how we might best prepare its practitioners. Do we give students a broad, shallow curriculum to allow them to make connections, without necessarily having the background to “drill deeper” if required, or do we stick to the “old model” of “first degree” and subsequent training? My own intuition is that it is far better to intensively train in a single field at the outset, and then offer the opportunity to “cherry pick” topics from a different discipline at a later stage. This educational debate is, however, not one that should be the sole preserve of synthetic biology!

2. Expectation Management

Even when biologists and (say) computer scientists can agree a suitable shared terminology, there is still the risk of a mismatch occurring in terms of expectations of what might be achieved. For example, the notion of “scalability” might mean very different things to a computer scientist and a microbiologist. To the former, it means being able to increase by several orders of magnitude the number of data items processed by an algorithm, or double the (already vast) number of transistors we may place on the surface of a computer chip. To a biologist, the idea of scalability might currently be very different:

“What's needed to make synthetic biology successful, Rabaey said, are the same three elements that made microelectronics successful. These are a scalable, reliable manufacturing process; a scalable design methodology; and a clear understanding of a computational model. "This is not biology, this is not physics, this is hard core engineering," Rabaey said.

In electronics, photolithography provides a scalable, reliable manufacturing process for designs involving millions of elements. Biology has a long way to go. What's needed, Rabaey said, is a way to generate thousands of genes reliably in a very short time period with very few errors. The difference between what's available and what's needed is about a trillion to one.”

3. Conceptual Issues

As the leading nanotechnologist (and FRS) Richard Jones has pointed out, his field was dominated from an early stage by often inappropriate analogies with mechanical engineering (e.g., cogs). It may well be that case that we are in danger of the same thing happening with synthetic biology, where computer scientists impose rigid circuit/software design principles on "softer", more “fuzzy” substrates. Jones quotes, on his blog, an article in the New York Times:

“Most people in synthetic biology are engineers who have invaded genetics. They have brought with them a vocabulary derived from circuit design and software development that they seek to impose on the softer substance of biology. They talk of modules — meaning networks of genes assembled to perform some standard function — and of “booting up” a cell with new DNA-based instructions, much the way someone gets a computer going.”


4. Complexity


The issue of "grey goo" has persistently dogged the field of nanotechnology, and it would be tempting to dismiss similar criticisms of synthetic biology as well-intentioned but ultimately uninformed. However, if synthetic biologists are to avoid the mistake that researchers in GM research made (that is, to appear arrogant and dismissive, leading to mass public protest and restrictive legislation), then we should acknowledge and address the very real possibility of the biological systems under study behaving in very unpredictable ways. Anyone who has any degree of contact with studying biosystems will understand the notion of complexity; components that are connected in an unknown fashion behave in unpredictable ways, which may include evasion of any control mechanisms that have been put in place. As Douglas Kell and his colleagues have observed, it is perfectly possible to alter parameters of a system on an individual basis, and see no effect, only to observe wild variations in behaviour when exactly the same tweak is applied to two or more parameters at the same time. Working in an interdisciplinary fashion may address this issue, at least in part, if modellers work closely with bench scientists in a cycle of cooperation. Once again invoking the issue of scalability, studying the behaviour of complex biosystems through modelling alone will quickly become infeasible, due the the combinatorial explosion in the size of the search space (of parameter values). By actually making or modifying the systems under study in the lab, the problem may be reduced to manageable proportions.

5. Hype

In my own book, Genesis Machines (Atlantic Books, 2006), I illustrate the risk of promising too much at an early stage by describing the story of the “AI winter”. In the 1960s, researchers in artificial intelligence (AI) had promised human-level intelligence “in a box” within twenty years. By issuing such wild predictions, AI researchers set themselves up for a monumental fall, and, when the promised benefits failed to accrue, funding was slashed and interest dwindled. This AI winter (by analogy with “nuclear winter”) affected the field for over 15 years, and it would be disappointing (to say the least) if the same thing were to happen to synthetic biology.

Hubristic claims for synthetic biology should be avoided wherever possible; without singling out particular groups, I have already seen several predictions (again, often conflated with ambitions) that have absolutely no realistic chance of coming to fruition in any meaningful time-scale (if at all). In this more “media savvy” age, perhaps practitioners in synthetic biology might benefit, as their AI counterparts did not, from media training (I have personally benefited (June 2004) from the course provided by the Royal Society, and perhaps the Society might consider a “mass participation” version for new entrants to the field).

Friday, August 17, 2007

For the love of ants

To be published next week, one book on my Amazon wishlist is titled The Ants Are My Friends. Students of popular music may recognise the phrase as one of the great misheard lyrics of our time, up there with "Beelzebub had a devil for a sideboard", rather than an expression of insect infatuation (the response being, of course, "blowing in the wind").

But I rather like the idea of ants being my friends. I've always held these misunderstood creatures in high regard, and was charmed by the story, recounted in Surely You're Joking, Mr. Feynman! (p. 91 in the Vintage edition), of how Richard Feynman investigated ant trail-following behaviour in his Princeton accomodation. He eventually used his findings to persuade an ant colony to leave his larder; "No poison, you gotta be humane to the ants!"

Anyone who has ever watched an ant colony at work cannot fail to be entranced by its beauty and efficiency. A single colony can strip an entire moose carcass in under two hours, and their work is coordinated in an inherently decentralised fashion (that is, there is no "head ant" giving out orders). An ant colony can be considered as a class of "super-organism", that is, a "virtual" organism made up of many other single organisms. Other examples include bacterial colonies and (arguably) the Earth itself.

Ants communicate remotely by way of pheromones, chemicals that generate some sort of response amongst members of the same species. When ants forage for food, they lay a particular pheromone on the ground once they've found a source. When this signal is detected by other ants, they follow the trail and reinforce it by laying pheromone themselves. Chemical signals also evaporate over time, which allows colonies to "forget" good solutions (i.e., paths) and construct new solutions if the environment changes (e.g., a stone falls onto an existing path).

By describing this mechanism in abstract terms, computer scientists have managed to harness the power of positive feedback in order to solve difficult computational problems. Perhaps the leading scientist in the field of ant colony optimization (ACO) is Marco Dorigo, and he has described how to use models of artificial ants to solve the problem of how to route text messages through a busy network of mobile base stations. We've also done some initial work on how ants build spatial structures, using an abstract model of pheromone deposition to explain how certain species can construct "bullseye"-like patterns of differently-sized objects.

Fundamentally, ongoing work in ACO reflects a wider interest in the notion of decentralised control. Rather than controlling everything from "on high" with global instructions, "bottom up" control emphasises the value of small, local interactions in keeping systems running smoothly. Software packages such as Netlogo have brought so-called agent-based modelling to a wider audience. I've just taken on a Ph.D. student to study the evacuation of tall buildings using this approach, and it's clear that, with ever-increasing computational power being available, the notion of simulating large systems of interacting entities will gain increasing influence.

Genesis Machines in the USA

I'm delighted to report that Atlantic have signed a deal to publish Genesis Machines in the USA. It's slated to appear on April 3rd of next year, and will be published by the Overlook Press (preorder here).

Friday, August 10, 2007

Molecules and Marx

My publisher kindly sends me copies of reviews of Genesis Machines that appear from time to time in the press. I was quite surprised to see the book featured in the June issue of the Marxist Review, the monthly theoretical magazine of the Workers Revolutionary Party. In his article, William Westwell invokes Richard Dawkins as the contemporary cheerleader of arch-reductionism and mechanical materialism. But, by concentrating purely on the first half of the book (which, by its very nature is comprised largely of historical background), Westwell ignores one of its fundamental arguments: that 21st century science cannot succeed by insisting on the top-down, reductionist paradigm. Science is still, to a large extent, a reductionist enterprise, but the emerging field of systems biology is providing a complementary approach (in a way, occupying the region where top-down meets bottom up). By arguing for a notion of "quality of computation", Westwell reminded me of conversations I have enjoyed in the past with Brian Goodwin, who has argued that "Biology is returning to notions of space-time organisation as an intrinsic aspect of the living condition... They are now described as complex networks of molecules that somehow read and make sense of genes. These molecular networks have intriguing properties, giving them some of the same characteristics as words in a language. Could it be that biology and culture are not so different after all; that both are based on historical traditions and languages that are used to construct patterns of relationship embodied in communities, either of cells or of individuals?" Unfortunately, Westwell appears to have ignored the later detailed discussion of such matters.

Tuesday, July 03, 2007

Royal Society and Synthetic Biology

I was gratified to see a healthy turn-out for the Cafe Scientifique event yesterday evening, and pleased by the range and depth of questions that were asked. I briefly mentioned at the end of the Q&A session that the Royal Society (the UK's preeminent learned society for the sciences) is soliciting views on the emerging field of synthetic biology.

The field's potential impact is huge, not just in terms of technological developments and scientific understanding, but in terms of social impact, legal issues and ethical concerns. It is vital that scientists engage with the wider community, so that the implications of their work may be considered at an early stage (and, indeed, throughout). The Royal Society is therefore asking for submissions (ranging from brief comments to draft discussion documents) on various aspects of synthetic biology, from potential applications to biosecurity risks and governance.

These exercises offer a real opportunity to help shape future policy, so I would encourage potentially interested parties to visit the call website and consider adding their voice to the debate. The deadline for submissions is August 27th.

Thursday, June 28, 2007

Event in Manchester

If you're in Manchester next Monday and are stuck for something to do in the evening, why not pop along to Cafe Scientifique, where I'll be speaking and (hopefully) generating some discussion? The format is pretty relaxed, with a 30-40 minute presentation from me, followed by a 15 minute break for refreshments, then an open-ended discussion session.

Genesis Machines: Engineering Life

Monday 2nd July 2007 at 6:30pm in Cafe Muse (directions are here).

Although anticipated as early as the 1950s, the idea that we could somehow build working computers from organic components was merely a theoretical notion until November 1994, when a scientist announced that he had built the world's first molecular computer. Emerging from a laboratory in Los Angeles, California, his collection of test tubes, gels and DNA lay at the heart of a totally new and unexplored region of the scientific landscape.

Millions of dollars are now being invested worldwide in molecular computing and synthetic biology research. DNA, the code of life, is right now being used at the heart of experimental computers. Living cells are being integrated with silicon nanotubes to create hybrid machines, as well as being routinely manipulated to add entirely new capabilities. Preparations are being made to build entirely new organisms, never seen before in nature.

This research raises amazing new questions. Does nature 'compute', and, if so, how? Can natural systems inspire entirely new ways of doing computation? How can humanity benefit from this potentially revolutionary new technology? What are the dangers? Could building computers with living components put us at risk from our own creations? What are the ethical implications of tinkering with nature's circuits? In this event we'll examine what it means to reprogram the logic of life.

Thursday, June 14, 2007

Two things

Apologies for the lack of recent activity on the blog; this time of year is always the busiest for academics, what with exam boards and so on. Hopefully things will calm down in the next week or two.

Two items of note: today marks both the publication of Genesis Machines in paperback, and the release of the programme for this year's Edinburgh International Book Festival.

Not much has changed in the paperback, apart from a few additional acknowledgements and some minor changes to the notes and references (which are available online here.)

As for the Book Festival, I'm privileged to be appearing on stage once again with Oliver Morton from Nature, where we'll be discussing "The Future of Nature" as part of the "Genes and Society" Festival theme. Oliver will be talking about "intelligent plants", and I'll be holding forth on biocomputing and synthetic biology (Craig Venter's recent patent swoop has given me lots of nice new discussion material). Anyway, we're appearing on Sunday the 26th of August, and full programme details are available here. The event is sponsored by the ESRC Genomics Policy and Research Forum, and it's one of five that they are supporting.

Friday, May 25, 2007

DNA hash pooling

The draft paper that came out of our trip to Paris has now been lodged with the arXiv e-print server.

DNA Hash Pooling and its Applications

Dennis Shasha (Courant Institute, New York University), Martyn Amos (Computing and Mathematics, Manchester Metropolitan University)

Abstract: In this paper we describe a new technique for the characterisation of populations of DNA strands. Such tools are vital to the study of ecological systems, at both the micro (e.g., individual humans) and macro (e.g., lakes) scales. Existing methods make extensive use of DNA sequencing and cloning, which can prove costly and time consuming. The overall objective is to address questions such as: (i) (Genome detection) Is a known genome sequence present at least in part in an environmental sample? (ii) (Sequence query) Is a specific fragment sequence present in a sample? (iii) (Similarity Discovery) How similar in terms of sequence content are two unsequenced samples?

We propose a method involving multiple filtering criteria that result in "pools" of DNA of high or very high purity. Because our method is similar in spirit to hashing in computer science, we call the method DNA hash pooling. To illustrate this method, we describe examples using pairs of restriction enzymes. The in silico empirical results we present reflect a sensitivity to experimental error. The method requires minimal DNA sequencing and, when sequencing is required, little or no cloning.

Available at http://www.arxiv.org/abs/0705.3597.

Tuesday, May 01, 2007

Spiked innovation survey

I was recently asked to contribute to the annual innovation survey from spiked. There appear to be many interpretations of the term "innovation", but notable entries (from my own perspective) include those by Scott Aaronson, Paul Rothemund and Jeffrey Shallit.

Here's the blurb from the spiked website:

"The internet, the alphabet, the discovery of nuclear fusion, x-rays, the brick, rockets, the eraser: all of these have been identified as the greatest innovations in history in a new survey.

Over 100 key thinkers and experts from the fields of science, technology and medicine - including six Nobel laureates - participated in the brand new spiked/Pfizer survey 'What's the Greatest Innovation?', which goes live on spiked today.

In his introduction to the survey, spiked's editor-at-large Mick Hume says: 'Some choose "sexy" looking innovations, others apologise for the apparent dullness of their arcane choices. But whatever the appearances, almost all of our respondents exude a sense of certainty about the improvement that innovations in their field are making to our world, and the potential for more of the same."

Wednesday, April 25, 2007

A new kind of firefighting

I managed to miss a potentially interesting edition of Horizon on the BBC after making the mistake of flicking over to watch the second half of the Manchester Utd/Milan match.

Anyway, I caught the last ten minutes, and managed to glean the basic facts: that fewer people would have died when the Twin Towers collapsed on September 11 had the authorities been in possession of a global picture of the state of the building (in terms of both its structure and the movement of its occupants). But having the raw data is not enough: it needs to be provided as input to predictive models that are capable of allowing firefighters to play "what if" games. These models are necessarily computationally complex and resource intensive, which is where Jose Torero comes in. He's in charge of Firegrid, an interdisciplinary project dedicated to using Grid-based computing to model and predict, in real-time, the evolution of fire emergencies.

This work is related to my own on evacuation modelling, and we've recently been awarded a Ph.D. studentship in order to develop our ideas on how crush conditions emerge in situations where people fail to follow a set evacuation plan. This work will be done in collaboration with Dr Steve Gwynne, who has worked for the last ten years on modelling people movement, and who helped develop the influential Exodus system. The position will be advertised shortly, so watch this space.

Monday, April 16, 2007

Warwick victorious!

Congratulations to one of my old institutions, Warwick, on winning the 2007 University Challenge. In a tight match, they eventually fought off the reigning champions, Manchester, both securing Warwick's first ever series win and preventing their opponents from gaining the first ever "back to back" run of titles.

I'm afraid, however, that most neutrals watching will remember it more for Prakash Patel's post-presentation lunge for Ann Widdecombe than anything else.