![]() |
We used to talk about "computing with soup". You know, chicken stock. A light broth. Clear. Delicious. Refreshing. Dilute molecules bumping into each other from time to time, reacting according to rules, changing color subtly, making decisions. Chemical reaction networks were programmable, we realized. But it's a new day. It's a new cookbook. It's a new flavor. Today, we are computing with stew: thick, mushy, everything always pressed together. It's complete goop in there! |
![]() |
People keep on telling you, oh yeah, dynamics is a lot more complicated than thermodynamics. Equilibrium is easy; kinetics can be really hard. But you don't really know what they're talking about until you get entangled in an example. Our example is non-pseudoknotted DNA secondary structure. We've got a decent thermodynamic model based on nearest-neighbor interactions. Melting temperatures, Boltzmann distributions, base pair probabilities -- just ask NUPACK or Vienna. But the kinetics of hybridization, hairpin folding, branch migration... well, good luck. You're about to encounter endless fascinating surprises... |
![]() |
Inspired by Ned's article DNA Nanotechnology at 40, Natasha and I thought it would be good to ask the community to step back for a minute and think about the past and think about the future. The resulting volume includes 22 chapters divided into sections on perspectives, chemistry and physics, structures, biochemical circuits, and spatial systems. We hoped Ned would appreciate it, but he never got the chance to see the final product. While it only touches on a smattering of the achievements and possibilities in DNA nanotechnology, and in that sense is an unfinished reflection of an unfinished field, I like to think that he would pleased to see his legacy in the strange, sprawling, chaotic, intermingled growth of the science and technology. One can try, of course, but in truth no one can predict where it's going. Not even Ned has a crystal ball! |
![]() |
For everything there is a season. There was a time when a single person could design a whole computer from transistors up to programming languages. Carver Mead and Lynn Conway taught classes at Caltech and MIT in the late 1970's with the explicit goal of giving students the skills to do this, which spawned their textbook, "Introduction to VLSI Systems", that changed the industry. Carver espoused the idea of a "tall thin computer engineer" who could produce better designs by optimizing across all scales of computer organization so that they worked better together. Molecular programming with DNA nanotechnology is now reaching a similar point -- not with respect to commercialization and applications, but with respect to the complexity of the design process. Has the time come for the tall thin molecular programmer? |
![]() |
The toehold-mediated strand displacement mechanism can be viewed as a way to selectively mine tunnels through a mountainous energy landscape, thus allowing fast kinetics for desired state changes while preserving barriers to undesired pathways. It has proved to be a central concept for dynamic DNA nanotechnology, foundational for molecular machines and computing systems both. As used in well-mixed DNA strand displacement circuits, the mechanism is essentially one-dimensional, as displacement acts from one end of a helix to the other. The recently discovered tile displacement mechanism shares many qualitative features with strand displacement, despite acting at a scale orders of magnitude larger in DNA origami tiles, but has one unique feature: it's behavior is fundamentally two-dimensional when acting within arrays of tiles. How important is this? Very! And how interesting? See for yourself how it makes beautiful connections to synchronous and asynchronous cellular automata, as well as how it highlights principles from the thermodynamics of computation. |
![]() |
In an age when it sometimes seems that everything is a sham, we bring you a SHAM that is no sham. Our SHAM is a self-assembling brain, you might say, a set of molecules that collectively make decisions and spell out their answers. "Is this pattern of concentrations a Horse?" you ask, and the SHAM spells out its nanoscale answer in the words that it is forming: "H!" It's a new way of thinking about neural computation, reflecting old ways. Where some might use Hebbian learning to dig basins in an energy landscape and thereby guide the neural computation thermodynamically down, our SHAM builds barriers in an energy landscape that must be surpassed and thereby challenges our neural computation kinetically up. Where some might take neural mathematics as a prescription for engineering an equivalent molecular circuit, our SHAM takes ubiquitous and inevitable molecular processes, nucleation and crystallization, and finds the neural mathematics within them. Where some might be content to contemplate the concepts and create in silico constructions, our SHAM is real: the molecules synthesized, the fluorescence monitored, the self-assembled shapes imaged by atomic force microscopy. We hope it brings you peace and comfort to know that not all that shimmers is a sham. |
![]() |
☯ The foundation for many machine learning models is a description of a probabilty distribution in terms of energies of configurations, coupled with generative processes designed to satisfy detailed balance such that the system equilibrates to the Boltzmann distribution over the energies. This enables concise analytic formulas, clear thinking, and systematic approaches to derivation of generative processes for manipulating probabilistic information -- such as establishing the equivalence of clamping variables and performing inference of conditional distributions. ☯ The foundation for chemical systems in equilibrium with a bath is again a description of the relevant energies, with respect to which a Boltzmann distribution of system states will emerge. The kinetics of such systems are physically required to satisfy detailed balance, down to the level of individual reaction in chemical networks. ☯ Our thesis is that these two things are two sides of the same coin. Any detailed balanced chemical reaction network can be viewed as, and used as, a machine learning model capable of representing a probability distribution and performing inference by clamping -- which now have physical and chemical interpretations to parallel their information processing interpretations. We show that the ability to encode complex distributions relies on exploiting reachability constraints within stochastic chemical reaction networks, and thus relies on the bath exchanging heat (canonical ensemble) but not material (grand canonical ensemble) for critical species. We further provide bounds on the thermodynamic cost of performing probabilistic inference, approaching zero in the limit as the speed also goes to zero, but related to the relative entropy for finite speed. ☯ |
![]() |
The secondary structure state space for multistranded nucleic acid systems suffers from combinatorial explosion, making exact methods for computing computational infeasible. So kinetic simulations are often used. But the secondary structure landscape for processes of interest often involves high-energy barrier states that make direct simulation of critical rare events computationally infeasible. So a variety of techniques from statistical physics perform biased sampling to efficiently approximate the rates of rare events. But if our goal is to infer model parameters based on data using some kind of gradient descent, very similar simulations would have to be run repeatedly, which is wasteful. The pathway elaboration method is our attempt to address these conflicting goals by "pre-computing" what is expected to be the most relevant part of the state space, then calculating parameter-dependent kinetics using fast matrix methods. |
![]() |
Formally, discrete chemical reaction networks (CRNs) have a finite number of species, but an infinite number of states -- the numbers of molecules in a test tube is unbounded. The point is, this can make it hard to verify that a proposed molecular implementation (a big complicated CRN) is really doing the job it was designed to do (a smaller simpler CRN). Nonetheless, a method based on Milner's notion of bisimulation in concurrency theory was developed in our work on "CRN bisimulation". When considering chemical systems involving polymers of potentially unbounded length, such as the DNA, RNA, and protein polymers in the central dogma of molecular biology, then the number of possible species becomes infinite as well. The point is, the extra infinities can make implementations of polymer reaction networks (PRNs) even harder to verify. And yet, by extending the bisimulation ideas, it can (at least sometimes) be done! |
![]() |
So-called domain-level models were introduced to describe how engineered DNA strand displacement systems should work, prior to assigning actual sequences to the domains. Here we describe our efforts to provide a model rich enough to describe arbitrary secondary structures (so long as no pseudoknots are involved) and to support reaction mechanisms such as association, dissociation, 3-way branch migration, and 4-way branch migration. We use a nifty time-scale separation to make the process tractable and to find a "condensed" reaction network representation. A sequence-independent kinetics model for domain-level reaction steps is developed, founded on rule-of-thumb biophysics, that (imperfectly, but reasonably) predicts reaction kinetics for a wide range of systems taken from the literature. |
![]() |
When chemical reactions are constrained to a surface, new possibilities arise and new constraints are imposed, relative to when they take place in a well-mixed solution. As proposed in our 2014 paper, DNA nanotechnology provides potential mechanisms for implementing arbitrary CRNs on a surface, and thus the "surface CRN" model can be considered as a programming language for future technologies. Here, we continue our exploration of how to think about programming surface CRNs. You might be interested in how we emulate synchronous cellular automata in two dimensions, despite the asynchronous and pairwise CRN reactions. Or how surface CRNs can implement continously active recurrent digital logic circuits -- giving another way to emulate cellular automata. Or how surface CRNs can be programmed to "grow" complex patterns. Or how we can think of programmable molecules as "robots" and explore ideas from swarm robotics, but at the nanometer scale. This is just a sampling... the most exciting ideas are yet to be discovered... by you? |
Impossible proofs make an invisible bridge to what can be done |
証明の 橋は掛かれり 見えずとも |
Sho-u-me-i-no Ha-shi-wa-ka-ka-re-ri Mi-e-zu-to-mo |
An unbridged gap awaits molecules to jump therein lies the proof |
証明す 橋なき川を 跳ぶ分子 |
Sho-u-me-i-su Ha-shi-na-ki-ka-wa-wo To-bu-bu-n-shi |
![]() |
You've heard them say, "it's a noisy business down there", and you looked at the inner workings of the cell, and yup, that's what you saw in the thick thermal bath. But Asimov told us that any sufficiently advanced technology is indistinguishable from magic, and SETI researchers remind us that any sufficiently compressed information is indistinguishable from randomness. So perhaps there's more to the noise, you wondered? In order to see the nearly-invisible, you have to know what you are looking for. Here we begin an exploration of how stochastic chemical reaction networks can shape noise with almost arbitrary finesse. |
![]() |
The world is hierarchical. So it seems. In geography, fractals are key to understanding that. In biology, perhaps P≠NP is the key. Whaaaat? Let me explain. That some functions are hard to compute but easy to verify means that a computationally-limited supervisor can verify the performance of a large number of workers doing hard things. And a bigger boss can verify the performance of a bunch of supervisors. From molecules to cells to organs to organisms, even from students to professors to deans to presidents, P≠NP makes hierarchical organization feasible, effective, and naturally self-repairing. In this paper, I argue that "programming by recognition" -- a computational architecture founded on the P≠NP distinction -- is natural for stochastic chemical systems and puts all that randomness to good use. The case is not entirely convincing, but hopefully will stimulate further thought. |
![]() |
Nucleic acid secondary structure rearrangements form the basis for a wide range of natural phenomena as well as an incredible diversity of engineered molecular machines. Improving the accuracy of secondary structure kinetics models will be of enormous benefit both for analysis and for design. One approach is to fit model parameters to experimental data, but when the model is a stochastic Monte Carlo simulator like Multistrand, this can be prohibitively time-consuming. Here we explore some approaches for speeding up parameter estimation. Can we develop even better techniques? |
![]() |
How many times have you been exhorted to do more with less? It's not a bad goal. For well-mixed chemical reaction networks, no reaction does "less" than A+B → B+A. In fact, it does absolutely nothing. So how can you do more with it? What these three undergraduates realized (much to my surprise) is that reactions like A+B → B+A are not so ineffective if they are operating confined to a surface. Now, the reaction means that A swaps positions with B. How much more can you do with that? For starters, you can build arbitrary feedforward Boolean logic circuits. It almost seems like getting something from nothing. |
![]() |
Hao Wang was right, although he was most famous for being wrong. Back in the early 1960s he was investigating mathematical logic from the perspective of automatic theorem proving, which entailed going head to head with Gödel's incompleteness and Turing's uncomputability. Looking for simpler cases to establish the line between solvable and unsolvable logics, he saw a connection between a subset of mathematical logic and a problem in mathematical geometry: determining whether a finite set of tiles could be used, with repetition, to tile the entire plane. He famously conjectured that any set of (what are now called) Wang tiles -- square tiles with matching rules on each side -- either tile the plane periodically, or else there is no infinite tiling with complete coverage. If that were true, his subset of mathematical logic would be decidable. But it wasn't true! Shocking him and many others, his student Robert Berger proved that there are tile sets that can tile the plane, but only aperiodically. Shortly thereafter, Roger Penrose came up with his beautiful aperiodic tilings using kites, darts, and other shapes. While periodic tilings were well-established as a mathematical basis for understanding real crystals, not many believed that the Penrose tilings had anything to do with real materials -- until 1984, when Dan Shechtman discovered quasicrystals in an aluminum-manganese alloy! Well, then, what about Hao Wang's other tilings? Perhaps most intriguingly, Wang had designed tiles that, if started with an initial "seed", directly simulate the behavior of a Turing machine. Does this "algorithmic order" within tilings also reflect a way that real molecules can self-organize? Yes, it seems to be so. |
![]() |
Coarse-graining is the act of lumping together states in a detailed model to obtain a more abstract model that -- if the coarse-graining is good -- can be used to accurately predict important behaviors of the original model. Design is the act of creating a detailed implementation based off of an abstract specification -- and if the implementation is good, coarse-graining will bring you right back to the abstract specification you started with (or somewhere close). Here we consider a framework for quantitatively evaluating such relationships in the context of DNA strand displacement systems. |
![]() |
Some theory is too good to be true. Lots of theory seems too good to be true. The most fun part of experimental work is to figure out which is which. In this paper, we make the remarkable discovery that our theory for how to design leakless DNA strand displacement cascades only seems too good to be true. But at it's core, it's the stuff of reality. |
![]() |
Can you accomplish the same task with less? That's the game here. Given an algorithmic tile set that performs well -- perhaps because it uses proofreading principles -- can you find a smaller tile set that does just as well? Fewer tiles, fewer glues. You might make the infeasible feasible again. Our DNA tile set compiler, Alhambra, can do this now. |
![]() |
The question of whether a molecular implementation is "for all intents and purposes" equivalent to the intended abstract specification is a fundamental question for molecular programming. Building on earlier work by Seung Woo Shin and Qing Dong, we show that bisimulation provides a surprisingly natural framework assessing the correctness of proposed implementations of formal chemical reaction networks (CRNs). While CRN bisimulation can be easily applied in many circumstances of interest, and an algorithm for doing so is frequently effective, we show that in general variants of the problem of finding and checking bisimulation are NP-complete or PSPACE-complete in this context. Differences with the alternative theory of pathway decomposition are also discussed. |
![]() |
What does it mean for two chemical reaction networks to be logically equivalent? The answer is not as obvious as it may seem -- in fact, we still can't give a fully satisfactory answer. There seem to be many possible notions one could entertain, each with its own strengths and weaknesses. Here we present a refinement of the ideas from Seung Woo's Masters Thesis that focus on the smallest coherent sequences of reactions out of which all other sequences can be built. A strength of the theory is that it can handle the "delayed choice" phenomenon, wherein a single species in a molecular implementation has already committed to participation in a reaction, but not yet fully committed to exactly which one. Combined with bisimulation ideas from Qing Dong's Masters Thesis, we hope that this theory can address most, if not all, methods to implement chemical reaction networks using DNA strand displacement systems -- and perhaps more. |
![]() |
In days gone by, did you ever build electrical circuits on an old-fashioned breadboard? Maybe you plugged in a few capacitors, a few resistors, an inductor, and transistor -- and voila! an AM radio! The wonderful thing was that if you could draw the circuit and analyze the equations, you could build it. And with a little tuning, it would work. Will biochemistry ever be that straightforward? We take a small step, by showing that the standard chemical reaction network formalism can serve as a programming language for DNA strand displacement systems. We provide a CRN-to-DNA compiler and build a three-reaction oscillator experimentally. What will you build? |
![]() |
Science is often blown forward by the wind of dreams. Can you imagine a mail room smaller than a single virus? Can you imagine a box of mail packages, each being just a single molecule? Can you imagine a single molecule robot wandering around this room, picking up molecular packages and placing each one in the right molecular bin, according to the zip code on its address? Can you imagine a cargo-sorting DNA robot? We dreamed this dream seven years ago. It wasn't particularly realistic, then. But now it is real. |
![]() |
Models that capture a substantial fraction of the known thermodynamics of multistranded DNA molecules at the analytically-tractable secondary structure level, such as NUPACK, have proven invaluable for the analysis of both biological and artificial nucleic acid systems. Extending such models to predict the kinetics of secondary structure rearrangements and interactions between molecules, as done for example in the Multistrand simulation, has the potential to address a wider range of phenomena. However, thermodynamics does not dictate kinetics -- there is an infinite family of kinetic models that are perfectly consistent with any given thermodynamic model -- and therefore the accuracy of naive kinetics models is limited. Here, we propose a parameterization of multistranded secondary structure kinetics that is based on the Arrhenius model for elementary base-pairing changes, and we prototype a Bayesian Markov Chain Monte Carlo (MCMC) inference method to obtain an ensemble of improved parameter sets -- resulting in markedly increased accuracy when evaluated on a database of experimentally-measured kinetics rates culled from the literature. |
![]() |
What is the difference between noise and information? What distinguishes stochasticity and thinking? These concepts may seem like opposites, but from a certain perspective, they are closely related. Here we present several theoretical constructions for chemical reaction networks whose stochasticity embodies meaningful information, and whose response to input constitutes perfect Bayesian inference. |
![]() |
What does it mean to compile to molecules? To some, this just means that a computer was used design the molecules, somehow. To others, a more rigorous process is implied: that the intended design is specified by a "high level" formal language, and that a systematic process is used to translate the design into the "low level" molecular construction. Here, we go further: the high-level specification language and the low-level implementation language must each have a semantics -- that is to say, the intended/expected behavior of a given system must be well-defined based on its description -- and more importantly, the behavior of the implementation produced by the compiler must come with a guarantee that it is effectively the same as the specification. These features are prototyped by the Nuskell compiler, which translates formal chemical reaction networks into domain-level DNA strand displacement systems. |
![]() |
We do our best to provide a gentle but solid introduction to some key physical principles for DNA tile self-assembly, including algorithmic growth, error rates, proofreading, and nucleation. We discuss how an understanding of these issues allows one to navigate from abstract models of tile assembly, such as the aTAM, to more realistic models of tile assembly, such as the single-crystal kTAM and the mass-action kTAM. Finally, we outline a "unified" model of tile self-asssembly that helps clarify the relationships between different levels of abstraction. |
![]() |
In earlier work, we proposed a theoretical model that combines tile self-assembly and chemical reaction networks, showing that fabrication tasks and computation tasks can be performed more efficiently than in either previous model, when measuring in terms of space used and program size. Here, we show that the CRN-TAM is also as fast or faster. In particular, we show that a class of search problems can be solved in polynomial time, using exponential space -- i.e. the CRN-TAM can efficiently use the inherent parallelism of chemistry for this class of problems. (Full proofs will appear in an expanded journal version of the paper, "soon".) |
![]() |
And so he looked at a tiny bubble bursting on the surface of an infinite ocean. Within it, molecules, their world torn asunder. And in that vigor, and in that endless churning, the origin of life. We followed him deep into this vision. |
![]() |
DNA strand displacement systems are powerful in theory, capable of simulating arbitrary formal chemical reaction network dynamics. In practice, they are plagued by problems, such as leak reactions where the outputs of an intended reaction are released -- to some degree -- even in the absence of the intended inputs. Here we present a systematic approach, involving what we call "double long domains", that aims to eliminate leak reactions. |
![]() |
The abstract chemical reaction network (CRN) model allows for the specification of complex dynamical behaviors in a well-mixed solution. CRN programs have a systematic implementation as DNA strand displacement cascades. The abstract tile assembly model (aTAM) allows for the specification of complex self-assembly processes within a single growing crystal. aTAM programs have a systematic implementation as DNA tile sets. The CRN-TAM provides a "minimal" integration of these two models, allowing CRN reactions to produce tiles, and allowing tile assembly steps to send signals back to the CRN. Although a compelling implementation is not yet available, we show that the CRN-TAM can do things neither previous model can do alone -- in particular, we show that concise CRN-TAM programs can ("optimally") construct arbitrary algorithmically-defined objects, without the sometimes-dramatic scale-up required in the aTAM. |
![]() |
Multistrand is a simulation package that performs random walks on the secondary structure energy landscape for test tubes of multiple (but not too many!) DNA strands. It has a powerful Python interface that allows setting up complex sets of simulations, as well as powerful analysis methods for making sense of them. (Multistrand was developed as the major component of Joseph's PhD thesis.) |
![]() |
Everyone makes mistakes. That applies to molecules too: during self-assembly, for example, sometimes the wrong molecule arrives at the wrong place, and sticks -- resulting in the growth of an ill-formed structure. But there's a solution: double check, triple check, quadruple check. In DNA tile self-assembly theory, there is a natural way to do this, using proofreading tile sets. Here we demonstrate, experimentally, that assembly error rates decrease exponentially with molecular designs that allow increased levels of proofreading. |
![]() |
Can you make a career of watching the grass grow? Probably. But we prefer watching nano-grass -- or, to be more precise, nanotubes self-assembled from DNA tiles. By watching how fast they grow, or how fast they shrink, as a function of temperature and concentration, we were able to extract the association rate constant as well as the enthalpy and entropy of binding. As usually happens when you watch closely and think carefully, we also observed several other interesting phenomena... enjoy the movies! |
![]() |
Abstract chemical reaction networks (CRNs) provide an all-encompassing formal language for modelling well-mixed chemical systems involving a finite number of species. Showing that arbitrary CRNs can in principle be implemented with DNA strand displacement cascades [*] was a major step toward proving the generality and universality of pure-DNA systems. However, well-mixed systems are far less powerful than molecular systems that exploit spatial geometry to encode information combinatorially. While I know of no all-encompassing formal language for molecular machines with geometry, a very broad class of systems can be described as discrete chemical reaction networks wherein each molecule is localized on a surface at grid points, and asynchronous reactions mediate movement and state change. Here we suggest a pure-DNA implementation using a combination of 3-way and 4-way branch migration. |
![]() |
Small is different. It's not different mechanistically, of course, but understanding it can be different because many of the simplifying assumptions we use to understand large systems -- systems with bijillions of copies of every type of molecule where macroscopic effects arise from the average case -- no longer work so well. Here we make tiny water-in-oil droplets, about the size of a cell, that contain a programmable biochemical oscillator. Because of stochastic effects, every droplet behaves differently. Can you guess why? We thought we knew, but our first guess was totally wrong. Read on... |
![]() |
When David and Georg realized that DNA strand displacement cascades can be designed to approximate arbitrary formal chemical reaction network (CRN) dynamics, that was theoretically exciting. Now that Yuan and collaborators have successfully demonstrated an example of Luca's version of the CRN-to-DNA implementation scheme in the laboratory, things are getting really exciting! |
![]() |
Toehold-mediated DNA strand displacement is at the heart of many dynamic DNA nanotechnology devices. And it works much much better than we thought it ought to. So we asked why? It all comes down to the relative rates of bimolecular collisions, zipping and fraying of the double-helix, branch migration steps -- and a previously unknown thermodynamic cost for branch migration intermediates. So now we can say that toehold-mediated DNA strand displacement works just about as well as it should. And we have insights about how to make it work better yet. |
![]() |
Theories of the logic and kinetics of algorithmic self-assembly make many idealizations that eliminate complexities and clarify essential insights. But those complexities are still there when one tries to create self-assembling systems in the laboratory. Which ones are most important, what effects do they have, and how can one design molecules and systems to minimize assembly errors? Examining these questions from both biophysical and combinatorial angles lead us to a DNA sequence design algorithm that may perform orders of magnitude better than previous methods. |
![]() |
DNA tile self-assembly provides a molecular architecture for algorithmically programming the growth of complex but static geometrical structures with molecular precision. DNA strand displacement circuitry provides a molecular architecture for algorithmically programming the temporal dynamics of well-mixed solutions by design of chemical reaction pathways. What might be achieved by integrating them to enable programmable control over spatial and temporal self-organization simultaneously? Here, we use an upstream strand-displacement catalytic circuit to control the timing of a downstream tile-assembly system to isothermally grow DNA nanotubes. |
![]() |
DNA nanotechnology provides the tools for engineering ``smart'' molecular motors that not only move from place to place, but can change their actions based on sensing molecules nearby. Molecular robots, if you will. But don't imagine a single molecular robot. Imagine a swarm of them. Walking on top of each other like bees or ants. What power is there in numbers? What power in the ability to move? On top of one another? Here we develop a theoretical model -- with roots in algorithmic tile self-assembly and L-systems and graph automata -- that can be used to explore what systems of interacting molecular robots can accomplish. As a hint of what's to come, here we examine the fabrication task: build an algorithmically-specified object. We show that our theoretical molecular robots (called ``nubots'') can fabricate objects exponentially faster and more compactly than can be done by passive self-assembly systems such as tile assembly. |
![]() |
Before he joined the DNA group, Niranjan was a skilled fighter in the world of bandits. They were an unruly and unpredictable bunch, some with eye patches, some with just one arm -- some with many arms! But Niranjan learned their ways, and figured out how to get the best of them. With these types, you have to poke and prod them a bit -- but be wise as to what you know and don't know about how they'll respond. Always aim to hit the place that might be the best spot. And when you see what happens, revise your thinking about what they could do. It's surprising effective that simple strategy can be. Posed mathematically by a proper Bayesian. |
![]() |
Dave wrote a delightful introduction to and review of the theory of tile-based algorithmic self-assembly -- complete with a motivational video! |
![]() |
To attain complexity, one must first master simplicity. That's been the mantra for much of our research engineering cell-free biochemical circuits. Previously we developed in vitro transcriptional systems as a bare-bones model of genetic regulatory networks that required only two essential enzymes. Knowing only how to implement negatively regulated "genelets", we constructed a bistable feedback circuit using two such genelets -- comprising six synthetic DNA oligonucleotides. Here, we show how positive regulation can be achieved, and build a single-switch positive feedback circuit that is similarly bistable. With only four synthetic DNA oligonucleotides. This simplicity encouraged us to characterize and model the system within a Bayesian inference framework. |
![]() |
It's nice to see what you're knowing. For over 15 years, our group has been modeling DNA tile self-assembly as a crystal growth process where individual tiles attach at a constant rate (independent of where and how they attach) and detach at a rate that depends exponentially on how many base pairs they are attached by. This model formed the basis for our theoretical investigations, for our experimental designs, and for our confidence that algorithmic self-assembly could work with low error rates. We took it for granted. But we never tested it directly. Now we have. |
![]() |
Sometimes a little flexibility makes a huge difference. For example, in Shannon's theory of information, the number of bits required to specify an element of a set with no possibility of error (the deterministic information) can be dramatically larger than the number of bits required if you can tolerate some chance of error, no matter how small (the probabilistic information). This work establishes a perhaps even more dramatic distinction for computation within well-mixed stochastic chemical reaction networks. Whereas it was previously known that, accepting an arbitrarily small probability that the output will be in error, well-mixed stochastic chemical reaction networks can simulate Turing machines, and that in contrast questions of reachability (rather than probability) are decidable, in this work the exact class of functions that can be deterministically computed by well-mixed stochastic chemical reaction networks (with no possibility of error) is elegantly characterized in terms of semilinear functions. This is a huge gap! |
![]() |
Rebecca Schulman, Bernard Yurke, and Erik Winfree. What is the simplest chemical self-replicator? Well, that depends on just what you mean by "self-replication". If your interest is ultimately in Darwinian evolution -- and perhaps the origin of life here on Earth or elsewhere -- then an interesting self-replicator must carry combinatorial information that guides its function and can be mutated to explore a vast range of functions. In 1965, Graham Cairns-Smith proposed that crystals -- specifically, polytypic minerals -- may have been the first such chemical self-replicators capable of Darwinian evolution. Here, we experimentally explore the general principles and mechanisms needed for Cairns-Smith's hypothesis, but using DNA tile crystals rather than mineral crystals, and find that they are sound in practice. |
![]() |
The abstract Tile Assembly Model examines a single tile-based crystal as it forms from a seed within an inexhaustible bath of free tiles. But in solution, many crystals may grow in parallel, and if they interact with each other by self-assembly, it is natural to think that this parallelism could be exploited to grow well-defined structures much more quickly. Somewhat remarkably, and counter-intuitively, the authors show that if basic elements of chemical kinetics are taken into account (low concentration species encounter each other less frequently than high concentration species) then the parallelism of hierarchical self-assembly provides no advantage at all (caveat, caveat). If you want to build large complex things quickly, look elsewhere. |
![]() |
We are all familiar with intrinsic universality. We know there exists a single Turing machine that can simulate any other Turing machine, step for step but with some linear spatial and temporal scaling. We know there exists a single cellular automata rule that can simulate any other cellular automaton, step for step but with some linear spatial and temporal scaling. This gives the models a sense of wholesomeness. They can simulate themselves efficiently; nothing's missing. But the abstract Tile Assembly Model, so dear to my heart, appeared able only to simulate certain subsets of itself -- an awkward incompleteness. Until now! |
![]() |
Asymptotic upper and lower bounds for the tile assembly complexity of NxN squares are known. But the precise number? Well, it's related to the Kolmogorov complexity of the number N, and that's uncomputable. However, the relation is not tight, and a polynomial time algorithm to find the smallest tile program is known for "temperature" T=2. Here, the authors generalize that result for all temperatures. |
![]() |
The brilliance of genetic regulatory networks is that they are programmable manufacturing plants. They can build all sorts of things. Protein motors, self-assembling scaffolds, enzymes, ion channels... they even build biochemical circuits, and transcription factors that regulate what is built, and when. Can chemical engineers learn to build non-biological molecular systems that operate according to similarly powerful principles, as programmable molecular manufacturing plants? As an exercise toward that goal, here we use in vitro transcriptional systems -- a stripped-down model for genetic regulatory networks in which RNA rather than protein plays the important signalling and control roles, with only two essential enzymes needed for RNA production and degradation -- to construct a synthetic biochemical clock that controls downstream nucleic acid machinery. |
![]() |
At last, we have a small test tube of over a hundred DNA strands, playing games with you, trying to read your mind, acting like a lovely tiny brain. |
![]() |
Benoit Mandelbrot famously noted that the geometry of natural objects is often fractal -- that self-similar structures appear at many scales, resulting in a fractional dimensional exponent. Self-similarity can appear not only in static objects, but in dynamic behavior patterns, or even in concepts. Here, Sungwook and Paul show that the principles of sequence-specific binding, which we enjoy at the nanometer scale in DNA and RNA, reappear at a ten times larger scale when engineering nanostructures (and eventually nanomachines) using DNA origami. |
![]() |
People sometimes say that DNA computing is useless. They're probably just thinking about the wrong application. A postdoc in our lab often wears a T-shirt with a classic XKCD comic. Inspired by the failure of math to explain romance, ( √ ♥ = ? ) we decided to try our latest and greatest DNA technology on the square root problem. It was challenging: designing 130 DNA strands to all work together as a circuit in a single test tube! Not only did it get all the answers right for inputs less than three ( √ < 3 ), but it even computed up to the square root of 15 (rounded down). Is that useless? |
![]() |
What is the core of life? Is it, as the poets sometimes say, the ebb and flow of the tides, the in and out of breathing, the ups and downs of fortune? Or is it, as the electrical engineers might say, all in the circuitry, the processing of information, the dynamics of behavior? In this work, we present a simple DNA device, called the "seesaw gate", that should make both poets and engineers happy. It operates by a simple back-and-forth process -- a strand comes in on one side, pushing that side "down" and kicking off the one on the other side as it goes "up". And then visa versa, like a seesaw. Released strands are then free to interact with other seesaw gates, allowing networks to be designed systematically. Remarkably, the ebb and flow of activity in these networks can perform computations of arbitrary complexity in principle -- in fact, we describe a compiler that translates digital logic circuits into functionally equivalent seesaw gate networks, and we argue that the simplicity of the motif should make networks containing thousands of gates possible. If this theoretical proposal pans out experimentally, could it become a core technology for embedding circuitry in synthetic biochemical systems? |
![]() |
The first* purely chemical oscillator was discovered a little over 60 years ago. Belousov's detailed paper was rejected twice, appearing later only as an abbreviated conference abstract and as a full paper only posthumously. Thus ingloriously began the scientific study of how chemical circuitry can lead to non-trivial dynamical behavior. Eventually, of course, the study of simple chemical oscillators and complex biological clocks flowered into a rich interdisciplinary field. In this paper, we probe the middle ground by showing how cell-free transcription and degradation can be systematically designed to create a variety of oscillator architectures -- merging the programmability of biology with the simplicity of in vitro reactions. |
![]() |
The concept of a "toehold" for initiating branch migration and strand displacement has become central to our thinking about kinetic control in DNA nanomachines and DNA circuits. Toeholds were originally formulated as single-stranded extensions contiguous with the double-stranded domain where branch migration occurs, so that a complementary strand can bind to the toehold and initiate branch migration up to six orders of magnitude faster than in the absense of a toehold. Here, Genot and colleagues show that toeholds need not be contiguous with the double-stranded branch migration domain; localizing the invading strand at a remote site within the target molecule can also serve to significantly accelerate the targeted strand displacement reaction. The resulting design flexibility is essential for some complex DNA nanomachines; furthermore, the intervening region between the toehold and the branch migration domain can be used as a knob to modulate the reaction rate. An idea worth knowing about. |
![]() |
This is what happens when you write a really good introduction to your thesis: It metamorphoses into a top notch review article. |
![]() |
When two Watson-Crick oligonucleotides hybridize to each other, they cancel each other out. This is useful. For example, it allows those oligonucleotides to serve as opposing signals in their single-stranded form. What if you want to build a DNA strand displacement circuit in which two dynamic signals cancel each other out, but the two signals must have unrelated sequences? Well, Dave's got an elegant trick for you. And it gets better. When the signals cancel out, they also release a new strand with a new sequence juxtaposition that can be fed into downstream circuit components. He demonstrates these features by constructing elegant Boolean logic gates and analog concentration comparators. Even more useful! |
![]() |
Everyone knows what happens when an astronaut hops out of his spaceship a little too close to a black hole. As he drifts toward the black hole, the gravitational pull becomes stronger and stronger, until it rips him apart, legs from his body, toes from his feet -- even molecules and atoms are eventually rent asunder. Interestingly, it's not the magnitude of the gravitational force that is so disruptive, but the change in gravity with distance, increasing as 1/R2. (An equally strong, but constant, force would result in no worse than a relatively pleasant free-fall sensation.) As it is, the astronaut's toes are eventually pulled so much harder than his knees, that the differential force is greater than the forces holding the tissue together -- and pop! Loosely speaking, Rizal and Bernie have built a microfluidic black hole, and rent asunder billions of DNA nanotube astronauts. |
![]() |
Well, it's Christmas season, and the Grinch isn't giving you any extra tiles. Success is the best revenge, so how do you do the most with what you've got? Remarkably, when you optimize the stoichiometry of the different tile types -- and no, it's not quite as simple as using them in direct proportion to how often they occur in the final assembled shape, but it's almost that simple -- you simultaneously get the best error rate and the best assembly time. So there, Grinch! |
![]() |
It's great that an entropy-driven DNA catalyst can amplify a signal to detect the presence of a target nucleic acid molecule. The amplified DNA signal can then trigger some downstream chemical activity, just under the right circumstances. But what if the circumstance you're trying to detect isn't the presence or absence of a specific single species, but rather a specific pattern of analog concentration levels for a spectrum of nucleic acid targets? One option is to implement a threshold for each species, then process those YES/NO answers with digital logic -- AND and OR gates. Another option, explored here, is to first implement analog logic that appropriately combines the signals from each species, then takes a threshold to yield the final YES/NO answer. To do so, Dave and Georg propose a scheme for modifying our entropy-driven catalyst system to serve as a signal multiplier, coupled with a mechanism for taking a difference and effecting a threshold. The resulting neural network-like architecture could be quite handy for chemical pattern recognition. |
![]() |
You've gotta love those little cartoons of Turing machines, you know, with the robotic handcar rolling along on the track writing zeros and ones all over the place? It's a seminal model of abstract computing machines, sure, but real computers aren't little motorized machines like toy trains. Right? Computers don't even need moving parts -- just a smooth and soundless flow of electrons through a staggering maze of wires carved into a cold piece of silicon. The Turing machine model is great for theorists, but of little use for anything else. Right? Wrong! The Turing machine is a wonderful, practical, implementable model of an autonomous molecular machine that modifies a linear heteropolymer. Charles Bennett realized this many years ago during his studies of the thermodynamics of computing. We don't quite figure out how to design a DNA Turing machine here, but we get pretty close -- we give a theoretical construction for its close cousin, the stack machine. |
![]() |
How pervasive is the principle of evolution? How complex must be the seed of life? Here we present our take on the simple origins of complexity, extending Graham Cairns-Smith's ideas: that the logic itself of crystallization can in principle give rise to unbounded complexity in pattern formation -- with very simple crystal repeat blocks, in very simple environments, and without any intrinsic chemical or mechanical function. Perhaps life is logically inevitable. |
![]() |
Dave describes his motivations and methods for designing the DNA sequences of molecular machines. These rule-of-thumb principles have worked out pretty well in his hands ( see his other publications ) so there must be some merit to them. It's an insightful paper, and practical. Plus, he wrote some quick-and-easy software that he's been using for a few years. Feel free to try it out. |
![]() |
How small can a robot be? When I was in school, I built a small robot by taking a remote-control car, adding some light sensors and bump sensors, and installing a tiny on-board computer. I programmed it up to scurry around following walls. Lots of fun. But while I played, researchers in academia and industry were building pint-sized robots that carried out collective tasks in distributed "swarms", or that connected to each other in reconfigurable ways to form "meta-robots". Researchers wanted to make the robots smaller and smaller, so that they could have lots and lots of them, so that they could do more and more interesting things. Some even began using silicon chip fabrication techniques to build MEMS (microelectromechanical systems) robots only millimeters on a side. Once you get that small, there's not really much room for a CPU and memory, much less sensors and motors! How can you program a robot -- or a swarm of robots -- when each one has such limited capabilities? It's a deep and important question. Fast forward to today. Can you make a robot from a single molecule? Can a single molecule sense its environment, make decisions, and take actions? The answer is a resounding yes! But how? Our first baby steps -- born of Milan Stojanovic's seminal work on "smart and agile" deoxyribozyme "spiders" -- are reported here in a collaboration lead by Milan that involved chemical synthesis and characterization at Columbia, atomic force microscopy at Arizona State University, and real-time single-molecule fluorescence microscopy at Michigan State University. |
![]() |
If chemistry is programmable, what's the programming language? Well, there are many possibilities... but let's not go with a fad. For well over one hundred years, chemists have been using the elegant and refined mathematical language of mass action chemical kinetics. Traditionally, it has been used descriptively, as a means of describing chemical systems encountered in the real world. As chemical engineers interested in programming chemistry, we wish to use the language of chemical kinetics prescriptively, as a means of describing the behaviors we aim to achieve. It's a rich language, capable of expressing behaviors as varied and sophisticated as stable attractors, oscillators, chaotic dynamics, signal processing and computation -- but much of the evidence for this has come from theoretical studies of chemical reaction networks that are possible in principle, rather than from the study of real systems. That is, for the past hundred years, many intriguing chemical reaction networks existed only in a kind of theoretical dream world, with no physical implementation known. Finally, here we show how any chemical reaction network you dream up can be implemented using systems of DNA logic gates, transforming chemical kinetics from a modeling language into a programming language. Go ahead, make your dreams come true! |
![]() |
What makes a molecule tick? There's one way to find out: Poke it. Prod it. Dave thoroughly harrassed his entropy-driven DNA catalyst system and found out a lot that way. A few surprises: * The system is exquisitely sensitive to mutations in the catalyst strand sequence, but relatively insensitive to mutations in the fuel strand sequence. * The system is almost unaffected by a high background of random poly-{A,C,T} strands, but performance rapidly degrades with high concentrations of random poly-{A,C,G,T} strands. These results can be explained in terms of design features of the entropy-driven catalyst system. You'll also find information about modeling, leak reactions, poisoning reactions, strand purity, 5'/3' orientation, and bulky add-ons to the catalyst. The bottom line: this is stuff you'll want to know if you plan on designing robust and modular strand displacement circuits. |
![]() |
Of central importance in the theory of computation -- and in life in general -- is how long it takes to get things done. For standard models such as digital circuits and Turing machines, time complexity is very well understood. Does our understanding of classical computation directly yield results for computation in biochemical circuits? Of course, it depends on how you do things. For example, as shown here, circuits built from stochiometric gates (one output molecule per input molecule) are asymptotically slower than those using catalytic gates (each input molecule can produce many output molecules). This suggests that the time complexity of formal chemical reaction networks is likely to develop into an interesting, rich -- and probably suprising -- theory. |
![]() |
When writing a computer program, it is easy -- so trivial you don't even think about it -- to make sure that your algorithm starts with the correct input and proceeds from there. Models of algorithmic self-assembly usually assume that assembly begins with a "seed tile" which provides the input information. But molecular self-assembly is an inherently parallel process; assembly begins anywhere it is energetically favored. In this paper we show that tile sets can be constructed that introduce arbitrarily large barriers to spurious nucleation (starting self-assembly from something other that the seed tile). By combining this construction with previous methods for proofreading tile sets, we can now construct tile sets that are robust to all major types of errors, without introducing significant slow-down. Control of nucleation also has some very interesting applications, such as enzyme-free detection and amplification, and enzyme-free in-vitro evolution. |
![]() |
Carbon nanotubes are amazing molecules -- rolled up sheets of hexagonal carbon mesh with astounding thermal and electrical properties, they're the nanocircuit engineer's dream. But they're oh so hard to handle! Too small and too slippery to pick up and put them where you want, most researchers either study individual nanotubes, make do with regular arrays of nanotubes, look for chance circuits in randomly scattered piles of nanotubes, or rely on bulk properties of tangled tubes. Here we suggest a way to self-assemble complex nanotube circuits in parallel -- by sticking them in precise locations on DNA origami nanobreadboards. In the future, we envision this technique being expanded from the two-nanotube devices demonstrated in this paper, to multiple nanotube circuits on individual origami, to large-scale nanotube circuits on origami placed on lithographically-patterned surfaces. Dream on! |
![]() |
Jongmin's very readable review about principles of network architecture in cells, synthetic networks in vivo, synthetic networks in vitro, design challenges related to enzyme saturation and waste products, and the prospects for constructing a minimal cell. |
![]() |
Computer programs work by transferring control from one line of the program to another. Many DNA programs work by transferring control from one part of a molecule to another. That's what toehold exchange does. It is an essential mechanism in many of our DNA strand displacement circuits. Here we study the device physics of toehold exchange and show that kinetics can be predicted remarkably accurately from the thermodynamics of toehold binding. With this understanding, we can make our DNA programs transition smoothly from step to step. |
![]() |
RJ Kershner, LD Bozano, CM Micheel, AH Hung, AR Fornof, JN Cha, CT Rettner, M Bersani, J Frommer, PWK Rothemund, GM Wallraff. Electrical engineers excel at making things using top-down patterning, such as using lithograph to etch circuits on large wafers of silicon, but it is quite difficult to achieve feature sizes below 20nm. Molecular engineers excel at making things using bottom-up self-assembly, such as using DNA hybridization to fold a virus genome into DNA origami, but it is quite difficult to put all the precisely-constructed molecules in the right places on a large scale. Patterning from 10cm down to 20nm. Patterning from 100nm down to 3Å. A match made in heaven. (Joint Caltech/IBM work.) |
![]() |
What is a seed? The tiny seed of a giant sequoia tree, sprouting after the fire. The invisible seed of an idea, from which a thousand possibilities grow. A crystal seed, determining the order of all that grows from it. The seed of man and woman, carrying with it the future of humanity. Clearly, it's important stuff. Why? The seed carries the information, the creative part, the inspiration -- and what follows is mere mechanism, the consequences, the algorithm. In this work, we use DNA origami as a highly effective seed for growing DNA tile crystals. Arbitrary information can be put on the seed; it directs the growth of DNA crystals and determines their morphology... much as a genome determines phenotype of an organism... or even, as an idea creates the future. |
![]() |
It's possible to learn. But there's a conundrum. If the thing you're trying to learn is really complex, you'll need a really complex model. That means you'll need a lot of data to distinguish between the good models and the bad ones. Can you know when you've got enough data? Remarkably, yes... most of the time. But not until you've seen the data -- a distinction that explains why other researchers (with different learning assumptions) have claimed that this is impossible. |
![]() |
"When you're hot, you're hot; when you're not, you're not." Some find in that phrase solace for fickle fortune. Others see an engineering principle. It's the ideal for a switchable catalyst: when it's ON, it works full speed ahead, but when it's OFF, nothing doing. In this paper, we design and demonstrate a modification of the catalyst for a previously studied DNA hybrization reaction (Zhang et al 2007) that can be turned ON and OFF by an exogenous DNA strand. This is accomplished by designing the catalyst so that it has two conformations -- switchable by the exogenous strand -- one of which hides the toehold critical for initiating the catalytic reaction. I can almost hear those poor DNA molecules' lament. |
![]() |
So you'd like to program chemistry, would you? Well, it's tough. Suppose you already had a handle on the molecular design problem: "No problem," you say, "given a specification for a set of chemical reaction equations, I can construct molecules that react according to plan." Even that's not good enough, because there are limits to what a system can do when the parts are bouncing around like a bag of marbles. So, it's tough. But we can give you some hints. In this paper, we compare stochastic chemical reaction networks to a variety of known models of computation (including one of our eccentric favorites, John Conway's FRACTRAN) and examine the limits of computability. You'll meet molecular counts, probabilities, vector addition systems, primitive recursive functions... and discover just when you can and can't perform uniform computation with chemistry. And all this is for well-mixed solutions with a fixed number of chemical species -- if you want to include polymers and geometrical structures, that's a whole other game. Maybe easier, actually. |
![]() |
To form a good crystal, must the monomer unit be a compact, well-folded structure? Or can it be a floppy mess? Peng shows that what's really important is the structure in the context of the resulting crystal -- so yes, a monomer can be a floppy mess, and it's still alright! This philosophy leads to the "single-stranded tile" (SST) motif. Because each tile consists of just one strand with four domains, each specifying connectivity in one of the four lattice directions. This allows a straightforward programming of crystals ribbons and tubes with specified width. Lovely. |
![]() |
With the pieces floating all over the place, how can self-assembly be restricted just to the active attachment sites where you want it? How can you prevent the pieces from sticking to each other to soon? Well, one possibility is to put a lock on their binding sites, and design the active attachment sites (and no others!) to serve as a key. Here, we present two possible implementations of this concept for the algorithmic self-assembly of DNA tiles. They are analyzed by theory and simulation; surprisingly, for our designs, locking both the input and output sides of tiles provides substantially greater error suppression than just locking one side. |
![]() |
Biology is robust. Step on it, squash it, zap it, shake it -- odds are, it will just keep going. What makes biological systems so robust, and what are the implications of this robustness? To study this question, forget the arms and legs and blood and guts -- biological organisms are just chemistry, intricate networks of chemical reactions. What's special about robust chemical reaction networks? That should be a simpler question. In this paper, David makes a surprising connection: robustness allows stochastic chemical reaction networks to be simulated efficiently. In fact, this insight leads to an elegant and fast algorithm for simulating stochastic chemical reaction networks and even results in formal statements limiting how fast a simulator could possibly be. The result provides a new and fundamental framework for analyzing the efficiency of stochastic simulation algorithms. The bottom line is, robustness doesn't only make life easier for the organism, it makes life easier for those of us studying the organism. |
![]() |
In biological morphogenesis, genetic information is expressed as biochemical processes that create an organism. In algorithmic self-assembly, information in DNA is expressed as complex folding and crystallization processes that construct intricately patterned supramolecular objects. Here, we combine several techniques developed in this lab -- Erik's DNA tiles, Paul's DNA origami, Rebecca's ribbon crystals, and Rob's (as yet unpublished) origami seeds -- to self-assemble a fixed-width cellular automaton pattern related to Sierpinski triangles. (One commentator calls them "snakeskin nanobelts", an odd but evocative phrase.) Along the way, we gained some insights about assembly errors and how to prevent them -- about how algorithmic crystals aggregate and grow together and about lattice defect and computation error rates. |
![]() |
The entropy of the universe is always increasing. That sounds like a force -- something that keeps increasing can push something else, can't it? The problem, of course, is that using entropy to do work sounds like trying to plow a field by herding stray cats -- it just ends in chaos. But does it have to? Nope: chemists, in fact, are quite familiar with entropy-driven reactions. In this paper, we show how to design systems of DNA molecules with catalytic reactions that are driven by entropy. And it doesn't end in chaos, far from it: we argue that our reactions can be wired together into arbitrary analog or digital circuits, which means they can process information and thus create order. So entropy can drive the production of order? Yup. No wonder the universe is such a beautiful place! |
![]() |
Some people think of chemistry as a bag of colored marbles. Shake really hard. When the marbles hit each other, they change colors, according to rules. So there's a bit of structure, but it's a chaotic mess -- at any given moment, it's anyone's guess what will happen next. Can chemistry do computation, then? At least since Jacob & Monod, and perhaps before, it has been generally recognized that biochemical systems, such as genetic regulatory networks, can operate much like electrical circuits -- the concentration of some chemical species can carry an ON/OFF signal. Arbitrary digital circuit logic can be performed. If enough marbles turn green, we'll say the output was "1". Bennett even realized that if we string the marbles together like a necklace of beads, then Turing-universal computation can be performed. That's strictly more powerful, theoretically, than digital circuits. What we show here is that finite stochastic chemical reaction networks -- bags of marbles without strings -- can also perform Turing-universal computation. Reasonably quickly, too! This result holds if we accept some probability, no matter how small, that the chemistry will produce the wrong output... but remarkably, the result fails if we insist that the chemistry always and without exception produces the correct output. It pays to be tolerant, if even ever so slightly. |
![]() |
Rock candy is my favorite form of sugar. It sparkles with beautiful flat facets. It's tasty. And it practically makes itself, growing right on the stick -- almost as you watch! All you have to do is slowly cool a pot-full of supersaturated hot sugar water. Ahh, the beauty of crystal growth. But if the solid form is thermodynamically favorable (it is growing, after all) why does it grow just on the stick, and not precipitate out of solution everywhere? The answer, of course, is that there is a kinetic barrier to nucleation: small crystals are more likely to melt than to grow. But your stick had on it some pre-formed large crystals (sugar grains) that were more likely to grow than to melt. Seed crystals, we call them. They suck up all the sugar, and voila, big beautiful rock candy! All goes to show: control over nucleation is wonderful to have. But rock candy is just a metaphor. Control over nucleation is critical for a whole slew of biological processes -- such as the growth of the cytoskeleton -- where a structure needs to be built in the right place and at the right time, but nowhere else. It's natural to think that control over nucleation will be key to self-assembling molecular technologies as well. That's what our paper is about. We design DNA tiles that crystallize into ribbons -- but there's a big kinetic barrier to nucleation unless we add a seed. Hopefully, this mechanism for controlling nucleation can be used as a general subroutine for programming complex crystal growth processes. |
![]() |
A single bit flip in the source code and the program won't compile. A single mistake in calculating the digits of pi, and every subsequent digit is garbage. This is what happens in brittle algorithms: corrupted information gets used to compute more information, which is thus corrupted, and the infection grows. Our initial efforts to embed algorithmic logic into DNA self-assembly were brittle. A single DNA tile that attached at the wrong site would utterly disrupt the algorithmic pattern being formed. So we invented a way to design "proofreading" tile sets that allowed a few errors to occur without disrupting algorithmic growth. Unfortunately, our construction turned out to still have problems when crystals grew with large facets -- it did not suppress errors occurring from spontaneous nucleation of new layers of tiles on the facets. A few years ago, Ho-Lin and Ashish proposed an improved construction for DNA tile sets, which they call "snaked proofreading" tile sets, that has excellent theoretical properties in this regard. Here, Ho-Lin and Rebecca demonstrate experimentally that the fundamental principle of the new construction is sound: we can design tile sets that logically suppress nucleation on facets. In effect, those nasty bit flips will now most of the time just flip right back. |
![]() |
Can a DNA molecule walk? Can we design a molecular motor from scratch? If so, how could it work? Consider macroscopic motors for a minute: they come in all varieties, using all sorts of principles -- internal combustion engines, steam engines, Wankel rotary engines, electric motors, pneumatic motors, solenoids, rockets, jets... each best suited to different tasks. The molecular world has similar diversity. In biology, we see rotary motors like ATPase and the flagellar motor, walking motors like kinesin, linear motors like RNA polymerase and the ribosome, and waving motors like cilia. Not to mention muscle. Among the most mind-bending are the polymerization motors of pathogenic bacteria, such as Rickettsia and Listeria, that live inside eukaryotic host cells. By displaying proteins that catalyze the polymerization of the host cell's actin, these bacteria create a "comet tail" behind them that pushes them forcefully through the cell and even into neighboring cells. This was the motor principle targeted in our work, which was lead by Niles' group. Perhaps most fascinating is that the DNA polymers grow by insertion between the polymer tail and the DNA catalyst strands anchored on the "surrogate bacterial cell". This insertion takes place by a series of conformational rearrangements without ever the two sides losing their strong attachment to each other. Quite a dance! |
![]() |
In 1965, Graham Cairns-Smith proposed that the first (and simplest) Darwinian entities were clay crystals, and that as they evolved to reproduce faster and more robustly, they catalyzed organic synthesis... and eventually their invented technology took over the reins... and this "genetic takeover" resulted in us, their organic brainchildren. Sounds crazy, doesn't it? Especially since crystal evolution hasn't been demonstrated experimentally. Well, read his books. In any case, a few years ago we realized that we might be able to evolve our DNA crystals in the lab -- it's not clay, but the principles might be similar. So we asked ourselves, what might be an interesting selective pressure that our crystals could be subjected to, so that evolution leads to increasing functional complexity? The answer was quite surprising: without any novel chemistry or physics, purely informational aspects of crystal growth are in principle sufficient for a crystal to "intelligently" respond to the presence or absence in the environment of material required for growth -- and crystals containing information that enables them to respond more adeptly will grow and reproduce faster. This is intriguing. |
![]() |
If you cut a crystal, does it bleed? Normally, you'd think not... but algorithmic crystal growth has some surprisingly life-like properties... so it makes you wonder. For example, there are crystal monomer solutions that, depending on the information contained in a seed crystal, can grow into any algorithmically describable shape -- much like a biological developmental program. But what if the growth process is error-prone? What if the crystal suffers damage from a hostile environment (i.e. you cut it)? Previous work showed that even under such conditions (considered separately), algorithmic crystals can grow and maintain their integrity. But different constructions and different arguments were used in the two cases. Here, we introduce a new construction and a new proof technique that shows that both challenges can be met simultaneously! No blood. |
![]() |
If you're thinking about the molecular machines that may have inhabited the RNA World in times gone by, if you're planning on engineering a modern-day DNA World, or even if you're just curious, you may want to know what happens when you toss a collection of strands together in a test tube. Will they fold up? Into what shapes? Will they interact? Form multi-stranded assemblies? With what concentrations? For single strands, existing dynamic programming algorithms can find the minimum-free-energy fold and even calculate the partition function exactly, so that the probabilities of alternative folds can be compared. (At least, with respect to the standard nearest-neighbor model of secondary structure.) The bimolecular case has also been tackled. Here, Niles & his boys lead the charge to solve the full problem for any number of strands. An exact solution, with fast algorithms, is provided. The math is elegant, involving symmetry groups, convex optimization, and what seem like remarkable "coincidences". Nice. |
![]() |
If you can effectively describe a shape -- e.g. write a computer program that draws it -- then can the shape be self-assembled by spontaneous local processes from just a few kinds of pieces? The answer may be "no" if you are concerned about the exact size of the shape, but if it is only the form of the shape that matters, then the answer is always "yes". In fact, the Kolmogorov complexity of a shape provides upper and lower bounds for the number of tile types necessary to self-assemble it (at some scale). Proof of this result makes use of a construction for programmable "blocks" that execute a Turing machine simulation to guide the growth process, and introduces a new proof technique for establishing when growth is independent of the order in which tiles are added. This work suggests that scale plays the same role in self-assembly as does time in computability: when you ignore it and just ask, "can it be done at all?", then the theory becomes clean and elegant. |
![]() |
How simple can life be? To some, the answer to this question is a small bungalow on Fiji, a hammock, and the sound of waves sloshing and slurping against the sand. Others are more inclined to think about the origin of life, about minimal organisms, about what it would take to create life from scratch. Being broad-minded, we appreciate both perspectives. While our research on the first aspect of the problem is still in progress, this paper reports our efforts in the second direction -- efforts to recreate, in simplified form, the information-processing heart of cellular processes: genetic regulatory networks. We do it in vitro using just DNA, RNA, and a small handful of enzymes... though not without encountering a few challenges and surprises! |
![]() |
It's a game of molecular dominoes. One falls, bumps into the next, triggering its fall. Sometimes two separate cascades converge on a single domino, and only when both hit at once does the domino fall. Or maybe either input is enough. In fact, you can build any kind of logic circuit out of dominoes. You can make a domino cascade that multiplies binary numbers, if you want. It's fun. But maybe not so deep scientifically, not so useful. But suppose you now do the same thing with molecules, say DNA strands. Now you have a way to embed logical control in chemistry, to sense and react to molecular events. Now you're at the same frontier of molecular information processing that biology has been exploiting and refining for almost 4 billion years. We're a bit behind in the game. |
![]() |
DNA machines need fuel. Whether the machine is a motor, a manipulator, or a computer, if it is to be capable of driving a system to perform a periodic behavior -- and what respectable machine can't do that? -- then it needs fuel. ATP is an excellent fuel: small, high energy content, simple... but, how can we put it?... holding it requires small hands. Not even the rain has such small hands. So certainly not human-designed DNA machines, yet. So for DNA machines, we will make do with a bigger fuel -- but one that can be specifically targeted. That, more or less, was the motivation for this work. (The original work pursuing this line of reasoning was Turberfield et al 2003 [ref 1] and further progress was made in "DNA Hybridization Catalysts and Catalyst Circuits" [ref 37, see below]. Here, we finally have a system that works quite nicely.) {apologies to ee cummings} |
![]() |
Deborah's group decides to ligate DNA nanotubes. They are stable under a fluorescence microscope up to 70C and, unlike noncovalent tubes, resist opening under AFM. That's very, very useful! |
![]() |
DNA replication, when one becomes two, marks the fundamental moment in life. How simple can the machinery for replication be? Can it be designed from scratch? Dave and Bernie present here a proposal for a clock-work machine -- an information-bearing superstructure made entirely of multi-base DNA building-block motifs -- that replicates its overall structure when exposed to the proper sequence of raw building blocks and fuel. So it's theoretically possible. This throws down the gauntlet: Can an autonomous scheme be devised? Can a simpler implementation be found, one practical enough for experimental demonstration? Can a chemical self-replicator be rationally designed? |
![]() |
Paul sends a swarm of staple strands to tie viral DNA in knots...thereby self-assembling 100 x 100 nm objects with roughly 6 nm resolution from the 7 kilobase single-stranded genomic DNA of M13mp18. Rectangles. Squares. Triangles. Stars. Even a smiley-face. About 50 billion copies of each, in a typical reaction, and with very high yields. It works like magic. We did some calculations... Paul's smiley faces constitute the most concentrated happiness ever experienced on earth. Each spot in such a structure contains a unique address and can be addressed as such by DNA hybridization, allowing one to "write" on the DNA origami objects. Words. Pictures. Snowflakes. A map of North and South America. We did some more calculations... Paul probably made more maps than have ever been produced in the history of mankind -- we're definitely talking quantity over quality here. The applications of this technology are likely to be less whimsical. For example, it can be used as a "nanobreadboard" for attaching almost arbitrary nanometer-scale components, and there are few other ways to obtain such precise control over the arrangement of components at this scale. You'll never look at M13 phage DNA the same way again... |
![]() |
Paul talks a little about the design software and future possibilities. |
![]() |
Paul makes a DNA origami especially for Ned Seeman, and sketches how polygonal networks and polygonal three-dimensional structures can be created. |
![]() |
Wandering in the dreamy fields of abstraction and metaphor, I like to think of algorithmic self-assembly as a toy model of organismal development. A tile set can direct the growth of an algorithmic pattern from a seed assembly "in the same way" as genetic DNA directs the growth of a salamander from an egg. Similar principles seem to be at work in both systems: information flow is essential, sophisticated programming is possible, the physics and chemistry is messy but fault-tolerance can be achieved. The adult salamander, however, can do something that has not previously been shown in algorithmic self-assembly -- after losing its tail, the salamander can regrow a new one! Perhaps more remarkable, actually, is the ability of almost all multicellular organisms to heal severe wounds within their bodies. In this paper I investigate a (somewhat) analogous issue for algorithmic self-assembly, and show that self-healing tile sets are possible. |
![]() |
Here we demonstrate crystals that count and tubes that copy. This is interesting as the second example of algorithmic self-assembly. Counting is a useful primitive for bottom-up fabrication tasks such as constructing a memory chip with address demultiplexers. Copying is a useful primitive for Darwinian evolution. The error rates in this work, however, strongly motivate research into "proofreading" and other methods for fault-tolerant self-assembly. |
![]() |
Presumably, there is some cost for obtaining low error rates during self-assembly. Maybe assembly must be slower. Or the same pattern can be assembled more reliably, but using more tile types. Or the pattern must be assembled at a larger scale. Or all of the above. In fact, Chen and Goel (2005) proved that if you can accept "just a little slower", "just a few more tiles", and "just a slightly larger scale", then any self-assembled pattern can be made as reliable as you want. But suppose you insist that the pattern remain at exactly the original scale, as do Reif, Sahu, and Yin (2005) ? We give evidence that for some patterns -- which we term "robust" -- it is still possible to do arbitrarily strong proofreading with modest costs in speed and tile set size, but for other patterns -- which we term "fragile" -- it is not... at least, not with any proofreading scheme that uses a certain kind of redundancy. |
![]() |
What is the simplest physical system capable of extensive Darwinian evolution?
Such a system might be a candidate for the origin of life.
One would expect that it must have certain features:
it must be capable of of reproducing as an organism to
populate an environmental niche; of storing an unbounded amount of
information in its "genome"; of copying that information with low
error rates; and of expressing that information as a phenotype with
selective advantages or disadvantages.
Graham Cairns-Smith has argued, since the 1960's, that mineral
crystals (such as clays) provide the most compelling example of a
simple and abundant physical system capable of extensive Darwinian
evolution: Clay crystals can store information as a pattern of
inhomogeneities that are propagated from layer to layer, with few
errors; they can reproduce by random fragmentation; and they can
express a variety of morphological phenotypes.
To be blunt, he proposes that your
great-great-great-great-great-....-grandparent was a lump of clay. I
think he may be right. But the evidence is thin: to date, no one has
managed to catch clay in the act of replicating information or
evolving. In this paper, we propose using experimentally-tractable
DNA tile self-assembly to investigate how evolution could conceivably
work on information-bearing crystals.
Won the conference best paper award! (of 94 papers) |
![]() |
In a series of experimental papers, Benenson et al demonstrated how information encoded in DNA can be processed autonomously by a molecular automaton consisting of a type IIS restriction enzyme (FokI) and a collection of DNA "rule molecules". They constructed several two-symbol, two-state finite state machines as well as several 4-input AND/NOT gates. Ultimately, applications envisioned include diagnosis of mRNA and delivery of a theraputic molecule. Here, we formalize the class of logical computations that can be performed by such systems and show that, in our abstraction, the finite-input computations that can be performed by Benenson automata can be exactly characterized by a connection to branching programs -- which indicates that (if restriction enzymes slightly more powerful than FokI can be found) Benenson automata are considerably more powerful than we had originally imagined! |
![]() |
Our first demonstration of algorithmic crystals, wherein molecularly-encoded information directs the growth process to create a complex pattern. The DNA crystals are, at the molecular level, a two-dimensional woven fabric of short DNA strands. Both because this programmable growth could be considered a super-simplified toy model of organismal development, and because DNA is the central information molecule in biology, I like to call it "weaving the tapestry of life". This is a substantial personal victory for me: I proposed that this should be possible in 1995 as a graduate student -- nearly 10 years later, Paul's efforts made it actually happen. |
![]() |
DNA tiles designed to make sheets sometimes roll up into tubes that are abstractly analogous to protein microtubules that self-assemble from tubulin. Way cool! Fortuitously discovered during Paul's work on the DNA Sierpinski triangles, DNA nanotubes have opened up a whole host of interesting possibilities that we never dreamed of before... |
![]() |
Is it possible to design biochemical circuits in the same way we design electrical circuits -- where the information processing and dynamics is what we're interested in, rather than the chemistry per se? What kinds of dynamical and decision-making behavior can be achieved this way? In this paper, we give a concrete proposal for how to use synthetic DNA molecules and a few enzymes (RNA polymerase and some ribonucleases) to make general-purpose threshold gates that can be wired together to make arbitrary circuits. The mathematical description turns out to be closely related to continuous-time analog neural networks. We analyze a few example circuits. |
![]() |
Energy stored in metastable configurations of DNA molecules can in principle be used to power DNA nanomachines and other molecular processes. Essential to using this energy productively is the ability to regulate how and when the energy is released. Starting with a previously-demonstrated DNA hybridization catalyst for doing exactly that, we introduce schemes for an improved catalyst and for networks of catalysts that operate as computing circuits. We show preliminary experimental results investigating these proposals -- which uncover a few interesting surprises! |
![]() |
What constitutes a good design criterion for choosing DNA sequences that (should) fold up into target nanostructures? We argue that both positive (the target structure should have low energy) and negative (all other structures should have high energy) design criteria should be considered. Secondary structure formalisms for describing and analyzing DNA molecules allow these criteria to be rigorously treated computationally. |
![]() |
Experimental studies of algorithmic self-assembly have observed error rates between 1 and 10 %, measured as the fraction of DNA tiles that fail to perfectly match their neighbors. It is difficult to perform extended computations when every elemental logical operation is so likely to be wrong. Here, we provide a general-purpose technique for designing tile sets that are robust to assembly errors. Kinetic and thermodynamic models of self-assembly are key to our investigations. It is suggested that proofreading tile sets can square the error rate for comparable physical conditions -- i.e. bring 10% to 1% and bring 1% to .01%. |
![]() |
Can DNA self-assembly be used for patterning, as a scaffold for functional devices such as molecular electronic circuits? We show that several circuit patterns, including demultiplexers, random-access memory, and Hadamard matrix transforms, can be self-assembled (in principle) from a small number of tile types. |
![]() |
Discusses our progress toward using DNA tiles to self-assembly the boundary for a self-assembled Sierpinski triangle. Experimental results show that individual tiles behave well, with the expected tile-tile interactions, but size and shape distributions of self-assembled structures were sometimes surprising and unsatisfactory. We understand this by analyzing models of 1D polymerization, for example distinguishing accretion from a seed vs. parallel aggregation. This work identifies control over nucleation as an important target for future study in algorithmic self-assembly. |
![]() |
A brief review of algorithmic self-assembly of DNA. |
![]() |
Shows that a commonly-used formulation for protein design -- amino acid sequence selection to stabilize a given backbone fold using a pairwise energy model -- is NP-Hard if the energy function is considered part of the specification of the problem (as it is in common practice, since this is the implicit formulation of the target backbone fold). Therefore, it may be advantageous to exploit specific properties of protein folding energetics, rather than relying on general-purpose (worst-case) exponential-time algorithms. |
![]() |
Can one find the minimal number of tiles required to uniquely produce a given shape? Here we prove that the problem is NP-complete. Also, for a large class of tile sets we give a method to find concentrations of tiles that yield the fastest time of self-assembly. |
![]() |
Overview discussion of Wang tiles, computation by self-assembly, and previous experimental results, as well as some additional experiments using the system described in Nature 394, 539-544, Aug. 6, 1998 -- see below. |
![]() |
This brief review, written by Ned Seeman, describes the state of DNA nanotechnology in the year 2000, including the recent results described in Nature 394, 539-544, Aug. 6, 1998 -- see below. |
![]() |
How much computation can be done with linear self-assembly? With some tricky strand routing through complex DNA tiles, you can do a lot! |
![]() |
Are there small self-assembly programs for building squares of a particular size? Yes! Also contains a nice presentation of our model of self-assembly. |
![]() |
2D crystals can be made from fancier DNA complexes that allow for baroque strand routing. |
![]() |
Here Paul used macroscopic plastic tiles to test ideas about algorithmic self-assembly. The plastic tiles self-assemble at the interface of oil and water, and hydrophilic and hydrophobic patches on the edges of the tiles mediate the specific binding interactions between them. Tiles in this paper encode binding interactions for creating Sierpinski triangles as well as Penrose tilings. Paul had mild success creating unnucleated patterns with Sierpinski tiles and perhaps created the most complex set of specific capillary bonds ever made. |
![]() |
Describes experimental characterization of three-helix DNA complexes, shows that multiple TX can assemble around a template strand, and illustrates how they could be used to self-assemble an addition table. |
![]() |
Update on Roweis et al: shows that misclassification can be ameliorated by repeated separation, and strand loss can be recovered by PCR, resulting in reliable computation at the cost of moderately more time and space. *IF* there is no sequence-specific bias to the errors. |
![]() |
Periodic lattices of DNA double-crossover molecules, of the type proposed in "On the Computational Power of DNA Annealing and Ligation", are demonstrated experimentally. |
![]() |
A kinetic and thermodynamic analysis of error rates in a simple model of the two-dimensional self-assembly process proposed in my first paper on this topic, "Annealing and Ligation". |
![]() |
Some comments and extensions to the "one-pot" PCR-like technique of [Hagiya,Arita,Kiga,Sakamoto,Yokoyama in DNA Based Computers III], including a a way for a single DNA strand to simulate the execution of circuit. |
![]() |
This follow-up to "Annealing and Ligation" (below) examines the relationship of DNA self-assembly to formal language theory -- finding ties to regular, context-free, and recursively enumerable languages. Additionally, preliminary laboratory investigations into the self-assembly required for universal computation is reported. |
![]() |
A new representation for encoding bits in DNA is presented and examined, and
generally applicable methods for decreasing separation error rates are
discussed.
[in DNA Based Computers II, pgs 1-29, 1998 (26 pages):
stickers.pdf, 611 KB, or
stickers.ps.gz, 231 KB, or
stickers.ps, 1.2 MB] This appeared in journal form as two papers:
|
![]() |
An algorithm for breaking DES is designed for the Stickers model. Size, space, and error rates of the resulting machine are considered. |
![]() |
Here I show how one might create a "one-pot" mixture of DNA which can perform universal computation by self-assembly of DNA double crossover units. A.k.a. "weaving the tapestry of life". |
![]() |
Here I show some limits on what can be computed using some proposed operations on DNA. In particular, how affinity separation w/ or w/o amplification relates to branching programs and non-deterministic branching programs. These limits have since been overcome by the inclusion of additional operations, such as ligation. |
![]() |
Here Paul gives a construction for simulating Minsky's 4 symbol, 7 state universal Turing machine using DNA, the type IIS restriction enzyme Fok I, and ligase. The encoding of machine state and current symbol used in this paper was later used by Kobi Benenson of Udi Shapiro's group to create DNA finite state machines and hence such machines have been called Rothemund-Shapiro machines. |
![]() |
Excitable media support directed wave propagation that, unlike linear waves, are directed and self-sustaining but annihilate each other when they collide. Examples include, in one dimension, action potential propagation along a neuron; in two dimensions, a slow-moving fire burning across a quick-growing plain of grass leaving a thin zone of burnt grass in its wake; in three dimensions, the electrical-contractile waves that make heart muscle beat. The Belousov-Zhabotinski reaction provides a simple chemical analog in any dimension, and the Greenberg-Hastings cellular automaton provides a simple discrete model that exhibits many of the same qualitative behaviors. The most striking behavior of excitable media in two dimensions is that after part of a wavefront is rubbed out, a spiral pattern develops as the wavefront swirls around its end. In three dimensions, one observes a wavesheet rather than a wavefront -- and if one rubs out a hole in that wavesheet, what develops is a remarkable three dimensional spiral. In fact there are an infinite variety of topologically distinct three dimensional spirals, each corresponding to the Seifert surface of a distinct knot. We simulated two such spirals in a three dimensional Greenberg-Hastings cellular automaton. |
DNA Based Computers: DIMACS Workshop, held April 4, 1995 (eds Richard J. Lipton and Eric B. Baum) American Mathematical Society, 1996.
DNA Based Computers II: DIMACS Workshop, held June 10-12, 1996 (eds Laura F. Landweber and Eric B. Baum) American Mathematical Society, 1998.
DNA Based Computers III: DIMACS Woskhop, held June 23-25, 1997 (eds Harvey Rubin and David H. Wood) American Mathematical Society, 1999.
Proceedings of the Fourth DIMACS Meeting on DNA Based Computers, held at the University of Pennsylvania, June 16-19, 1998. (Never published as a book; selected articles appear in https://www.sciencedirect.com/journal/biosystems/vol/52/issue/1 but alas not mine.)
DNA Based Computers V: DIMACS Workshop, held June 14-15, 1999. (eds. Erik Winfree and David K. Gifford) American Mathematical Society, 2000.
DNA Based Computers VI: held June 13-17, 2000. (eds. Anne Condon and Grzegorz Rozenberg) Lecture Notes in Computer Science 2054, Springer, 2001.
Send comments to
winfree@caltech.edu