Matthew Griffin, described as “The Adviser behind the Advisers” and a “Young Kurzweil,” is the founder and CEO of the World Futures Forum and the 311 Institute, a global Futures and Deep Futures consultancy working between the dates of 2020 to 2070, and is an award winning futurist, and author of “Codex of the Future” series. Regularly featured in the global media, including AP, BBC, Bloomberg, CNBC, Discovery, RT, Viacom, and WIRED, Matthew’s ability to identify, track, and explain the impacts of hundreds of revolutionary emerging technologies on global culture, industry and society, is unparalleled. Recognised for the past six years as one of the world’s foremost futurists, innovation and strategy experts Matthew is an international speaker who helps governments, investors, multi-nationals and regulators around the world envision, build and lead an inclusive, sustainable future. A rare talent Matthew’s recent work includes mentoring Lunar XPrize teams, re-envisioning global education and training with the G20, and helping the world’s largest organisations envision and ideate the future of their products and services, industries, and countries. Matthew's clients include three Prime Ministers and several governments, including the G7, Accenture, Aon, Bain & Co, BCG, Credit Suisse, Dell EMC, Dentons, Deloitte, E&Y, GEMS, Huawei, JPMorgan Chase, KPMG, Lego, McKinsey, PWC, Qualcomm, SAP, Samsung, Sopra Steria, T-Mobile, and many more.
WHY THIS MATTERS IN BRIEF
The future of computing is a lot more sci-fi than you might think and we’re building that future today …
Love the Exponential Future? Join our XPotential Community, future proof yourself with courses from XPotential University, read about exponential tech and trends, connect, watch a keynote, or browse my blog.
Name your favourite computing platform … I bet you didn’t say a bacteria. And yet when we look into the deep future of computing increasingly the computing platforms we use will have less silicon and more DNA, and that’s before we discuss other future computing platforms such as the development of new chemical, liquid, neuromorphic, and quantum computers, some of which could pack more computing power than we have on the entire planet today into something the size of a test tube …
E. coli thrives in our guts, sometimes to unfortunate effect, and it facilitates scientific advances – in DNA, biofuels, and Pfizer’s covid vaccine, to name but a few. Now this multitalented bacterium has a new trick: it can solve a classic computational maze problem using distributed computing – dividing up the necessary calculations among different types of genetically engineered cells.
This neat feat is a credit to synthetic biology, which aims to rig up biological circuitry much like electronic circuitry and to program organic cells as easily as computers.
The maze experiment is part of what some researchers consider a promising direction in the field: rather than engineering a single type of cell to do all the work, they design multiple types of cells, each with different functions, to get the job done. Working in concert, these engineered microbes might be able to “compute” and solve problems more like multi-cellular networks in the wild, and as I’ll discuss later in this article these sci-fi like biological computers, some of which have already been used elsewhere to store and replay information, could pack huge amounts of computing power into tiny tiny packages.
So far though, for better or worse, fully harnessing biology’s powerful designs have eluded, and frustrated, synthetic biologists.
“Nature can do this – think about a brain – but we don’t yet know how to design at that overwhelming level of complexity using biology,” says Pamela Silver, a synthetic biologist at Harvard.
The study using E. coli as maze solvers, led by biophysicist Sangram Bagh at the Saha Institute of Nuclear Physics in Kolkata, is a simple and fun “toy” problem. But it also serves as a proof of principle for distributed computing among organic cells, demonstrating how more complex and practical computational problems might be solved in a similar way. And, if this approach works at larger scales, it could unlock applications for everything from pharmaceuticals to agriculture to space travel.
“As we move into solving more complex problems with engineered biological systems, spreading out the load like this is going to be an important capacity to establish,” says David McMillen, a bioengineer at the University of Toronto.
Unsurprisingly getting E. coli to solve the maze problem required some ingenuity. The bacteria didn’t wander through a palace labyrinth of well-pruned hedges. Rather, the bacteria analyzed various maze configurations. The setup: one maze per test tube, with each maze generated by a different chemical concoction.
The chemical recipes were abstracted from a 2×2 grid representing the maze problem. The grid’s top left square is the start of the maze, and the bottom right square is the destination. Each square on the grid can be either an open path or blocked, yielding 16 possible mazes.
Bagh and his colleagues mathematically translated this problem into a truth table composed of 1s and 0s, showing all possible maze configurations. Then they mapped those configurations onto 16 different concoctions of four chemicals. The presence or absence of each chemical corresponds to whether a particular square is open or blocked in the maze.
The team engineered multiple sets of E. coli with different genetic circuits that detected and analyzed those chemicals. Together, the mixed population of bacteria functions as a distributed computer; each of the various sets of cells perform part of the computation, processing the chemical information and solving the maze.
Running the experiment, the researchers first put the E. coli in 16 test tubes, added a different chemical-maze concoction in each, and left the bacteria to grow. Then, after 48 hours if the E. coli detected no clear path through the maze – that is, if the requisite chemicals were absent – then the system remained dark. If the correct chemical combination was present, corresponding circuits turned “on” and the bacteria collectively expressed fluorescent proteins, in yellow, red, blue, or pink, to indicate solutions.
“If there is a path, a solution, the bacteria glow,” says Bagh.
What Bagh found particularly exciting was that in churning through all 16 mazes, the E. coli provided physical proof that only three were solvable.
“Calculating this with a mathematical equation is not straightforward,” he says. “With this experiment, you can visualize it very simply.”
Bagh envisions such a biological computer helping in cryptography or steganography – the art and science of hiding information – which use mazes to encrypt and conceal data, respectively. But the implications extend beyond those applications to synthetic biology’s loftier ambitions.
The idea of synthetic biology dates to the 1960s, but the field emerged concretely in 2000 with the creation of synthetic biological circuits (specifically, a toggle switch and an oscillator) that made it increasingly possible to program cells to produce desired compounds or react intelligently within their environments.
Biology, however, has not been the most cooperative of collaborators. One limiting factor is how many changes you can make to a cell without destroying its viability. “The cell has its own interests,” says McMillen, whose lab is developing a yeast-based system that detects malaria antibodies in blood samples and a similar system for covid. When inserting human-designed components into a biological system, he says, “you’re fighting against natural selection and entropy, which are two of the biggies in terms of forces of nature.”
If a cell is overloaded with too many doodads, for instance, there’s risk of interference and crosstalk—hindering performance and limiting the system’s capabilities. With the computational maze solvers, Bagh says, the algorithm could have been programmed into only one type of E. coli cell. But the system ran best when the necessary circuit functions were distributed among six types of cells.
“There is a physical limit on how many genetic parts can be used in a single cell,” says Karen Polizzi, a chemical engineer at Imperial College London, who develops cellular biosensors to monitor the manufacture of therapeutic proteins and vaccines. “This limits the sophistication of the computing concepts that can be developed.”
“Distributed computing might actually be a way to achieve some of [synthetic biology’s] really lofty goals,” she adds. “Because there’s no way you’re going to get a cell to completely do a complex task by itself.”
Chris Voigt, a synthetic biologist at MIT, and editor in chief of ACS Synthetic Biology, which published Bagh’s result, believes distributed computing is the direction that synthetic biology needs to go.
In Voigt’s view, grand ambitions for microbial cells are justified – and he’s run calculations to back that up.
“One teaspoon of bacteria has more logic gates in it than, I think, 2 billion Xeon processors,” he says. “And it has more memory in the DNA than the entire internet. Biology has incredible computing capacity, and it does it by distributing it over an enormous number of cells.” But there are caveats: “A gate takes 20 minutes to process, so it’s very slow.”
Last year, Voigt and collaborators succeeded in programming DNA with a calculator’s algorithm and generating a digital display with fluorescing E. coli. They used software created by Voigt’s lab, called Cello. Cello takes files from Verilog—a programming language used to describe and model circuits—and converts them into DNA, so the same sorts of configurations can be run in cells. All the circuitry for the E. coli calculator, however, was crammed into a single type of cell, a single colony.
“We’ve kind of hit the limitation on that,” he admits. “We need to figure out how to make larger designs.”
Even if researchers wanted to run something as low-fi by today’s standards as the Apollo 11 guidance systems in bacteria, Voigt says, it couldn’t be done in a single engineered cell.
“The capacity is there,” he says. “We just need ways to break up the algorithm across cells and then link cells up to efficiently share information so that they can collectively perform the calculation.”
Indeed, Voigt questions whether directly mimicking traditional electronic computing is the best approach for harnessing biology’s computational power and solving complex bio-based problems.
In his search for the right approach, Bagh also recently devised a type of artificial neural network architecture for bacterial hardware. And he’s interested in exploring an approach that involves “fuzzy logic,” going beyond the constraints of binary 0s and 1s toward a continuum more aligned with the noise and mess of living biological systems.
Synthetic biologist Ángel Goñi-Moreno, at the Technical University of Madrid, is thinking along similar lines.
“If we are going to play with living technology, we need to play by the rules of living systems,” he says.
Goñi-Moreno imagines breaking with the electronic circuit analogy by capitalizing on how cells sense and respond and adapt to their environment, using natural selection itself as a tool to push computational designs along. Evolution, he says, is a biological process that computes information over time, optimizing cellular systems to accomplish a diversity of tasks.
Goñi-Moreno believes this approach could ultimately culminate in what he calls “cellular supremacy.” The term draws a deliberate parallel with “quantum supremacy” – the point at which quantum computers exceed the capabilities of conventional computing in certain domains.
Biocomputers that have evolved to such an extent, Goñi-Moreno says, might offer superior problem-solving savvy in areas such as enhancing agricultural production, think soil bacteria that can adjust the chemicals they make based shifting conditions, and targeting disease therapeutics.
Just don’t expect E. coli to help surf the internet just yet – for that we’ll still need good old fashioned computers.