Scientists Want to Make Robots with Genomes that Help Grow their Minds
Lina Zeldovich has written about science, medicine and technology for Popular Science, Smithsonian, National Geographic, Scientific American, Reader’s Digest, the New York Times and other major national and international publications. A Columbia J-School alumna, she has won several awards for her stories, including the ASJA Crisis Coverage Award for Covid reporting, and has been a contributing editor at Nautilus Magazine. In 2021, Zeldovich released her first book, The Other Dark Matter, published by the University of Chicago Press, about the science and business of turning waste into wealth and health. You can find her on http://linazeldovich.com/ and @linazeldovich.
One day in recent past, scientists at Columbia University’s Creative Machines Lab set up a robotic arm inside a circle of five streaming video cameras and let the robot watch itself move, turn and twist. For about three hours the robot did exactly that—it looked at itself this way and that, like toddlers exploring themselves in a room full of mirrors. By the time the robot stopped, its internal neural network finished learning the relationship between the robot’s motor actions and the volume it occupied in its environment. In other words, the robot built a spatial self-awareness, just like humans do. “We trained its deep neural network to understand how it moved in space,” says Boyuan Chen, one of the scientists who worked on it.
For decades robots have been doing helpful tasks that are too hard, too dangerous, or physically impossible for humans to carry out themselves. Robots are ultimately superior to humans in complex calculations, following rules to a tee and repeating the same steps perfectly. But even the biggest successes for human-robot collaborations—those in manufacturing and automotive industries—still require separating the two for safety reasons. Hardwired for a limited set of tasks, industrial robots don't have the intelligence to know where their robo-parts are in space, how fast they’re moving and when they can endanger a human.
Over the past decade or so, humans have begun to expect more from robots. Engineers have been building smarter versions that can avoid obstacles, follow voice commands, respond to human speech and make simple decisions. Some of them proved invaluable in many natural and man-made disasters like earthquakes, forest fires, nuclear accidents and chemical spills. These disaster recovery robots helped clean up dangerous chemicals, looked for survivors in crumbled buildings, and ventured into radioactive areas to assess damage.
Now roboticists are going a step further, training their creations to do even better: understand their own image in space and interact with humans like humans do. Today, there are already robot-teachers like KeeKo, robot-pets like Moffin, robot-babysitters like iPal, and robotic companions for the elderly like Pepper.
But even these reasonably intelligent creations still have huge limitations, some scientists think. “There are niche applications for the current generations of robots,” says professor Anthony Zador at Cold Spring Harbor Laboratory—but they are not “generalists” who can do varied tasks all on their own, as they mostly lack the abilities to improvise, make decisions based on a multitude of facts or emotions, and adjust to rapidly changing circumstances. “We don’t have general purpose robots that can interact with the world. We’re ages away from that.”
Robotic spatial self-awareness – the achievement by the team at Columbia – is an important step toward creating more intelligent machines. Hod Lipson, professor of mechanical engineering who runs the Columbia lab, says that future robots will need this ability to assist humans better. Knowing how you look and where in space your parts are, decreases the need for human oversight. It also helps the robot to detect and compensate for damage and keep up with its own wear-and-tear. And it allows robots to realize when something is wrong with them or their parts. “We want our robots to learn and continue to grow their minds and bodies on their own,” Chen says. That’s what Zador wants too—and on a much grander level. “I want a robot who can drive my car, take my dog for a walk and have a conversation with me.”
Columbia scientists have trained a robot to become aware of its own "body," so it can map the right path to touch a ball without running into an obstacle, in this case a square.
Jane Nisselson and Yinuo Qin/ Columbia Engineering
Today’s technological advances are making some of these leaps of progress possible. One of them is the so-called Deep Learning—a method that trains artificial intelligence systems to learn and use information similar to how humans do it. Described as a machine learning method based on neural network architectures with multiple layers of processing units, Deep Learning has been used to successfully teach machines to recognize images, understand speech and even write text.
Trained by Google, one of these language machine learning geniuses, BERT, can finish sentences. Another one called GPT3, designed by San Francisco-based company OpenAI, can write little stories. Yet, both of them still make funny mistakes in their linguistic exercises that even a child wouldn’t. According to a paper published by Stanford’s Center for Research on Foundational Models, BERT seems to not understand the word “not.” When asked to fill in the word after “A robin is a __” it correctly answers “bird.” But try inserting the word “not” into that sentence (“A robin is not a __”) and BERT still completes it the same way. Similarly, in one of its stories, GPT3 wrote that if you mix a spoonful of grape juice into your cranberry juice and drink the concoction, you die. It seems that robots, and artificial intelligence systems in general, are still missing some rudimentary facts of life that humans and animals grasp naturally and effortlessly.
How does one give robots a genome? Zador has an idea. We can’t really equip machines with real biological nucleotide-based genes, but we can mimic the neuronal blueprint those genes create.
It's not exactly the robots’ fault. Compared to humans, and all other organisms that have been around for thousands or millions of years, robots are very new. They are missing out on eons of evolutionary data-building. Animals and humans are born with the ability to do certain things because they are pre-wired in them. Flies know how to fly, fish knows how to swim, cats know how to meow, and babies know how to cry. Yet, flies don’t really learn to fly, fish doesn’t learn to swim, cats don’t learn to meow, and babies don’t learn to cry—they are born able to execute such behaviors because they’re preprogrammed to do so. All that happens thanks to the millions of years of evolutions wired into their respective genomes, which give rise to the brain’s neural networks responsible for these behaviors. Robots are the newbies, missing out on that trove of information, Zador argues.
A neuroscience professor who studies how brain circuitry generates various behaviors, Zador has a different approach to developing the robotic mind. Until their creators figure out a way to imbue the bots with that information, robots will remain quite limited in their abilities. Each model will only be able to do certain things it was programmed to do, but it will never go above and beyond its original code. So Zador argues that we have to start giving robots a genome.
How does one do that? Zador has an idea. We can’t really equip machines with real biological nucleotide-based genes, but we can mimic the neuronal blueprint those genes create. Genomes lay out rules for brain development. Specifically, the genome encodes blueprints for wiring up our nervous system—the details of which neurons are connected, the strength of those connections and other specs that will later hold the information learned throughout life. “Our genomes serve as blueprints for building our nervous system and these blueprints give rise to a human brain, which contains about 100 billion neurons,” Zador says.
If you think what a genome is, he explains, it is essentially a very compact and compressed form of information storage. Conceptually, genomes are similar to CliffsNotes and other study guides. When students read these short summaries, they know about what happened in a book, without actually reading that book. And that’s how we should be designing the next generation of robots if we ever want them to act like humans, Zador says. “We should give them a set of behavioral CliffsNotes, which they can then unwrap into brain-like structures.” Robots that have such brain-like structures will acquire a set of basic rules to generate basic behaviors and use them to learn more complex ones.
Currently Zador is in the process of developing algorithms that function like simple rules that generate such behaviors. “My algorithms would write these CliffsNotes, outlining how to solve a particular problem,” he explains. “And then, the neural networks will use these CliffsNotes to figure out which ones are useful and use them in their behaviors.” That’s how all living beings operate. They use the pre-programmed info from their genetics to adapt to their changing environments and learn what’s necessary to survive and thrive in these settings.
For example, a robot’s neural network could draw from CliffsNotes with “genetic” instructions for how to be aware of its own body or learn to adjust its movements. And other, different sets of CliffsNotes may imbue it with the basics of physical safety or the fundamentals of speech.
At the moment, Zador is working on algorithms that are trying to mimic neuronal blueprints for very simple organisms—such as earthworms, which have only 302 neurons and about 7000 synapses compared to the millions we have. That’s how evolution worked, too—expanding the brains from simple creatures to more complex to the Homo Sapiens. But if it took millions of years to arrive at modern humans, how long would it take scientists to forge a robot with human intelligence? That’s a billion-dollar question. Yet, Zador is optimistic. “My hypotheses is that if you can build simple organisms that can interact with the world, then the higher level functions will not be nearly as challenging as they currently are.”
Lina Zeldovich has written about science, medicine and technology for Popular Science, Smithsonian, National Geographic, Scientific American, Reader’s Digest, the New York Times and other major national and international publications. A Columbia J-School alumna, she has won several awards for her stories, including the ASJA Crisis Coverage Award for Covid reporting, and has been a contributing editor at Nautilus Magazine. In 2021, Zeldovich released her first book, The Other Dark Matter, published by the University of Chicago Press, about the science and business of turning waste into wealth and health. You can find her on http://linazeldovich.com/ and @linazeldovich.
In the 1966 movie "Fantastic Voyage," actress Raquel Welch and her submarine were shrunk to the size of a cell in order to eliminate a blood clot in a scientist's brain. Now, 55 years later, the scenario is becoming closer to reality.
California-based startup Bionaut Labs has developed a nanobot about the size of a grain of rice that's designed to transport medication to the exact location in the body where it's needed. If you think about it, the conventional way to deliver medicine makes little sense: A painkiller affects the entire body instead of just the arm that's hurting, and chemotherapy is flushed through all the veins instead of precisely targeting the tumor.
"Chemotherapy is delivered systemically," Bionaut-founder and CEO Michael Shpigelmacher says. "Often only a small percentage arrives at the location where it is actually needed."
But what if it was possible to send a tiny robot through the body to attack a tumor or deliver a drug at exactly the right location?
Several startups and academic institutes worldwide are working to develop such a solution but Bionaut Labs seems the furthest along in advancing its invention. "You can think of the Bionaut as a tiny screw that moves through the veins as if steered by an invisible screwdriver until it arrives at the tumor," Shpigelmacher explains. Via Zoom, he shares the screen of an X-ray machine in his Culver City lab to demonstrate how the half-transparent, yellowish device winds its way along the spine in the body. The nanobot contains a tiny but powerful magnet. The "invisible screwdriver" is an external magnetic field that rotates that magnet inside the device and gets it to move and change directions.
The current model has a diameter of less than a millimeter. Shpigelmacher's engineers could build the miniature vehicle even smaller but the current size has the advantage of being big enough to see with bare eyes. It can also deliver more medicine than a tinier version. In the Zoom demonstration, the micorobot is injected into the spine, not unlike an epidural, and pulled along the spine through an outside magnet until the Bionaut reaches the brainstem. Depending which organ it needs to reach, it could be inserted elsewhere, for instance through a catheter.
"The hope is that we can develop a vehicle to transport medication deep into the body," says Max Planck scientist Tian Qiu.
Imagine moving a screw through a steak with a magnet — that's essentially how the device works. But of course, the Bionaut is considerably different from an ordinary screw: "At the right location, we give a magnetic signal, and it unloads its medicine package," Shpigelmacher says.
To start, Bionaut Labs wants to use its device to treat Parkinson's disease and brain stem gliomas, a type of cancer that largely affects children and teenagers. About 300 to 400 young people a year are diagnosed with this type of tumor. Radiation and brain surgery risk damaging sensitive brain tissue, and chemotherapy often doesn't work. Most children with these tumors live less than 18 months. A nanobot delivering targeted chemotherapy could be a gamechanger. "These patients really don't have any other hope," Shpigelmacher says.
Of course, the main challenge of the developing such a device is guaranteeing that it's safe. Because tissue is so sensitive, any mistake could risk disastrous results. In recent years, Bionaut has tested its technology in dozens of healthy sheep and pigs with no major adverse effects. Sheep make a good stand-in for humans because their brains and spines are similar to ours.
The Bionaut device is about the size of a grain of rice.
Bionaut Labs
"As the Bionaut moves through brain tissue, it creates a transient track that heals within a few weeks," Shpigelmacher says. The company is hoping to be the first to test a nanobot in humans. In December 2022, it announced that a recent round of funding drew $43.2 million, for a total of 63.2 million, enabling more research and, if all goes smoothly, human clinical trials by early next year.
Once the technique has been perfected, further applications could include addressing other kinds of brain disorders that are considered incurable now, such as Alzheimer's or Huntington's disease. "Microrobots could serve as a bridgehead, opening the gateway to the brain and facilitating precise access of deep brain structure – either to deliver medication, take cell samples or stimulate specific brain regions," Shpigelmacher says.
Robot-assisted hybrid surgery with artificial intelligence is already used in state-of-the-art surgery centers, and many medical experts believe that nanorobotics will be the instrument of the future. In 2016, three scientists were awarded the Nobel Prize in Chemistry for their development of "the world's smallest machines," nano "elevators" and minuscule motors. Since then, the scientific experiments have progressed to the point where applicable devices are moving closer to actually being implemented.
Bionaut's technology was initially developed by a research team lead by Peer Fischer, head of the independent Micro Nano and Molecular Systems Lab at the Max Planck Institute for Intelligent Systems in Stuttgart, Germany. Fischer is considered a pioneer in the research of nano systems, which he began at Harvard University more than a decade ago. He and his team are advising Bionaut Labs and have licensed their technology to the company.
"The hope is that we can develop a vehicle to transport medication deep into the body," says Max Planck scientist Tian Qiu, who leads the cooperation with Bionaut Labs. He agrees with Shpigelmacher that the Bionaut's size is perfect for transporting medication loads and is researching potential applications for even smaller nanorobots, especially in the eye, where the tissue is extremely sensitive. "Nanorobots can sneak through very fine tissue without causing damage."
In "Fantastic Voyage," Raquel Welch's adventures inside the body of a dissident scientist let her swim through his veins into his brain, but her shrunken miniature submarine is attacked by antibodies; she has to flee through the nerves into the scientist's eye where she escapes into freedom on a tear drop. In reality, the exit in the lab is much more mundane. The Bionaut simply leaves the body through the same port where it entered. But apart from the dramatization, the "Fantastic Voyage" was almost prophetic, or, as Shpigelmacher says, "Science fiction becomes science reality."
This article was first published by Leaps.org on April 12, 2021.
How the Human Brain Project Built a Mind of its Own
In 2009, neuroscientist Henry Markram gave an ambitious TED talk. “Our mission is to build a detailed, realistic computer model of the human brain,” he said, naming three reasons for this unmatched feat of engineering. One was because understanding the human brain was essential to get along in society. Another was because experimenting on animal brains could only get scientists so far in understanding the human ones. Third, medicines for mental disorders weren’t good enough. “There are two billion people on the planet that are affected by mental disorders, and the drugs that are used today are largely empirical,” Markram said. “I think that we can come up with very concrete solutions on how to treat disorders.”
Markram's arguments were very persuasive. In 2013, the European Commission launched the Human Brain Project, or HBP, as part of its Future and Emerging Technologies program. Viewed as Europe’s chance to try to win the “brain race” between the U.S., China, Japan, and other countries, the project received about a billion euros in funding with the goal to simulate the entire human brain on a supercomputer, or in silico, by 2023.
Now, after 10 years of dedicated neuroscience research, the HBP is coming to an end. As its many critics warned, it did not manage to build an entire human brain in silico. Instead, it achieved a multifaceted array of different goals, some of them unexpected.
Scholars have found that the project did help advance neuroscience more than some detractors initially expected, specifically in the area of brain simulations and virtual models. Using an interdisciplinary approach of combining technology, such as AI and digital simulations, with neuroscience, the HBP worked to gain a deeper understanding of the human brain’s complicated structure and functions, which in some cases led to novel treatments for brain disorders. Lastly, through online platforms, the HBP spearheaded a previously unmatched level of global neuroscience collaborations.
Simulating a human brain stirs up controversy
Right from the start, the project was plagued with controversy and condemnation. One of its prominent critics was Yves Fregnac, a professor in cognitive science at the Polytechnic Institute of Paris and research director at the French National Centre for Scientific Research. Fregnac argued in numerous articles that the HBP was overfunded based on proposals with unrealistic goals. “This new way of over-selling scientific targets, deeply aligned with what modern society expects from mega-sciences in the broad sense (big investment, big return), has been observed on several occasions in different scientific sub-fields,” he wrote in one of his articles, “before invading the field of brain sciences and neuromarketing.”
"A human brain model can simulate an experiment a million times for many different conditions, but the actual human experiment can be performed only once or a few times," said Viktor Jirsa, a professor at Aix-Marseille University.
Responding to such critiques, the HBP worked to restructure the effort in its early days with new leadership, organization, and goals that were more flexible and attainable. “The HBP got a more versatile, pluralistic approach,” said Viktor Jirsa, a professor at Aix-Marseille University and one of the HBP lead scientists. He believes that these changes fixed at least some of HBP’s issues. “The project has been on a very productive and scientifically fruitful course since then.”
After restructuring, the HBP became a European hub on brain research, with hundreds of scientists joining its growing network. The HBP created projects focused on various brain topics, from consciousness to neurodegenerative diseases. HBP scientists worked on complex subjects, such as mapping out the brain, combining neuroscience and robotics, and experimenting with neuromorphic computing, a computational technique inspired by the human brain structure and function—to name just a few.
Simulations advance knowledge and treatment options
In 2013, it seemed that bringing neuroscience into a digital age would be farfetched, but research within the HBP has made this achievable. The virtual maps and simulations various HBP teams create through brain imaging data make it easier for neuroscientists to understand brain developments and functions. The teams publish these models on the HBP’s EBRAINS online platform—one of the first to offer access to such data to neuroscientists worldwide via an open-source online site. “This digital infrastructure is backed by high-performance computers, with large datasets and various computational tools,” said Lucy Xiaolu Wang, an assistant professor in the Resource Economics Department at the University of Massachusetts Amherst, who studies the economics of the HBP. That means it can be used in place of many different types of human experimentation.
Jirsa’s team is one of many within the project that works on virtual brain models and brain simulations. Compiling patient data, Jirsa and his team can create digital simulations of different brain activities—and repeat these experiments many times, which isn’t often possible in surgeries on real brains. “A human brain model can simulate an experiment a million times for many different conditions,” Jirsa explained, “but the actual human experiment can be performed only once or a few times.” Using simulations also saves scientists and doctors time and money when looking at ways to diagnose and treat patients with brain disorders.
Compiling patient data, scientists can create digital simulations of different brain activities—and repeat these experiments many times.
The Human Brain Project
Simulations can help scientists get a full picture that otherwise is unattainable. “Another benefit is data completion,” added Jirsa, “in which incomplete data can be complemented by the model. In clinical settings, we can often measure only certain brain areas, but when linked to the brain model, we can enlarge the range of accessible brain regions and make better diagnostic predictions.”
With time, Jirsa’s team was able to move into patient-specific simulations. “We advanced from generic brain models to the ability to use a specific patient’s brain data, from measurements like MRI and others, to create individualized predictive models and simulations,” Jirsa explained. He and his team are working on this personalization technique to treat patients with epilepsy. According to the World Health Organization, about 50 million people worldwide suffer from epilepsy, a disorder that causes recurring seizures. While some epilepsy causes are known others remain an enigma, and many are hard to treat. For some patients whose epilepsy doesn’t respond to medications, removing part of the brain where seizures occur may be the only option. Understanding where in the patients’ brains seizures arise can give scientists a better idea of how to treat them and whether to use surgery versus medications.
“We apply such personalized models…to precisely identify where in a patient’s brain seizures emerge,” Jirsa explained. “This guides individual surgery decisions for patients for which surgery is the only treatment option.” He credits the HBP for the opportunity to develop this novel approach. “The personalization of our epilepsy models was only made possible by the Human Brain Project, in which all the necessary tools have been developed. Without the HBP, the technology would not be in clinical trials today.”
Personalized simulations can significantly advance treatments, predict the outcome of specific medical procedures and optimize them before actually treating patients. Jirsa is watching this happen firsthand in his ongoing research. “Our technology for creating personalized brain models is now used in a large clinical trial for epilepsy, funded by the French state, where we collaborate with clinicians in hospitals,” he explained. “We have also founded a spinoff company called VB Tech (Virtual Brain Technologies) to commercialize our personalized brain model technology and make it available to all patients.”
The Human Brain Project created a level of interconnectedness within the neuroscience research community that never existed before—a network not unlike the brain’s own.
Other experts believe it’s too soon to tell whether brain simulations could change epilepsy treatments. “The life cycle of developing treatments applicable to patients often runs over a decade,” Wang stated. “It is still too early to draw a clear link between HBP’s various project areas with patient care.” However, she admits that some studies built on the HBP-collected knowledge are already showing promise. “Researchers have used neuroscientific atlases and computational tools to develop activity-specific stimulation programs that enabled paraplegic patients to move again in a small-size clinical trial,” Wang said. Another intriguing study looked at simulations of Alzheimer’s in the brain to understand how it evolves over time.
Some challenges remain hard to overcome even with computer simulations. “The major challenge has always been the parameter explosion, which means that many different model parameters can lead to the same result,” Jirsa explained. An example of this parameter explosion could be two different types of neurodegenerative conditions, such as Parkinson’s and Huntington’s diseases. Both afflict the same area of the brain, the basal ganglia, which can affect movement, but are caused by two different underlying mechanisms. “We face the same situation in the living brain, in which a large range of diverse mechanisms can produce the same behavior,” Jirsa said. The simulations still have to overcome the same challenge.
Understanding where in the patients’ brains seizures arise can give scientists a better idea of how to treat them and whether to use surgery versus medications.
The Human Brain Project
A network not unlike the brain’s own
Though the HBP will be closing this year, its legacy continues in various studies, spin-off companies, and its online platform, EBRAINS. “The HBP is one of the earliest brain initiatives in the world, and the 10-year long-term goal has united many researchers to collaborate on brain sciences with advanced computational tools,” Wang said. “Beyond the many research articles and projects collaborated on during the HBP, the online neuroscience research infrastructure EBRAINS will be left as a legacy even after the project ends.”
Those who worked within the HBP see the end of this project as the next step in neuroscience research. “Neuroscience has come closer to very meaningful applications through the systematic link with new digital technologies and collaborative work,” Jirsa stated. “In that way, the project really had a pioneering role.” It also created a level of interconnectedness within the neuroscience research community that never existed before—a network not unlike the brain’s own. “Interconnectedness is an important advance and prerequisite for progress,” Jirsa said. “The neuroscience community has in the past been rather fragmented and this has dramatically changed in recent years thanks to the Human Brain Project.”
According to its website, by 2023 HBP’s network counted over 500 scientists from over 123 institutions and 16 different countries, creating one of the largest multi-national research groups in the world. Even though the project hasn’t produced the in-silico brain as Markram envisioned it, the HBP created a communal mind with immense potential. “It has challenged us to think beyond the boundaries of our own laboratories,” Jirsa said, “and enabled us to go much further together than we could have ever conceived going by ourselves.”