AI and you: Is the promise of personalized nutrition apps worth the hype?
As a type 2 diabetic, Michael Snyder has long been interested in how blood sugar levels vary from one person to another in response to the same food, and whether a more personalized approach to nutrition could help tackle the rapidly cascading levels of diabetes and obesity in much of the western world.
Eight years ago, Snyder, who directs the Center for Genomics and Personalized Medicine at Stanford University, decided to put his theories to the test. In the 2000s continuous glucose monitoring, or CGM, had begun to revolutionize the lives of diabetics, both type 1 and type 2. Using spherical sensors which sit on the upper arm or abdomen – with tiny wires that pierce the skin – the technology allowed patients to gain real-time updates on their blood sugar levels, transmitted directly to their phone.
It gave Snyder an idea for his research at Stanford. Applying the same technology to a group of apparently healthy people, and looking for ‘spikes’ or sudden surges in blood sugar known as hyperglycemia, could provide a means of observing how their bodies reacted to an array of foods.
“We discovered that different foods spike people differently,” he says. “Some people spike to pasta, others to bread, others to bananas, and so on. It’s very personalized and our feeling was that building programs around these devices could be extremely powerful for better managing people’s glucose.”
Unbeknown to Snyder at the time, thousands of miles away, a group of Israeli scientists at the Weizmann Institute of Science were doing exactly the same experiments. In 2015, they published a landmark paper which used CGM to track the blood sugar levels of 800 people over several days, showing that the biological response to identical foods can vary wildly. Like Snyder, they theorized that giving people a greater understanding of their own glucose responses, so they spend more time in the normal range, may reduce the prevalence of type 2 diabetes.
The commercial potential of such apps is clear, but the underlying science continues to generate intriguing findings.
“At the moment 33 percent of the U.S. population is pre-diabetic, and 70 percent of those pre-diabetics will become diabetic,” says Snyder. “Those numbers are going up, so it’s pretty clear we need to do something about it.”
Fast forward to 2022,and both teams have converted their ideas into subscription-based dietary apps which use artificial intelligence to offer data-informed nutritional and lifestyle recommendations. Snyder’s spinoff, January AI, combines CGM information with heart rate, sleep, and activity data to advise on foods to avoid and the best times to exercise. DayTwo–a start-up which utilizes the findings of Weizmann Institute of Science–obtains microbiome information by sequencing stool samples, and combines this with blood glucose data to rate ‘good’ and ‘bad’ foods for a particular person.
“CGMs can be used to devise personalized diets,” says Eran Elinav, an immunology professor and microbiota researcher at the Weizmann Institute of Science in addition to serving as a scientific consultant for DayTwo. “However, this process can be cumbersome. Therefore, in our lab we created an algorithm, based on data acquired from a big cohort of people, which can accurately predict post-meal glucose responses on a personal basis.”
The commercial potential of such apps is clear. DayTwo, who market their product to corporate employers and health insurers rather than individual consumers, recently raised $37 million in funding. But the underlying science continues to generate intriguing findings.
Last year, Elinav and colleagues published a study on 225 individuals with pre-diabetes which found that they achieved better blood sugar control when they followed a personalized diet based on DayTwo’s recommendations, compared to a Mediterranean diet. The journal Cell just released a new paper from Snyder’s group which shows that different types of fibre benefit people in different ways.
“The idea is you hear different fibres are good for you,” says Snyder. “But if you look at fibres they’re all over the map—it’s like saying all animals are the same. The responses are very individual. For a lot of people [a type of fibre called] arabinoxylan clearly reduced cholesterol while the fibre inulin had no effect. But in some people, it was the complete opposite.”
Eight years ago, Stanford's Michael Snyder began studying how continuous glucose monitors could be used by patients to gain real-time updates on their blood sugar levels, transmitted directly to their phone.
The Snyder Lab, Stanford Medicine
Because of studies like these, interest in precision nutrition approaches has exploded in recent years. In January, the National Institutes of Health announced that they are spending $170 million on a five year, multi-center initiative which aims to develop algorithms based on a whole range of data sources from blood sugar to sleep, exercise, stress, microbiome and even genomic information which can help predict which diets are most suitable for a particular individual.
“There's so many different factors which influence what you put into your mouth but also what happens to different types of nutrients and how that ultimately affects your health, which means you can’t have a one-size-fits-all set of nutritional guidelines for everyone,” says Bruce Y. Lee, professor of health policy and management at the City University of New York Graduate School of Public Health.
With the falling costs of genomic sequencing, other precision nutrition clinical trials are choosing to look at whether our genomes alone can yield key information about what our diets should look like, an emerging field of research known as nutrigenomics.
The ASPIRE-DNA clinical trial at Imperial College London is aiming to see whether particular genetic variants can be used to classify individuals into two groups, those who are more glucose sensitive to fat and those who are more sensitive to carbohydrates. By following a tailored diet based on these sensitivities, the trial aims to see whether it can prevent people with pre-diabetes from developing the disease.
But while much hope is riding on these trials, even precision nutrition advocates caution that the field remains in the very earliest of stages. Lars-Oliver Klotz, professor of nutrigenomics at Friedrich-Schiller-University in Jena, Germany, says that while the overall goal is to identify means of avoiding nutrition-related diseases, genomic data alone is unlikely to be sufficient to prevent obesity and type 2 diabetes.
“Genome data is rather simple to acquire these days as sequencing techniques have dramatically advanced in recent years,” he says. “However, the predictive value of just genome sequencing is too low in the case of obesity and prediabetes.”
Others say that while genomic data can yield useful information in terms of how different people metabolize different types of fat and specific nutrients such as B vitamins, there is a need for more research before it can be utilized in an algorithm for making dietary recommendations.
“I think it’s a little early,” says Eileen Gibney, a professor at University College Dublin. “We’ve identified a limited number of gene-nutrient interactions so far, but we need more randomized control trials of people with different genetic profiles on the same diet, to see whether they respond differently, and if that can be explained by their genetic differences.”
Some start-ups have already come unstuck for promising too much, or pushing recommendations which are not based on scientifically rigorous trials. The world of precision nutrition apps was dubbed a ‘Wild West’ by some commentators after the founders of uBiome – a start-up which offered nutritional recommendations based on information obtained from sequencing stool samples –were charged with fraud last year. The weight-loss app Noom, which was valued at $3.7 billion in May 2021, has been criticized on Twitter by a number of users who claimed that its recommendations have led to them developed eating disorders.
With precision nutrition apps marketing their technology at healthy individuals, question marks have also been raised about the value which can be gained through non-diabetics monitoring their blood sugar through CGM. While some small studies have found that wearing a CGM can make overweight or obese individuals more motivated to exercise, there is still a lack of conclusive evidence showing that this translates to improved health.
However, independent researchers remain intrigued by the technology, and say that the wealth of data generated through such apps could be used to help further stratify the different types of people who become at risk of developing type 2 diabetes.
“CGM not only enables a longer sampling time for capturing glucose levels, but will also capture lifestyle factors,” says Robert Wagner, a diabetes researcher at University Hospital Düsseldorf. “It is probable that it can be used to identify many clusters of prediabetic metabolism and predict the risk of diabetes and its complications, but maybe also specific cardiometabolic risk constellations. However, we still don’t know which forms of diabetes can be prevented by such approaches and how feasible and long-lasting such self-feedback dietary modifications are.”
Snyder himself has now been wearing a CGM for eight years, and he credits the insights it provides with helping him to manage his own diabetes. “My CGM still gives me novel insights into what foods and behaviors affect my glucose levels,” he says.
He is now looking to run clinical trials with his group at Stanford to see whether following a precision nutrition approach based on CGM and microbiome data, combined with other health information, can be used to reverse signs of pre-diabetes. If it proves successful, January AI may look to incorporate microbiome data in future.
“Ultimately, what I want to do is be able take people’s poop samples, maybe a blood draw, and say, ‘Alright, based on these parameters, this is what I think is going to spike you,’ and then have a CGM to test that out,” he says. “Getting very predictive about this, so right from the get go, you can have people better manage their health and then use the glucose monitor to help follow that.”
Why we need to get serious about ending aging
It is widely acknowledged that even a small advance in anti-aging science could yield benefits in terms of healthy years that the traditional paradigm of targeting specific diseases is not likely to produce. A more youthful population would also be less vulnerable to epidemics. Approximately 93 percent of all COVID-19 deaths reported in the U.S. occurred among those aged 50 or older. The potential economic benefits would be tremendous. A more youthful population would consume less medical resources and be able to work longer. A recent study published in Nature estimates that a slowdown in aging that increases life expectancy by one year would save $38 trillion per year for the U.S. alone.
A societal effort to understand, slow down, arrest or even reverse aging of at least the size of our response to COVID-19 would therefore be a rational commitment. In fact, given that America’s older population is projected to grow dramatically, and the cost of healthcare with it, it is not an overstatement to say that the future welfare of the country may depend on solving aging.
This year, the kingdom of Saudi Arabia has announced that it will spend up to 1 billion dollars per year on science with the potential to slow down the aging process. We have also seen important investments from billionaires like Google co-founder Larry Page, Amazon founder Jeff Bezos, business magnate Larry Ellison, and PayPal co-founder Peter Thiel.
The U.S. government, however, is lagging: The National Institutes of Health spent less than one percent of its $43 billion budget for the fiscal year of 2021 on the National Institute on Aging’s Division of Aging Biology. When you visit the division’s webpage you find that their mission statement carefully omits any mention of the possibility of slowing down the aging process.
There is a lack of political will and leadership on the issue, and the idea that we should seek to do something about aging is generally met with a great deal of suspicion and trepidation. In a large representative study conducted by the Pew Research Center in 2013, only 38% of the respondents said that they would want a treatment that could slow the aging process and allow them to live at least 120 years. Apparently, most people prefer, or at least do not mind, to age and die within a natural lifespan. This result has been confirmed by smaller studies and it is, I think, surprising. Are we not supposed to live in a youth-culture? Are people not supposed to want to stay young and alive forever? Is self-preservation not the strong drive we have always assumed it to be?
We are inundated and saturated with an ideology of death-acceptance.
In my book, The Case against Death, I suggest that we have been culturally conditioned to think that it is virtuous to accept aging and death. We are taught to believe that although aging and death seem gruesome, they are what is best for us, all things considered. This is what we are supposed to think, and the majority accept it. I call this the Wise View because death acceptance has been the dominant view of philosophers since the beginning. Socrates compared our earthly life to an illness and a prison and described death as a healer and a liberator. The Buddha taught that life is suffering and that the way to escape suffering is to end the cycle of birth, death and rebirth. Stoic philosophers from Zeno to Marcus Aurelius believed that everything that happens in accordance with nature is good, and that therefore we should not only accept death but welcome it as an aspect of a perfect totality.
Epicureans agreed with these rival schools and famously argued that death cannot harm us because where we are, death is not, and where death is we are not. We cannot be harmed if we are not, so death is harmless. The simple view that death actually can harm us greatly is one of the least philosophical views one can hold.
In The Case Against Death, philosopher Ingemar Patrick Linden argues that we frown on using science to prolong healthy life only because we're culturally conditioned to think that way.
Many of the stories we tell promote the Wise View. One of the earliest known pieces of literature, the Epic of Gilgamesh, follows Gilgamesh on a quest for eternal life ending with the wisdom that death is the destiny of man. Today we learn about the tedium of immortality from the children’s book Tuck Everlasting by Natalie Babbitt, and we are warned about the vice of wanting to resist death in other books and films such as J.K Rowling’s Harry Potter, where Voldemort must kill Harry as a step towards his own immortality; C.S. Lewis’ The Chronicles of Narnia where the White Witch has gained immortal youth and madness in equal measures; J.R.R. Tolkien’s Lord of the Rings trilogy where the ring extends the wearer’s life but can also destroy them, as exemplified by the creep Gollum; and Doctor Strange where life extension is the one magical power that is taboo. In Star Wars, Yoda, a stereotype of the sage, teaches us the wisdom handed down by philosophers and prophets: “Death is a natural part of life. Rejoice for those around you who transform into the Force. Mourn them do not. Miss them do not.”
We are inundated and saturated with an ideology of death-acceptance. Can the dear reader name one single story where the hero is pursuing anti-aging, longevity or immortality and the villain tries to stop her?
The Wise View resonates with us partly because we think that there is nothing we can do about aging and death, so we do not want to wish for what we cannot have. Youth and immortality are sour grapes to us. Believing that death is, all things considered, not such a bad thing, protects us from experiencing our aging and approaching death as a gruesome tragedy. This need to escape the thought that we are heading towards a personal catastrophe explains why many are so quick to accept arguments against radical life extension, despite their often glaring weaknesses.
One of the most common objections to radical life extension is that aging and death are natural. The problem with this argument is that many things that are natural are very bad, such as cancer, and other things that are not natural are very good, such as a cure for cancer. Why are we so sure that cancer is bad? Because we assume that it is bad to die. Indeed, nothing is more natural than wanting to live. We seem to need philosophers and story tellers to talk us out of it and, in the words of a distinguished bioethicist, “instruct and somewhat moderate our lust for life.”
Another standard objection is that we need a deadline, and that without death we could postpone every action forever. “Death brings urgency and seriousness to life,” say proponents of this view, but there are several problems with this argument. Even if our lives were endless, there would still be many things we would have to do at a certain time, and that could not be redone, for example, saving our planet from being destroyed, or becoming the first person on Venus. And if we prefer pleasant endless lives over unpleasant endless ones, we will have to exercise, eat right, keep our word, develop our talents, show up for time at work, pay our taxes by the due date, remember birthdays, and so on.
The Wise View provides us with a feel-good bromide for the anxiety created by the foreknowledge of our decay and death by telling us that these are not evils, but blessings in disguise. Once perhaps an innocuous delusion, today the view stands in the way of a necessary societal commitment to research that can prolong our healthy life.
Besides, even if we succeeded in ending aging, we would still die from other causes. Given the rate of accidental deaths we would be fortunate to live to age 2000 all things equal. So even if, contrary to what I have argued, we do need a deadline, we can still argue that the natural lifespan that we now labor under is inhuman and that it forces each human to limit her ambitions and to become only a fragment of all that she that could have been. Our tight time constraint imposes tragic choices and inflated opportunity costs. Death does not make life matter; it makes time matter.
The perhaps most awful argument against radical life extension is grounded in a pessimism that holds life in such little regard that it says that best of all is never to have born. This view was expressed by Ecclesiastes in the Hebrew Bible, by Sophocles and several other ancient Greeks, by the German philosopher Arthur Schopenhauer, and recently by, among others, the South African philosopher David Benatar who argues that it is wrong to bring children into the world and that we should euthanize all sentient life. Pessimism, one suspects, largely appeals to some for reasons having to do with personal temperament, but insofar as it is built on factual beliefs, they can be addressed by providing a less negatively biased understanding of the world, by pointing out that curing aging would decrease the badness that they are so hypersensitive to, and by reminding them that if life really becomes unbearable, they are free to quit at any time. Other means of persuasion could include recommending sleep, exercise and taking long brisk walks in nature.
The Wise View provides us with a feel-good bromide for the anxiety created by the foreknowledge of our decay and death by telling us that these are not evils, but blessings in disguise. Once perhaps an innocuous delusion, today the view stands in the way of a necessary societal commitment to research that can prolong our healthy life. We need abandon it and openly admit that aging is a scourge that deserves to be fought with the combined energies equaling those expended on fighting COVID-19, Alzheimer’s disease, cancer, stroke and all the other illnesses for which aging is the greatest risk factor. The fight to end aging transcends ordinary political boundaries and is therefore the kind of grand unifying enterprise that could re-energize a society suffering from divisiveness and the sense of a lack of a common purpose. It is hard to imagine a more worthwhile cause.
Could a tiny fern change the world — again?
More than 50 million years ago, the Arctic Ocean was the opposite of a frigid wasteland. It was a gigantic lake surrounded by lush greenery brimming with flora and fauna, thanks to the humidity and warm temperatures. Giant tortoises, alligators, rhinoceros-like animals, primates, and tapirs roamed through nearby forests in the Arctic.
This greenhouse utopia abruptly changed in the early Eocene period, when the Arctic Ocean became landlocked. A channel that connected the Arctic to the greater oceans got blocked. This provided a tiny fern called Azolla the perfect opportunity to colonize the layer of freshwater that formed on the surface of the Arctic Ocean. The floating plants rapidly covered the water body in thick layers that resembled green blankets.
Gradually, Azolla colonies migrated to every continent with the help of repeated flooding events. For around a million years, they captured more than 80 percent of atmospheric carbon dioxide that got buried at the bottom of the Arctic Ocean as billions of Azolla plants perished.
This “Arctic Azolla event” had devastating impacts on marine life. To date, scientists are trying to figure out how it ended. But they documented that the extraordinary event cooled down the Arctic by at least 40 degrees Fahrenheit — effectively freezing the poles and triggering several cycles of ice ages. “This carbon dioxide sequestration changed the climate from greenhouse to white house,” says Jonathan Bujak, a paleontologist who has researched the Arctic through expeditions since 1973.
Some farmers and scientists, such as Bujak, are looking to this ancient fern, which manipulated the Earth’s climate around 49 million years ago with its insatiable appetite for carbon dioxide, as a potential solution to our modern-day agricultural and environmental challenges. “There is no other plant like Azolla in the world,” says Bujak.
Decoding the Azolla plant
Azolla lives in symbiosis with a cyanobacterium called Anabaena that made the plant’s leaf cavities its permanent home at an early stage in Earth's history. This close relationship with Anabaena enables Azolla to accomplish a feat that is impossible for most plants: directly splitting dinitrogen molecules that make up 78 percent of the Earth’s atmosphere.
A dinitrogen molecule consists of two nitrogen atoms tightly locked together in one of the strongest bonds in nature. The semi-aquatic fern’s ability to split nitrogen, called nitrogen-fixing, made it a highly revered plant in East Asia. Rice farmers used Azolla as a biofertilizer since the 11th century in Vietnam and China.
For decades, scientists have attempted to decode Azolla’s evolution. Cell biologist Francisco Carrapico, who worked at the University of Lisbon, has analyzed this distinctive symbiosis since the 1980s. To his amazement, in 1991, he found that bacteria are the third partner of the Azolla-Anabaena symbiosis.
“Azolla and Anabaena cannot survive without each other. They have co-evolved for 80 million years, continuously exchanging their genetic material with each other,” says Bujak, co-author of The Azolla Story, which he published with his daughter, Alexandra Bujak, an environmental scientist. Three different levels of nitrogen fixation take place within the plant, as Anabaena draws down as much as 2,200 pounds of atmospheric nitrogen per acre annually.
“Using Azolla to mitigate climate change might sound a bit too simple. But that is not the case,” Bujak says. “At a microscopic level, extremely complicated biochemical reactions are constantly occurring inside the plant’s cells that machines or technology cannot replicate yet.”
In 2018, researchers based in the U.S. managed to sequence Azolla’s complete genome — which is four times larger than the human genome — through a crowdfunded study, further increasing our understanding of this plant. “Azolla is a superorganism that works efficiently as a natural biotechnology system that makes it capable of doubling in size within three to five days,” says Carrapico.
Making Azolla mainstream again in agriculture
While scientific groups in the Global North have been working towards unraveling the tiny fern’s inner workings, communities in the Global South are busy devising creative ways to return to their traditional agricultural roots by tapping into Azolla’s full potential.
Pham Gia Minh, an entrepreneur living in Hanoi, Vietnam, is one such citizen scientist who believes that Azolla could be a climate savior. More than two decades after working in finance and business development, Minh is now focusing on continuing his grandfather’s legacy, an agricultural scientist who conducted Azolla research until the 1950s. “Azolla is our family’s heritage,” says Minh.
Pham Gia Minh, an entrepreneur and citizen scientist in Hanoi, Vietnam, believes that Azolla could be a climate savior
Pham Gia Minh
Since the advent of chemical fertilizers in the early 1900s, farmers in Asia abandoned Azolla to save on time and labor costs. But rice farmers in the country went back to cultivating Azolla during the Vietnam War after chemical trade embargoes made chemical fertilizers far too expensive and inaccessible.
By 1973, Azolla cultivation in rice paddy fields was established on half a million hectares in Vietnam. By injecting nitrogen into the soil, Azolla improves soil fertility and also increases rice yields by at least 27 percent compared to urea. The plants can also reduce a farm’s methane emissions by 40 percent.
“Unfortunately, after 1985, chemical fertilizers became cheap and widely available in Vietnam again. So, farmers stopped growing Azolla because of the time-consuming and labor-intensive cultivation process,” says Minh.
Minh has invested in a rural farm where he is proving that modern technology can make the process less burdensome. He uses a pump and drying equipment for harvesting Azolla in a small pond, and he deploys a drone for spraying insecticides and fertilizers on the pond at regular intervals.
As Azolla lacks phosphorus, farmers in developing countries still find it challenging to let go of chemical fertilizers completely. Still, Minh and Bujak say that farmers can use Azolla instead of chemical fertilizers after mixing it with dung.
In the last few years, the fern’s popularity has been growing in other developing countries like India, Palestine, Indonesia, the Philippines, and Bangladesh, where local governments and citizens are trying to re-introduce Azolla integrated farming by growing the ferns in small ponds.
Replacing soybeans with Azolla
In Ecuador, Mariano Montano Armijos, a former chemical engineer, has worked with Azolla for more than 20 years. Since 2008, he has shared resources and information for growing Azolla with 3,000 farmers in Ecuador. The farmers use the harvested plants as a bio-fertilizer and feed for livestock.
“The farmers do not use urea anymore,” says Armijos. “This goes against the conventional agricultural practices of using huge amounts of synthetic nitrogen on a hectare of rice or corn fields.”
He insists that Azolla’s greatest strength is that it is a rich source of proteins, making it highly nutritious for human beings as well. After growing Azolla on a small scale in ponds, Armijos and his business partner, Ivan Noboa, are now building a facility for cultivating the ferns as a superfood on an industrial scale.
According to Armijos, one hectare of Azolla in Ecuador can produce seven tons of proteins. Whereas soybeans produce only one ton of protein per hectare. “If we switch to Azolla, it could help in reducing deforestation in the Amazon. But taming Azolla and turning it into a crop is not easy,” he adds.
Henriette Schluepmann, a molecular plant biologist at Utrecht University in the Netherlands, believes that Azolla could replace soybeans and chemical fertilizers someday — only if researchers can achieve yield stability in controlled environments over long durations.
“In a country like the Netherlands that is surrounded by water with high levels of phosphates, it makes sense to grow Azolla as a substitute for soybeans,” says Schluepmann. “For that to happen, we need massive investments to understand these ferns’ reproductive system and how to replicate that within aquaculture systems on a large scale.”
Pollution control and carbon sequestration
Currently, Schluepmann and her team are growing Azolla in a plant nursery or closed system before transferring the ferns to flooded fields. So far, they have been able to continuously grow Azolla without any major setbacks for a total of 155 days. Taking care of these plants’ well-being is an uphill struggle.
Unlike most plants, Azolla does not grow from seeds because it contains female and male spores that tend to split instead of reproducing. To add to that, growing Azolla on a large scale in controlled environments makes the floating plants extremely vulnerable to insect infestations and fungi attacks.
“Even though it is easier to grow Azolla on a non-industrial scale, the long and tedious cultivation process is often in conflict with human rights,” she says. Farms in developing countries such as Indonesia sometimes use child labor for cultivating Azolla.”
History has taught us that the uncontrolled growth of Azolla plants deprives marine ecosystems of sunlight and chokes life underneath them. But researchers like Schluepmann and Bujak are optimistic that even on a much smaller scale, Azolla can put up a fight against human-driven climate change.
Schluepmann discovered an insecticide that can control Azolla blooms. But in the wild, this aquatic fern grows relentlessly in polluted rivers and lakes and has gained a notorious reputation as an invasive weed. Countries like Portugal and the UK banned Azolla after experiencing severe blooms in rivers that snuffed out local marine life.
“Azolla has been misunderstood as a nuisance. But in reality, it is highly beneficial for purifying water,” says Bujak. Through a process called phytoremediation, Azolla locks up pollutants like excess nitrogen and phosphorus and stops toxic algal blooms from occurring in rivers and lakes.
A 2018 study found that Azolla can decrease nitrogen and phosphorus levels in wastewater by 33 percent and 40.5 percent, respectively. While harmful algae like phytoplankton produce toxins and release noxious gases, Azolla automatically blocks any toxins that its cyanobacteria, Anabaena, might produce.
“In our labs, we observed that Azolla works effectively in treating wastewater,” explains Schluepmann. “Once we gain a better understanding of Azolla aquaculture, we can also use it for carbon capture and storage. But in Europe, we would have to use the entire Baltic Sea to make a difference.”
Planting massive amounts of these prehistoric ferns in any of the Northern great water bodies is out of the question. After all, history has taught us that the uncontrolled growth of Azolla plants deprives marine ecosystems of sunlight and chokes life underneath them. But researchers like Schluepmann and Bujak are optimistic that even on a much smaller scale, Azolla can put up a fight against human-driven climate change.
Traditional carbon capture and storage methods are not only expensive but also inefficient and could increase air pollution. According to Bujak’s estimates, Azolla can sequester 10 metric tonnes of carbon dioxide per hectare annually, which is 10 times the average capacity of grasslands.
“Anyone can set up their own DIY carbon capture and storage system by growing Azolla in shallow water. After harvesting and compressing the plants, carbon dioxide gets stored permanently,” says Bujak.
He envisions scaling up this process by setting up “Azolla hubs” in mega-cities where the plants are grown in shallow trays stacked on top of each other with vertical farming systems built within multi-story buildings. The compressed Azolla plants can then be converted into a biofuel, fertilizer, livestock feed, or biochar for sequestering carbon dioxide.
“Using Azolla to mitigate climate change might sound a bit too simple. But that is not the case,” Bujak adds. “At a microscopic level, extremely complicated biochemical reactions are constantly occurring inside the plant’s cells that machines or technology cannot replicate yet.”
Through Azolla, scientists hope to work with nature by tapping into four billion years of evolution.