Smartwatches can track COVID-19 symptoms, study finds
If a COVID-19 infection develops, a wearable device may eventually be able to clue you in. A study at the University of Michigan found that a smartwatch can monitor how symptoms progress.
The study evaluated the effects of COVID-19 with various factors derived from heart-rate data. This method also could be employed to detect other diseases, such as influenza and the common cold, at home or when medical resources are limited, such as during a pandemic or in developing countries.
Tracking students and medical interns across the country, the University of Michigan researchers found that new signals embedded in heart rate indicated when individuals were infected with COVID-19 and how ill they became.
For instance, they discovered that individuals with COVID-19 experienced an increase in heart rate per step after the onset of their symptoms. Meanwhile, people who reported a cough as one of their COVID-19 symptoms had a much more elevated heart rate per step than those without a cough.
“We previously developed a variety of algorithms to analyze data from wearable devices. So, when the COVID-19 pandemic hit, it was only natural to apply some of these algorithms to see if we can get a better understanding of disease progression,” says Caleb Mayer, a doctoral student in mathematics at the University of Michigan and a co-first author of the study.
People may not internally sense COVID-19’s direct impact on the heart, but “heart rate is a vital sign that gives a picture of overall health," says Daniel Forger, a University of Michigan professor.
Millions of people are tracking their heart rate through wearable devices. This information is already generating a tremendous amount of data for researchers to analyze, says co-author Daniel Forger, professor of mathematics and research professor of computational medicine and bioinformatics at the University of Michigan.
“Heart rate is affected by many different physiological signals,” Forger explains. “For instance, if your lungs aren’t functioning properly, your heart may need to beat faster to meet metabolic demands. Your heart rate has a natural daily rhythm governed by internal biological clocks.” While people may not internally sense COVID-19’s direct impact on the heart, he adds that “heart rate is a vital sign that gives a picture of overall health.”
Among the total of 2,164 participants who enrolled in the student study, 72 undergraduate and graduate students contracted COVID-19, providing wearable data from 50 days before symptom onset to 14 days after. The researchers also analyzed this type of data for 43 medical interns from the Intern Health Study by the Michigan Neuroscience Institute and 29 individuals (who are not affiliated with the university) from the publicly available dataset.
Participants could wear the device on either wrist. They also documented their COVID-19 symptoms, such as fever, shortness of breath, cough, runny nose, vomiting, diarrhea, body aches, loss of taste, loss of smell, and sore throat.
Experts not involved in the study found the research to be productive. “This work is pioneering and reveals exciting new insights into the many important ways that we can derive clinically significant information about disease progression from consumer-grade wearable devices,” says Lisa A. Marsch, director of the Center for Technology and Behavioral Health and a professor in the Geisel School of Medicine at Dartmouth College. “Heart-rate data are among the highest-quality data that can be obtained via wearables.”
Beyond the heart, she adds, “Wearable devices are providing novel insights into individuals’ physiology and behavior in many health domains.” In particular, “this study beautifully illustrates how digital-health methodologies can markedly enhance our understanding of differences in individuals’ experience with disease and health.”
Previous studies had demonstrated that COVID-19 affects cardiovascular functions. Capitalizing on this knowledge, the University of Michigan endeavor took “a giant step forward,” says Gisele Oda, a researcher at the Institute of Biosciences at the University of Sao Paulo in Brazil and an expert in chronobiology—the science of biological rhythms. She commends the researchers for developing a complex algorithm that “could extract useful information beyond the established knowledge that heart rate increases and becomes more irregular in COVID patients.”
Wearable devices open the possibility of obtaining large-scale, long, continuous, and real-time heart-rate data on people performing everyday activities or while sleeping. “Importantly, the conceptual basis of this algorithm put circadian rhythms at the center stage,” Oda says, referring to the physical, mental, and behavioral changes that follow a 24-hour cycle. “What we knew before was often based on short-time heart rate measured at any time of day,” she adds, while noting that heart rate varies between day and night and also changes with activity.
However, without comparison to a control group of people having the common flu, it is difficult to determine if the heart-rate signals are unique to COVID-19 or also occur with other illnesses, says John Torous, an assistant professor of psychiatry at Harvard Medical School who has researched wearable devices. In addition, he points to recent data showing that many wearables, which work by beaming light through the skin, may be less accurate in people with darker skin due to variations in light absorption.
While the results sound interesting, they lack the level of conclusive evidence that would be needed to transform how physicians care for patients. “But it is a good step in learning more about what these wearables can tell us,” says Torous, who is also director of digital psychiatry at Beth Israel Deaconess Medical Center, a Harvard affiliate, in Boston. A follow-up step would entail replicating the results in a different pool of people to “help us realize the full value of this work.”
It is important to note that this research was conducted in university settings during the early phases of the pandemic, with remote learning in full swing amid strict isolation and quarantine mandates in effect. The findings demonstrate that physiological monitoring can be performed using consumer-grade wearable sensors, allowing research to continue without in-person contact, says Sung Won Choi, a professor of pediatrics at the University of Michigan who is principal investigator of the student study.
“The worldwide COVID-19 pandemic interrupted a lot of activities that relied on face-to-face interactions, including clinical research,” Choi says. “Mobile technology proved to be tremendously beneficial during that time, because it allowed us to collect detailed physiological data from research participants remotely over an entire semester.” In fact, the researchers did not have any in-person contact with the students involved in the study. “Everything was done virtually," Choi explains. "Importantly, their willingness to participate in research and share data during this historical time, combined with the capacity of secure cloud storage and novel mathematical analytics, enabled our research teams to identify unique patterns in heart-rate data associated with COVID-19.”
Today’s Focus on STEM Education Is Missing A Crucial Point
I once saw a fascinating TED talk on 3D printing. As I watched the presenter discuss the custom fabrication, not of plastic gears or figurines, but of living, implantable kidneys, I thought I was finally living in the world of Star Trek, and I experienced a flush of that eager, expectant enthusiasm I felt as a child looking toward the future. I looked at my current career and felt a rejuvenation of my commitment to teach young people the power of science.
The well-rounded education of human beings needs to include lessons learned both from a study of the physical world, and from a study of humanity.
Whether we are teachers or not, those of us who admire technology and innovation, and who wish to support progress, usually embrace the importance of educating the next generation of scientists and inventors. Growing a healthy technological civilization takes a lot of work, skill, and wisdom, and its continued health depends on future generations of competent thinkers. Thus, we may find it encouraging that there is currently an abundance of interest in STEM– the common acronym for the study of science, technology, engineering, and math.
But education is as challenging an endeavor as science itself. Educating youth--if we want to do it right--requires as much thought, work, and expertise as discovering a cure or pioneering regenerative medicine. Before we give our money, time, or support to any particular school or policy, let's give some thought to the details of the educational process.
A Well-Balanced Diet
For one thing, STEM education cannot stand in isolation. The well-rounded education of human beings needs to include lessons learned both from a study of the physical world, and from a study of humanity. This is especially true for the basic education of children, but it is true even for college students. And even for those in science and engineering, there are important lessons to be learned from the study of history, literature, and art.
Scientists have their own emotions and values, and also need financial support. The fruits of their labor ultimately benefit other people. How are we all to function together in our division-of-labor society, without some knowledge of the way societies work? How are we to fully thrive and enjoy life, without some understanding of ourselves, our motives, our moral values, and our relationships to others? STEM education needs the humanities as a partner. That flourishing civilization we dream of requires both technical competence and informed life-choices.
Think for Yourself (Even in Science)
Perhaps even more important than what is taught, is the subject of how things are taught. We want our children to learn the skill of thinking independently, but even in the sciences, we often fail completely to demonstrate how. Instead of teaching science as a thinking process, we indoctrinate, using the grand discoveries of the great scientists as our sacred texts. But consider the words of Isaac Newton himself, regarding rote learning:
A Vulgar Mechanick can practice what he has been taught or seen done, but if he is in an error he knows not how to find it out and correct it, and if you put him out of his road he is at a stand. Whereas he that is able to reason nimbly and judiciously about figure, force, and motion, is never at rest till he gets over every rub.
What's the point of all this formal schooling in the first place? Is it, as many of the proponents of STEM education might argue, to train students for a "good" career?
If our goal is to help students "reason nimbly" about the world around them, as the great scientists themselves did, are we succeeding? When we "teach" middle school students about DNA or cellular respiration by presenting as our only supporting evidence cartoon pictures, are we showing students a process of discovery based on evidence and hard work? Or are we just training them to memorize and repeat what the authorities say?
A useful education needs to give students the skill of following a line of reasoning, of asking rational questions, and of chewing things through in their minds--even if we regard the material as beyond question. Besides feeding students a well-balanced diet of knowledge, healthy schooling needs to teach them to digest this information thoroughly.
Thinking Training
Now step back for a moment and think about the purpose of education. What's the point of all this formal schooling in the first place? Is it, as many of the proponents of STEM education might argue, to train students for a "good" career? That view may have some validity for young adults, who are beginning to choose electives in favored subjects, and have started to choose a direction for their career.
But for the basic education of children, this way of thinking is presumptuous and disastrous. I would argue that the central purpose of a basic education is not to teach children how to perform this or that particular skill, but simply to teach them to think clearly. We should not be aiming to provide job training, but thinking training. We should be helping children learn how to "reason nimbly" about the world around them, and breathing life into their thinking processes, by which they will grapple with the events and circumstances of their lives.
So as we admire innovation, dream of a wonderful future, and attempt to nurture the next generation of scientists and engineers, instead of obsessing over STEM education, let us focus on rational education. Let's worry about showing children how to think--about all the important things in life. Let's give them the basic facts of human existence -- physical and humanitarian -- and show them how to fluently and logically understand them.
Some students will become the next generation of creators, and some will follow other careers, but together -- if they are educated properly -- they will continue to grow their inheritance, and to keep our civilization healthy and flourishing, in body and in mind.
Do New Tools Need New Ethics?
Scarcely a week goes by without the announcement of another breakthrough owing to advancing biotechnology. Recent examples include the use of gene editing tools to successfully alter human embryos or clone monkeys; new immunotherapy-based treatments offering longer lives or even potential cures for previously deadly cancers; and the creation of genetically altered mosquitos using "gene drives" to quickly introduce changes into the population in an ecosystem and alter the capacity to carry disease.
The environment for conducting science is dramatically different today than it was in the 1970s, 80s, or even the early 2000s.
Each of these examples puts pressure on current policy guidelines and approaches, some existing since the late 1970s, which were created to help guide the introduction of controversial new life sciences technologies. But do the policies that made sense decades ago continue to make sense today, or do the tools created during different eras in science demand new ethics guidelines and policies?
Advances in biotechnology aren't new of course, and in fact have been the hallmark of science since the creation of the modern U.S. National Institutes of Health in the 1940s and similar government agencies elsewhere. Funding agencies focused on health sciences research with the hope of creating breakthroughs in human health, and along the way, basic science discoveries led to the creation of new scientific tools that offered the ability to approach life, death, and disease in fundamentally new ways.
For example, take the discovery in the 1970s of the "chemical scissors" in living cells called restriction enzymes, which could be controlled and used to introduce cuts at predictable locations in a strand of DNA. This led to the creation of tools that for the first time allowed for genetic modification of any organism with DNA, which meant bacteria, plants, animals, and even humans could in theory have harmful mutations repaired, but also that changes could be made to alter or even add genetic traits, with potentially ominous implications.
The scientists involved in that early research convened a small conference to discuss not only the science, but how to responsibly control its potential uses and their implications. The meeting became known as the Asilomar Conference for the meeting center where it was held, and is often noted as the prime example of the scientific community policing itself. While the Asilomar recommendations were not sufficient from a policy standpoint, they offered a blueprint on which policies could be based and presented a model of the scientific community setting responsible controls for itself.
But the environment for conducting science changed over the succeeding decades and it is dramatically different today than it was in the 1970s, 80s, or even the early 2000s. The regime for oversight and regulation that has provided controls for the introduction of so-called "gene therapy" in humans starting in the mid-1970s is beginning to show signs of fraying. The vast majority of such research was performed in the U.S., U.K., and Europe, where policies were largely harmonized. But as the tools for manipulating humans at the molecular level advanced, they also became more reliable and more precise, as well as cheaper and easier to use—think CRISPR—and therefore more accessible to more people in many more countries, many without clear oversight or policies laying out responsible controls.
There is no precedent for global-scale science policy, though that is exactly what this moment seems to demand.
As if to make the point through news headlines, scientists in China announced in 2017 that they had attempted to perform gene editing on in vitro human embryos to repair an inherited mutation for beta thalassemia--research that would not be permitted in the U.S. and most European countries and at the time was also banned in the U.K. Similarly, specialists from a reproductive medicine clinic in the U.S. announced in 2016 that they had performed a highly controversial reproductive technology by which DNA from two women is combined (so-called "three parent babies"), in a satellite clinic they had opened in Mexico to avoid existing prohibitions on the technique passed by the U.S. Congress in 2015.
In both cases, genetic changes were introduced into human embryos that if successful would lead to the birth of a child with genetically modified germline cells—the sperm in boys or eggs in girls—with those genetic changes passed on to all future generations of related offspring. Those are just two very recent examples, and it doesn't require much imagination to predict the list of controversial possible applications of advancing biotechnologies: attempts at genetic augmentation or even cloning in humans, and alterations of the natural environment with genetically engineered mosquitoes or other insects in areas with endemic disease. In fact, as soon as this month, scientists in Africa may release genetically modified mosquitoes for the first time.
The technical barriers are falling at a dramatic pace, but policy hasn't kept up, both in terms of what controls make sense and how to address what is an increasingly global challenge. There is no precedent for global-scale science policy, though that is exactly what this moment seems to demand. Mechanisms for policy at global scale are limited–-think UN declarations, signatory countries, and sometimes international treaties, but all are slow, cumbersome and have limited track records of success.
But not all the news is bad. There are ongoing efforts at international discussion, such as an international summit on human genome editing convened in 2015 by the National Academies of Sciences and Medicine (U.S.), Royal Academy (U.K.), and Chinese Academy of Sciences (China), a follow-on international consensus committee whose report was issued in 2017, and an upcoming 2nd international summit in Hong Kong in November this year.
These efforts need to continue to focus less on common regulatory policies, which will be elusive if not impossible to create and implement, but on common ground for the principles that ought to guide country-level rules. Such principles might include those from the list proposed by the international consensus committee, including transparency, due care, responsible science adhering to professional norms, promoting wellbeing of those affected, and transnational cooperation. Work to create a set of shared norms is ongoing and worth continued effort as the relevant stakeholders attempt to navigate what can only be called a brave new world.