The Promise of Pills That Know When You Swallow Them
Dr. Sara Browne, an associate professor of clinical medicine at the University of California, San Diego, is a specialist in infectious diseases and, less formally, "a global health person." She often travels to southern Africa to meet with colleagues working on the twin epidemics of HIV and tuberculosis.
"This technology, in my opinion, is an absolute slam dunk for tuberculosis."
Lately she has asked them to name the most pressing things she can help with as a researcher based in a wealthier country. "Over and over and over again," she says, "the only thing they wanted to know is whether their patients are taking the drugs."
Tuberculosis is one of world's deadliest diseases; every year there are 10 million new infections and more than a million deaths. When a patient with tuberculosis is prescribed medicine to combat the disease, adherence to the regimen is important not just for the individual's health, but also for the health of the community. Poor adherence can lead to lengthier and more costly treatment and, perhaps more importantly, to drug-resistant strains of the disease -- an increasing global threat.
Browne is testing a new method to help healthcare workers track their patients' adherence with greater precision—close to exact precision even. They're called digital pills, and they involve a patient swallowing medicine as they normally would, only the capsule contains a sensor that—when it contacts stomach acid—transmits a signal to a small device worn on or near the body. That device in turn sends a signal to the patient's phone or tablet and into a cloud-based database. The fact that the pill has been swallowed has therefore been recorded almost in real time, and notice is available to whoever has access to the database.
"This technology, in my opinion, is an absolute slam dunk for tuberculosis," Browne says. TB is much more prevalent in poorer regions of the world—in Sub-Saharan Africa, for example—than in richer places like the U.S., where Browne's studies thus far have taken place. But when someone is diagnosed in the U.S., because of the risk to others if it spreads, they will likely have to deal with "directly observed therapy" to ensure that they take their medicines correctly.
DOT, as it's called, requires the patient to meet with a healthcare worker several days a week, or every day, so that the medicine intake can be observed in person -- an expensive and time-consuming process. Still, the Centers for Disease Control and Prevention website says (emphasis theirs), "DOT should be used for ALL patients with TB disease, including children and adolescents. There is no way to accurately predict whether a patient will adhere to treatment without this assistance."
Digital pills can help with both the cost and time involved, and potentially improve adherence in places where DOT is impossibly expensive. With the sensors, you can monitor a patient's adherence without a healthcare worker physically being in the room. Patients can live their normal lives and if they miss a pill, they can receive a reminder by text or a phone call from the clinic or hospital. "They can get on with their lives," said Browne. "They don't need the healthcare system to interrupt them."
A 56-year-old patient who participated in one of Browne's studies when he was undergoing TB treatment says that before he started taking the digital pills, he would go to the clinic at least once every day, except weekends. Once he switched to digital pills, he could go to work and spend time with his wife and children instead of fighting traffic every day to get to the clinic. He just had to wear a small patch on his abdomen, which would send the signal to a tablet provided by Browne's team. When he returned from work, he could see the results—that he'd taken the pill—in a database accessed via the tablet. (He could also see his heart rate and respiratory rate.) "I could do my daily activities without interference," he said.
Dr. Peter Chai, a medical toxicologist and emergency medicine physician at Brigham and Women's Hospital in Boston, is studying digital pills in a slightly different context, to help fight the country's opioid overdose crisis. Doctors like Chai prescribe pain medicine, he says, but then immediately put the onus on the patient to decide when to take it. This lack of guidance can lead to abuse and addiction. Patients are often told to take the meds "as needed." Chai and his colleagues wondered, "What does that mean to patients? And are people taking more than they actually need? Because pain is such a subjective experience."
The patients "liked the fact that somebody was watching them."
They wanted to see what "take as needed" actually led to, so they designed a study with patients who had broken a bone and come to the hospital's emergency department to get it fixed. Those who were prescribed oxycodone—a pharmaceutical opioid for pain relief—got enough digital pills to last one week. They were supposed to take the pills as needed, or as many as three pills per day. When the pills were ingested, the sensor sent a signal to a card worn on a lanyard around the neck.
Chai and his colleagues were able to see exactly when the patients took the pills and how many, and to detect patterns of ingestion more precisely than ever before. They talked to the patients after the seven days were up, and Chai said most were happy to be taking digital pills. The patients saw it as a layer of protection from afar. "They liked the fact that somebody was watching them," Chai said.
Both doctors, Browne and Chai, are in early stages of studies with patients taking pre-exposure prophylaxis, medicines that can protect people with a high-risk of contracting HIV, such as injectable drug users. Without good adherence, patients leave themselves open to getting the virus. If a patient is supposed to take a pill at 2 p.m. but the digital pill sensor isn't triggered, the healthcare provider can have an automatic message sent as a reminder. Or a reminder to one of the patient's friends or loved ones.
"Like Swallowing Your Phone"?
Deven Desai, an associate professor of law and ethics at Georgia Tech, says that digital pills sound like a great idea for helping with patient adherence, a big issue that self-reporting doesn't fully solve. He likes the idea of a physician you trust having better information about whether you're taking your medication on time. "On the surface that's just cool," he says. "That's a good thing." But Desai, who formerly worked as academic research counsel at Google, said that some of the same questions that have come up in recent years with social media and the Internet in general also apply to digital pills.
"Think of it like your phone, but you swallowed it," he says. "At first it could be great, simple, very much about the user—in this case, the patient—and the data is going between you and your doctor and the medical people it ought to be going to. Wonderful. But over time, phones change. They become 'smarter.'" And when phones and other technologies become smarter, he says, the companies behind them tend to expand the type of data they collect, because they can. Desai says it will be crucial that prescribers be completely transparent about who is getting the patients' data and for what purpose.
"We're putting stuff in our body in good faith with our medical providers, and what if it turned out later that all of a sudden someone was data mining or putting in location trackers and we never knew about that?" Desai asks. "What science has to realize is if they don't start thinking about this, what could be a wonderful technology will get killed."
Leigh Turner, an associate professor at the University of Minnesota's Center for Bioethics, agrees with Desai that digital pills have great promise, and also that there are clear reasons to be concerned about their use. Turner compared the pills to credit cards and social media, in that the data from them can potentially be stolen or leaked. One question he would want answered before the pills were normalized: "What kind of protective measures are in place to make sure that personal information isn't spilling out and being acquired by others or used by others in unexpected and unwanted ways?"
If digital pills catch on, some experts worry that they may one day not be a voluntary technology.
Turner also wonders who will have access to the pills themselves. Only those who can afford both the medicine plus the smartphones that are currently required for their use? Or will people from all economic classes have access? If digital pills catch on, he also worries they may one day not be a voluntary technology.
"When it comes to digital pills, it's not something that's really being foisted on individuals. It's more something that people can be informed of and can choose to take or not to take," he says. "But down the road, I can imagine a scenario where we move away from purely voluntary agreements to it becoming more of an expectation."
He says it's easy to picture a scenario in which insurance companies demand that patient medicinal intake data be tracked and collected or else. Refuse to have your adherence tracked and you risk higher rates or even overall coverage. Maybe patients who don't take the digital pills suffer dire consequences financially or medically. "Maybe it becomes beneficial as much to health insurers and payers as it is to individual patients," Turner says.
In November 2017, the FDA approved the first-ever digital pill that includes a sensor, a drug called Abilify MyCite, made by Otsuka Pharmaceutical Company. The drug, which is yet to be released, is used to treat schizophrenia, bipolar disorder, and depression. With a built-in sensor developed by Proteus Digital Health, patients can give their doctors permission to see when exactly they are taking, or not taking, their meds. For patients with mental illness, the ability to help them stick to their prescribed regime can be life-saving.
But Turner wonders if Abilify is the best drug to be a forerunner for digital pills. Some people with schizophrenia might be suffering from paranoia, and perhaps giving them a pill developed by a large corporation that sends data from their body to be tracked by other people might not be the best idea. It could in fact exacerbate their sense of paranoia.
The Bottom Line: Protect the Data
We all have relatives who have pillboxes with separate compartments for each day of the week, or who carry pillboxes that beep when it's time to take the meds. But that's not always good enough for people with dementia, mental illness, drug addiction, or other life situations that make it difficult to remember to take their pills. Digital pills can play an important role in helping these people.
"The absolute principle here is that the data has to belong to the patient."
The one time the patient from Browne's study forgot to take his pills, he got a beeping reminder from his tablet that he'd missed a dose. "Taking a medication on a daily basis, sometimes we just forget, right?" he admits. "With our very accelerated lives nowadays, it helps us to remember that we have to take the medications. So patients are able to be on top of their own treatment."
Browne is convinced that digital pills can help people in developing countries with high rates of TB and HIV, though like Turner and Desai she cautions that patients' data must be protected. "I think it can be a tremendous technology for patient empowerment and I also think if properly used it can help the medical system to support patients that need it," she said. "But the absolute principle here is that the data has to belong to the patient."
Many leaders at top companies are trying to get workers to return to the office. They say remote and hybrid work are bad for their employees’ mental well-being and lead to a sense of social isolation, meaninglessness, and lack of work-life boundaries, so we should just all go back to office-centric work.
One example is Google, where the company’s leadership is defending its requirement of mostly in-office work for all staff as necessary to protect social capital, meaning people’s connections to and trust in one another. That’s despite a survey of over 1,000 Google employees showing that two-thirds feel unhappy about being forced to work in the office three days per week. In internal meetings and public letters, many have threatened to leave, and some are already quitting to go to other companies with more flexible options.
Last month, GM rolled out a policy similar to Google’s, but had to backtrack because of intense employee opposition. The same is happening in some places outside of the U.S. For instance, three-fifths of all Chinese employers are refusing to offer permanent remote work options, according to a survey this year from The Paper.
For their claims that remote work hurts well-being, some of these office-centric traditionalists cite a number of prominent articles. For example, Arthur Brooks claimed in an essay that “aggravation from commuting is no match for the misery of loneliness, which can lead to depression, substance abuse, sedentary behavior, and relationship damage, among other ills.” An article in Forbes reported that over two-thirds of employees who work from home at least part of the time had trouble getting away from work at the end of the day. And Fast Company has a piece about how remote work can “exacerbate existing mental health issues” like depression and anxiety.
For his part, author Malcolm Gladwell has also championed a swift return to the office, saying there is a “core psychological truth, which is we want you to have a feeling of belonging and to feel necessary…I know it’s a hassle to come into the office, but if you’re just sitting in your pajamas in your bedroom, is that the work life you want to live?”
These arguments may sound logical to some, but they fly in the face of research and my own experience as a behavioral scientist and as a consultant to Fortune 500 companies. In these roles, I have seen the pitfalls of in-person work, which can be just as problematic, if not more so. Remote work is not without its own challenges, but I have helped 21 companies implement a series of simple steps to address them.
Research finds that remote work is actually better for you
The trouble with the articles described above - and claims by traditionalist business leaders and gurus - stems from a sneaky misdirection. They decry the negative impact of remote and hybrid work for wellbeing. Yet they gloss over the damage to wellbeing caused by the alternative, namely office-centric work.
It’s like comparing remote and hybrid work to a state of leisure. Sure, people would feel less isolated if they could hang out and have a beer with their friends instead of working. They could take care of their existing mental health issues if they could visit a therapist. But that’s not in the cards. What’s in the cards is office-centric work. That means the frustration of a long commute to the office, sitting at your desk in an often-uncomfortable and oppressive open office for at least 8 hours, having a sad desk lunch and unhealthy snacks, sometimes at an insanely expensive cost and, for making it through this series of insults, you’re rewarded with more frustration while commuting back home.
In a 2022 survey, the vast majority of respondents felt that working remotely improved their work-life balance. Much of that improvement stemmed from saving time due to not needing to commute and having a more flexible schedule.
So what happens when we compare apples to apples? That’s when we need to hear from the horse’s mouth: namely, surveys of employees themselves, who experienced both in-office work before the pandemic, and hybrid and remote work after COVID struck.
Consider a 2022 survey by Cisco of 28,000 full-time employees around the globe. Nearly 80 percent of respondents say that remote and hybrid work improved their overall well-being: that applies to 83 percent of Millennials, 82 percent of Gen Z, 76 percent of Gen Z, and 66 percent of Baby Boomers. The vast majority of respondents felt that working remotely improved their work-life balance.
Much of that improvement stemmed from saving time due to not needing to commute and having a more flexible schedule: 90 percent saved 4 to 8 hours or more per week. What did they do with that extra time? The top choice for almost half was spending more time with family, friends and pets, which certainly helped address the problem of isolation from the workplace. Indeed, three-quarters of them report that working from home improved their family relationships, and 51 percent strengthened their friendships. Twenty percent used the freed up hours for self-care.
Of the small number who report their work-life balance has not improved or even worsened, the number one reason is the difficulty of disconnecting from work, but 82 percent report that working from anywhere has made them happier. Over half say that remote work decreased their stress levels.
Other surveys back up Cisco’s findings. For example, a 2022 Future Forum survey compared knowledge workers who worked full-time in the office, in a hybrid modality, and fully remote. It found that full-time in-office workers felt the least satisfied with work-life balance, hybrid workers were in the middle, and fully remote workers felt most satisfied. The same distribution applied to questions about stress and anxiety. A mental health website called Tracking Happiness found in a 2022 survey of over 12,000 workers that fully remote employees report a happiness level about 20 percent greater than office-centric ones. Another survey by CNBC in June found that fully remote workers are more often very satisfied with their jobs than workers who are fully in-person.
Academic peer-reviewed research provides further support. Consider a 2022 study published in the International Journal of Environmental Research and Public Health of bank workers who worked on the same tasks of advising customers either remotely or in-person. It found that fully remote workers experienced higher meaningfulness, self-actualization, happiness, and commitment than in-person workers. Another study, published by the National Bureau of Economic Research, reported that hybrid workers, compared to office-centric ones, experienced higher satisfaction with work and had 35 percent more job retention.
What about the supposed burnout crisis associated with remote work? Indeed, burnout is a concern. A survey by Deloitte finds that 77 percent of workers experienced burnout at their current job. Gallup came up with a slightly lower number of 67 percent in its survey. But guess what? Both of those surveys are from 2018, long before the era of widespread remote work.
By contrast, in a Gallup survey in late 2021, 58 percent of respondents reported less burnout. An April 2021 McKinsey survey found burnout in 54 percent of Americans and 49 percent globally. A September 2021 survey by The Hartford reported 61 percent burnout. Arguably, the increase in full or part-time remote opportunities during the pandemic helped to address feelings of burnout, rather than increasing them. Indeed, that finding aligns with the earlier surveys and peer-reviewed research suggesting remote and hybrid work improves wellbeing.
Remote work isn’t perfect – here’s how to fix its shortcomings
Still, burnout is a real problem for hybrid and remote workers, as it is for in-office workers. Employers need to offer mental health benefits with online options to help employees address these challenges, regardless of where they’re working.
Moreover, while they’re better overall for wellbeing, remote and hybrid work arrangements do have specific disadvantages around work-life separation. To address work-life issues, I advise my clients who I helped make the transition to hybrid and remote work to establish norms and policies that focus on clear expectations and setting boundaries.
For working at home and collaborating with others, there’s sometimes an unhealthy expectation that once you start your workday in your home office chair, and that you’ll work continuously while sitting there.
Some people expect their Slack or Microsoft Teams messages to be answered within an hour, while others check Slack once a day. Some believe email requires a response within three hours, and others feel three days is fine. As a result of such uncertainty and lack of clarity about what’s appropriate, too many people feel uncomfortable disconnecting and not replying to messages or doing work tasks after hours. That might stem from a fear of not meeting their boss’s expectations or not wanting to let their colleagues down.
To solve this problem, companies need to establish and incentivize clear expectations and boundaries. They should develop policies and norms around response times for different channels of communication. They also need to clarify work-life boundaries – for example, the frequency and types of unusual circumstances that will require employees to work outside of regular hours.
Moreover, for working at home and collaborating with others, there’s sometimes an unhealthy expectation that once you start your workday in your home office chair, and that you’ll work continuously while sitting there (except for your lunch break). That’s not how things work in the office, which has physical and mental breaks built in throughout the day. You took 5-10 minutes to walk from one meeting to another, or you went to get your copies from the printer and chatted with a coworker on the way.
Those and similar physical and mental breaks, research shows, decrease burnout, improve productivity, and reduce mistakes. That’s why companies should strongly encourage employees to take at least a 10-minute break every hour during remote work. At least half of those breaks should involve physical activity, such as stretching or walking around, to counteract the dangerous effects of prolonged sitting. Other breaks should be restorative mental activities, such as meditation, brief naps, walking outdoors, or whatever else feels restorative to you.
To facilitate such breaks, my client organizations such as the University of Southern California’s Information Sciences Institute shortened hour-long meetings to 50 minutes and half-hour meetings to 25 minutes, to give everyone – both in-person and remote workers – a mental and physical break and transition time.
Very few people will be reluctant to have shorter meetings. After that works out, move to other aspects of setting boundaries and expectations. Doing so will require helping team members get on the same page and reduce conflicts and tensions. By setting clear expectations, you’ll address the biggest challenge for wellbeing for remote and hybrid work: establishing clear work-life boundaries.
In May 2022, Californian biotech Ultima Genomics announced that its UG 100 platform was capable of sequencing an entire human genome for just $100, a landmark moment in the history of the field. The announcement was particularly remarkable because few had previously heard of the company, a relative unknown in an industry long dominated by global giant Illumina which controls about 80 percent of the world’s sequencing market.
Ultima’s secret was to completely revamp many technical aspects of the way Illumina have traditionally deciphered DNA. The process usually involves first splitting the double helix DNA structure into single strands, then breaking these strands into short fragments which are laid out on a glass surface called a flow cell. When this flow cell is loaded into the sequencing machine, color-coded tags are attached to each individual base letter. A laser scans the bases individually while a camera simultaneously records the color associated with them, a process which is repeated until every single fragment has been sequenced.
Instead, Ultima has found a series of shortcuts to slash the cost and boost efficiency. “Ultima Genomics has developed a fundamentally new sequencing architecture designed to scale beyond conventional approaches,” says Josh Lauer, Ultima’s chief commercial officer.
This ‘new architecture’ is a series of subtle but highly impactful tweaks to the sequencing process ranging from replacing the costly flow cell with a silicon wafer which is both cheaper and allows more DNA to be read at once, to utilizing machine learning to convert optical data into usable information.
To put $100 genome in perspective, back in 2012 the cost of sequencing a single genome was around $10,000, a price tag which dropped to $1,000 a few years later. Before Ultima’s announcement, the cost of sequencing an individual genome was around $600.
Several studies have found that nearly 12 percent of healthy people who have their genome sequenced, then discover they have a variant pointing to a heightened risk of developing a disease that can be monitored, treated or prevented.
While Ultima’s new machine is not widely available yet, Illumina’s response has been rapid. Last month the company unveiled the NovaSeq X series, which it describes as its fastest most cost-efficient sequencing platform yet, capable of sequencing genomes at $200, with further price cuts likely to follow.
But what will the rapidly tumbling cost of sequencing actually mean for medicine? “Well to start with, obviously it’s going to mean more people getting their genome sequenced,” says Michael Snyder, professor of genetics at Stanford University. “It'll be a lot more accessible to people.”
At the moment sequencing is mainly limited to certain cancer patients where it is used to inform treatment options, and individuals with undiagnosed illnesses. In the past, initiatives such as SeqFirst have attempted further widen access to genome sequencing based on growing amounts of research illustrating the potential benefits of the technology in healthcare. Several studies have found that nearly 12 percent of healthy people who have their genome sequenced, then discover they have a variant pointing to a heightened risk of developing a disease that can be monitored, treated or prevented.
“While whole genome sequencing is not yet widely used in the U.S., it has started to come into pediatric critical care settings such as newborn intensive care units,” says Professor Michael Bamshad, who heads the genetic medicine division in the University of Washington’s pediatrics department. “It is also being used more often in outpatient clinical genetics services, particularly when conventional testing fails to identify explanatory variants.”
But the cost of sequencing itself is only one part of the price tag. The subsequent clinical interpretation and genetic counselling services often come to several thousand dollars, a cost which insurers are not always willing to pay.
As a result, while Bamshad and others hope that the arrival of the $100 genome will create new opportunities to use genetic testing in innovative ways, the most immediate benefits are likely to come in the realm of research.
Bigger Data
There are numerous ways in which cheaper sequencing is likely to advance scientific research, for example the ability to collect data on much larger patient groups. This will be a major boon to scientists working on complex heterogeneous diseases such as schizophrenia or depression where there are many genes involved which all exert subtle effects, as well as substantial variance across the patient population. Bigger studies could help scientists identify subgroups of patients where the disease appears to be driven by similar gene variants, who can then be more precisely targeted with specific drugs.
If insurers can figure out the economics, Snyder even foresees a future where at a certain age, all of us can qualify for annual sequencing of our blood cells to search for early signs of cancer or the potential onset of other diseases like type 2 diabetes.
David Curtis, a genetics professor at University College London, says that scientists studying these illnesses have previously been forced to rely on genome-wide association studies which are limited because they only identify common gene variants. “We might see a significant increase in the number of large association studies using sequence data,” he says. “It would be far preferable to use this because it provides information about rare, potentially functional variants.”
Cheaper sequencing will also aid researchers working on diseases which have traditionally been underfunded. Bamshad cites cystic fibrosis, a condition which affects around 40,000 children and adults in the U.S., as one particularly pertinent example.
“Funds for gene discovery for rare diseases are very limited,” he says. “We’re one of three sites that did whole genome sequencing on 5,500 people with cystic fibrosis, but our statistical power is limited. A $100 genome would make it much more feasible to sequence everyone in the U.S. with cystic fibrosis and make it more likely that we discover novel risk factors and pathways influencing clinical outcomes.”
For progressive diseases that are more common like cancer and type 2 diabetes, as well as neurodegenerative conditions like multiple sclerosis and ALS, geneticists will be able to go even further and afford to sequence individual tumor cells or neurons at different time points. This will enable them to analyze how individual DNA modifications like methylation, change as the disease develops.
In the case of cancer, this could help scientists understand how tumors evolve to evade treatments. Within in a clinical setting, the ability to sequence not just one, but many different cells across a patient’s tumor could point to the combination of treatments which offer the best chance of eradicating the entire cancer.
“What happens at the moment with a solid tumor is you treat with one drug, and maybe 80 percent of that tumor is susceptible to that drug,” says Neil Ward, vice president and general manager in the EMEA region for genomics company PacBio. “But the other 20 percent of the tumor has already got mutations that make it resistant, which is probably why a lot of modern therapies extend life for sadly only a matter of months rather than curing, because they treat a big percentage of the tumor, but not the whole thing. So going forwards, I think that we will see genomics play a huge role in cancer treatments, through using multiple modalities to treat someone's cancer.”
If insurers can figure out the economics, Snyder even foresees a future where at a certain age, all of us can qualify for annual sequencing of our blood cells to search for early signs of cancer or the potential onset of other diseases like type 2 diabetes.
“There are companies already working on looking for cancer signatures in methylated DNA,” he says. “If it was determined that you had early stage cancer, pre-symptomatically, that could then be validated with targeted MRI, followed by surgery or chemotherapy. It makes a big difference catching cancer early. If there were signs of type 2 diabetes, you could start taking steps to mitigate your glucose rise, and possibly prevent it or at least delay the onset.”
This would already revolutionize the way we seek to prevent a whole range of illnesses, but others feel that the $100 genome could also usher in even more powerful and controversial preventative medicine schemes.
Newborn screening
In the eyes of Kári Stefánsson, the Icelandic neurologist who been a visionary for so many advances in the field of human genetics over the last 25 years, the falling cost of sequencing means it will be feasible to sequence the genomes of every baby born.
“We have recently done an analysis of genomes in Iceland and the UK Biobank, and in 4 percent of people you find mutations that lead to serious disease, that can be prevented or dealt with,” says Stefansson, CEO of deCODE genetics, a subsidiary of the pharmaceutical company Amgen. “This could transform our healthcare systems.”
As well as identifying newborns with rare diseases, this kind of genomic information could be used to compute a person’s risk score for developing chronic illnesses later in life. If for example, they have a higher than average risk of colon or breast cancer, they could be pre-emptively scheduled for annual colonoscopies or mammograms as soon as they hit adulthood.
To a limited extent, this is already happening. In the UK, Genomics England has launched the Newborn Genomes Programme, which plans to undertake whole-genome sequencing of up to 200,000 newborn babies, with the aim of enabling the early identification of rare genetic diseases.
"I have not had my own genome sequenced and I would not have wanted my parents to have agreed to this," Curtis says. "I don’t see that sequencing children for the sake of some vague, ill-defined benefits could ever be justifiable.”
However, some scientists feel that it is tricky to justify sequencing the genomes of apparently healthy babies, given the data privacy issues involved. They point out that we still know too little about the links which can be drawn between genetic information at birth, and risk of chronic illness later in life.
“I think there are very difficult ethical issues involved in sequencing children if there are no clear and immediate clinical benefits,” says Curtis. “They cannot consent to this process. I have not had my own genome sequenced and I would not have wanted my parents to have agreed to this. I don’t see that sequencing children for the sake of some vague, ill-defined benefits could ever be justifiable.”
Curtis points out that there are many inherent risks about this data being available. It may fall into the hands of insurance companies, and it could even be used by governments for surveillance purposes.
“Genetic sequence data is very useful indeed for forensic purposes. Its full potential has yet to be realized but identifying rare variants could provide a quick and easy way to find relatives of a perpetrator,” he says. “If large numbers of people had been sequenced in a healthcare system then it could be difficult for a future government to resist the temptation to use this as a resource to investigate serious crimes.”
While sequencing becoming more widely available will present difficult ethical and moral challenges, it will offer many benefits for society as a whole. Cheaper sequencing will help boost the diversity of genomic datasets which have traditionally been skewed towards individuals of white, European descent, meaning that much of the actionable medical information which has come out of these studies is not relevant to people of other ethnicities.
Ward predicts that in the coming years, the growing amount of genetic information will ultimately change the outcomes for many with rare, previously incurable illnesses.
“If you're the parent of a child that has a susceptible or a suspected rare genetic disease, their genome will get sequenced, and while sadly that doesn’t always lead to treatments, it’s building up a knowledge base so companies can spring up and target that niche of a disease,” he says. “As a result there’s a whole tidal wave of new therapies that are going to come to market over the next five years, as the genetic tools we have, mature and evolve.”