For most patients, treatment guidelines are fuzzy at best and, at worst, nonexistent, forcing clinicians to rely on educated guesswork. But thanks to advances in computation, data processing and telecommunication, that may be about to change.
October 26, 2015 - By Bruce Goldman
It’s hardly a secret among medical practitioners: For most patients, clear treatment guidelines simply don’t exist.
Take Vera. She is a 55-year-old woman of Vietnamese descent who has asthma. You’re her doctor, and you’ve just learned that she also has high blood pressure. Vera’s case doesn’t fit the data from any clinical trials; there’s no medical literature on hypertension medications for middle-aged, asthmatic Vietnamese-American women. You want to treat her hypertension, but you have no guidelines. Vera is sitting in your exam room now. What do you do?
Suppose you could get some guidance simply by pressing a virtual button on a computer screen displaying Vera’s electronic medical record? This would trigger a search of millions of other electronic records and, in a matter of minutes, generate a succinct composite summary of the outcomes of 25 or 100 or perhaps 1,000 patients very similar to her — same race, same height, same age, same symptoms, lookalike lab-test results — who were given various antihypertensive medications. Patients similar to Vera, it turns out, respond particularly well to one drug type — something you would have been hard put to guess on your own.
Vera is a made-up patient, but there are plenty of people with conditions that are just as complicated. Scattered throughout millions of electronic medical records, such lookalike cases could point the way to effective treatments options for Vera and others if these could be plucked from the aggregate and formatted for easy interpretation. While some aspects of this approach, such as assuring patients that their privacy will be protected and making databases compatible across health-care systems, need to be worked out, Stanford medical researchers want to tackle those problems. The goal is a seamless system that quickly links physicians to the information they need in order to give their patients the best-validated treatments available.
A pediatrician, a cardiologist and a biomedical informaticist at Stanford Medicine have come up with an idea that could revolutionize the practice of medicine. Their brainchild, which they’ve dubbed the Green Button, will do that by tapping the huge volumes of data lying dormant in the EMRs of millions of patients.
The Green Button approach leverages the increasingly routine use of these records and the fast-paced progress taking place in computation and data transmission. It could enable a real-time solution to a big problem: the inadequacy of results from clinical trials — the foundation upon which treatment guidelines are built — for the vast majority of patients. Clinical trials are experiments in which new medications and procedures are tested on people. In order to achieve understandable results, investigators tend to select participants for trials who are a lot alike in terms of age, sex, ethnicity, medical conditions and treatment history. Yet the average patient walking into a doctor’s office seldom resembles a patient included in those trials.
“Every day I encounter patients for whom we just don’t have the best scientific evidence on how to treat them,” said Christopher Longhurst, MD, clinical professor of pediatrics in system medicine and chief medical information officer for Stanford Children's Health.
In a 2014 Health Affairs article, Longhurst along with Nigam Shah, MBBS, PhD, assistant professor of biomedical research and assistant director of the Stanford Center for Biomedical Informatics Research, and Robert Harrington, MD, professor and chair of medicine, outlined a vision for drawing medical guidance from day-to-day medical practice in hospitals and doctors’ offices. They called it the Green Button. The idea was to give doctors access to aggregate patient data, right there and then, from a vast collection of EMRs. This near-instant output isn’t a substitute for a clinical trial, but it’s a lot better than nothing — or than resorting to the physician’s own bias-prone memory of one or two previous encounters with similar patients.
“You don’t have to type anything in,” said Shah. “Just press the Green Button.”
From the gold standard to the Green Button
The randomized clinical trial is considered the gold standard of medical research. In a randomized clinical trial, a number of participants are randomly assigned to one of two — or sometimes more — groups. One group gets the drug or the procedure being tested; the other is given a placebo or undergoes a sham procedure. Ideally, the study is blinded — patients don’t know which option they’re getting — or even better, double-blinded — the investigators and their assistants don’t know, either. Once the trial’s active phase ends, rigorous statistical analysis determines whether the hypothesis, spelled out in advance of the trial, was fulfilled.
“It goes without saying that you should use randomized trial evidence when it’s available,” said Harrington, who also holds the Arthur L. Bloomfield Professorship of Medicine. “But a lot of times, it’s not.”
Shah concurred. “Clinical trials select only a small, artificial subset of the real population,” he said. “A regular, ordinary person who walks into the doctor’s office doesn’t usually fit.”
He continued: “Clinical trials are designed to prove one thing, and you’re testing it on people with just one thing: Type 2 diabetes, eczema, whatever. But most real-life people don’t have just one thing. They have three or four or five things.”
As a result, “only about 4 percent of the time have you got a clinical-trial-based guideline applicable to the patient facing you right now,” Shah said. The rest of the time, doctors must rely on their own judgment.
Yet even though there may not be clinical-trial evidence to guide a doctor’s choice of treatment options for a particular patient, “tons of applicable evidence” are locked away in health systems’ EMRs, Shah said. The inspiration for the 2014 Health Affairs paper in which he, Longhurst and Harrington elaborated their Green Button concept was a real-life, real-time data search conducted by Jennifer Frankovich, MD, now a clinical assistant professor of pediatric rheumatology at Stanford. A 13-year-old girl with lupus had been admitted to Lucile Packard Children’s Hospital Stanford with severe kidney and pancreatic inflammation. She was considered at risk for blood clots. While anticoagulants could counteract clotting, they would also increase her risk of bleeding from some procedures likely to be used during her hospital stay. There were no clear clinical-trial-based guidelines on whether to give the girl anticoagulants, and different clinicians had different thoughts about what was advisable.
But owing to a research project she was involved in, Frankovich had access to a Stanford database containing the EMRs of pediatric lupus patients admitted between 2004 and 2009. So she was able to perform an on-the-spot analysis of the outcomes of 98 kids who’d been in situations similar to the one confronting her patient. Within four hours, it was clear to Frankovich that kidney and pancreatic complications put kids with lupus at much higher risk of clotting.
Frankovich and her teammates decided to give the girl anticoagulants right away. The young patient suffered no clotting or other adverse events. Frankovich was the lead author of a 2011 article in The New England Journal of Medicine describing the story, of which Longhurst was a co-author.
That serendipitous result, said Longhurst, led to a follow-on question: “How can we go about doing this in a purposeful way on a continuing, case-by-case basis?”
With advancing technology, the kind of analysis Frankovich performed can be completed in considerably less than an hour today — soon enough for an outpatient finishing an appointment. But there are several obstacles to achieving this goal.
The stumbling blocks along the road to the Green Button’s realization aren’t primarily technical — the methodologies are available, and the infrastructure is buildable. But the more idiosyncratic your patient’s case is, the larger the initial pool of patient data needs to be. And achieving the scale necessary for generating enough records of lookalike patients to provide meaningful results presents some challenges.
As Shah put it: “What if you press the Green Button and nothing happens?” If you can’t access enough records of similar patients to begin with, you’re out of luck.
Assembling that huge data pool gets easier if numerous institutions can be coaxed into contributing to it. The numbers are certainly there: Stanford Health Care alone has close to 2 million patient EMRs. Kaiser Permanente, which has been using EMRs for a decade or more, has 9 million, and the University of California health system has 14 million. The U.S. Department of Veterans Affairs has 20 to 25 years’ worth of longitudinal data on many millions of veterans.
The key lies in integrating these disparate databases to yield valuable, personalized medical insights.
But sharing data across institutions is no simple matter. “Any hospital CEO today would kick you out of the office if you propose data-sharing,” Shah said. “That’s rational on their part. Sharing data puts you at risk of leaks, and compromised patient privacy can mean big financial and public-relations pitfalls.”
Federal law guards patients’ privacy, but it doesn’t make the data in their medical records totally off limits. For instance, as Longhurst points out, the law specifically allows the use of patient data for improving quality of care.
Yet even if the patient-privacy issue turns out to be insurmountable in the short run, there’s a workaround, Shah said: Health systems could share with one another descriptions of the kinds of patients they’re looking for, rather than request raw patient data. Thus, a health system that got a request for information on patients of Vietnamese descent with asthma and high blood pressure would, in accordance with such an arrangement, automatically search its own database and share only statistical summaries of what it found, such as the range of outcomes for certain medications given to this cohort.
The point is not to outsmart the physician.
Meanwhile, there is progress. Stanford researchers including Shah and Longhurst have published numerous studies establishing the power of pooling large volumes of data to derive clinically beneficial results. The Stanford Center for Population Health Sciences, directed by Mark Cullen, MD, professor of medicine, is putting in place a data library housing the records of some 10 million different patients, purchased from another institution.
These developments are keyed to efforts around precision health, Stanford Medicine’s push to anticipate and prevent disease in the healthy and precisely diagnose and treat disease in the ill. Precision health aims to give researchers and physicians better tools for predicting individual risks for specific diseases, developing approaches to early detection and prevention, and helping clinicians make real-time decisions about the best way to care for patients.
Asked whether the Green Button idea could meet resistance from medical practitioners who object to taking orders from an algorithm, Shah said, “The point is not to outsmart the physician. The point is to tell you the outcomes of the best guesses of 100 of your colleagues. You can choose to interpret or ignore it.”
Some smart money is betting that these stumbling blocks can be hurdled. Kyron Inc., a Palo Alto-based start-up Shah cofounded in 2013 with technologist Louis Monier, PhD, and Stanford-trained biomedical informaticist Noah Zimmerman, PhD, has raised several million dollars and has licensed informatics-associated technology from Stanford’s Office of Technology Licensing to do just that.
Build your own randomized trial
“Virtually 100 percent of the 3,000 kids who get diagnosed with cancer every year in the U.S. are in clinical trials,” said Longhurst. “How many adults with cancer are in clinical trials? Maybe 2 or 3 percent — we can’t possibly afford to put 100 percent of adults into trials. So the other 97 percent may be getting treated, but the health-care system isn’t learning anything from their outcomes.” For his part, cardiologist Harrington noted that fewer than 10 percent of heart-attack patients are actually enrolled in a clinical trial.
The Green Button will let clinicians learn more from the patients they’re caring for each time they see one of them.
The Green Button approach may be able not only to substitute for randomized trials, but to generate them. Suppose you’re a doctor, and a patient walks into your office. You take the patient’s history, perform a workup, update the patient’s EMR accordingly and hit the Green Button. But, it turns out, there’s not enough data on similar patients to provide decent information on which of two treatment options is best for this patient.
But that’s not the end of it. The Green Button now shifts gears from merely downloading outcomes of other patients to what Harrington has termed “point-of-care randomization”: You give this patient one of the two treatments, and the next patient who walks in the door and is similar to this one gets the other treatment. Keep alternating prescriptions to successive similar patients — while monitoring their responses to minimize the chances of either treatment doing them any harm — and you will have increasingly large cohorts fueling an informed conclusion.
“Applied this way, the Green Button will let clinicians learn more from the patients they’re caring for each time they see one of them,” said Shah. “Every patient becomes part of a scientific experiment.”
Stanford Medicine integrates research, medical education and health care at its three institutions - Stanford School of Medicine, Stanford Health Care, and Stanford Children's Health. For more information, please visit the Office of Communications website at http://mednews.stanford.edu.