Rizwan Malik had always had an interest in AI. As the lead radiologist at the Royal Bolton Hospital, run by the UK’s National Health Service (NHS), he saw its potential to make his job easier. In his hospital, patients often had to wait six hours or more for a specialist to look at their x-rays. If an emergency room doctor could get an initial reading from an AI-based tool, it could dramatically shrink that wait time. A specialist could follow up the AI system’s reading with a more thorough diagnosis later.
So in September of last year, Malik took it upon himself to design a conservative clinical trial that would help showcase the technology’s potential. He identified a promising AI-based chest x-ray system called qXR from the Mumbai-based company Qure.ai. He then proposed to test the system over six months. For all chest x-rays handled by his trainees, it would offer a second opinion. If those opinions consistently matched his own, he would then phase the system in permanently to double-check his trainees’ work for him. After four months of reviews from multiple hospital and NHS committees and forums, the proposal was finally approved.
More on coronavirus
Our most essential coverage of covid-19 is free, including:
Newsletter: Coronavirus Tech Report
Zoom show: Radio Corona
But before the trial could kick off, covid-19 hit the UK. What began as a pet interest suddenly looked like a blessing. Early research had shown that in radiology images, the most severe covid cases displayed distinct lung abnormalities associated with viral pneumonia. With shortages and delays in PCR tests, chest x-rays had become one of the fastest and most affordable ways for doctors to triage patients.
Within weeks, Qure.ai retooled qXR to detect covid-induced pneumonia, and Malik proposed a new clinical trial, pushing for the technology to perform initial readings rather than just double-check human ones. Normally, the updates to both the tool and the trial design would have initiated a whole new approval process. But without more months to spare, the hospital greenlighted the adjusted proposal immediately. “The medical director basically said, ‘Well, do you know what? If you think it’s good enough, crack on and do it,’” Malik recalls. “‘We’ll deal with all the rest of it after the event.’”
The Royal Bolton Hospital is among a growing number of health-care facilities around the world that are turning to AI to help manage the coronavirus pandemic. Many are using such technologies for the first time under the pressure of staff shortages and overwhelming patient loads. In parallel, dozens of AI firms have developed new software or revamped existing tools, hoping to cash in by cultivating new client relationships that will continue after the crisis is over.
The pandemic, in other words, has turned into a gateway for AI adoption in health care—bringing both opportunity and risk. On the one hand, it is pushing doctors and hospitals to fast-track promising new technologies. On the other, this accelerated process could allow unvetted tools to bypass regulatory processes, putting patients in harm’s way.
“At a high level, artificial intelligence in health care is very exciting,” says Chris Longhurst, the chief information officer at UC San Diego Health. “But health care is one of those industries where there are a lot of factors that come into play. A change in the system can have potentially fatal unintended consequences.”
“It’s really an attractive solution”
Before the pandemic, health-care AI was already a booming area of research. Deep learning, in particular, has demonstrated impressive results for analyzing medical images to identify diseases like breast and lung cancer or glaucoma at least as accurately as human specialists. Studies have also shown the potential of using computer vision to monitor elderly people in their homes and patients in intensive care units.
But there have been significant obstacles to translating that research into real-world applications. Privacy concerns make it challenging to collect enough data for training algorithms; issues related to bias and generalizability make regulators cautious to grant approvals. Even for applications that do get certified, hospitals rightly have their own intensive vetting procedures and established protocols. “Physicians, like everybody else—we’re all creatures of habit,” says Albert Hsiao, a radiologist at UCSD Health who is now trialing his own covid detection algorithm based on chest x-rays. “We don’t change unless we’re forced to change.”
As a result, AI has been slow to gain a foothold. “It feels like there’s something there; there are a lot of papers that show a lot of promise,” said Andrew Ng, a leading AI practitioner, in a recent webinar on its applications in medicine. But “it’s not yet as widely deployed as we wish.”
Pierre Durand, a physician and radiologist based in France, experienced the same difficulty when he cofounded the teleradiology firm Vizyon in 2018. The company operates as a middleman: it licenses software from firms like Qure.ai and a Seoul-based startup called Lunit and offers the package of options to hospitals. Before the pandemic, however, it struggled to gain traction. “Customers were interested in the artificial-intelligence application for imaging,” Durand says, “but they could not find the right place for it in their clinical setup.”
The onset of covid-19 changed that. In France, as caseloads began to overwhelm the health-care system and the government failed to ramp up testing capacity, triaging patients via chest x-ray—though less accurate than a PCR diagnostic—became a fallback solution. Even for patients who could get genetic tests, results could take at least 12 hours and sometimes days to return—too long for a doctor to wait before deciding whether to isolate someone. By comparison, Vizyon’s system using Lunit’s software, for example, takes only 10 minutes to scan a patient and calculate a probability of infection. (Lunit says its own preliminary study found that the tool was comparable to a human radiologist in its risk analysis, but this research has not been published.) “When there are a lot of patients coming,” Durand says, “it’s really an attractive solution.”
Vizyon has since signed partnerships with two of the largest hospitals in the country and says it is in talks with hospitals in the Middle East and Africa. Qure.ai, meanwhile, has now expanded to Italy, the US, and Mexico on top of existing clients. Lunit is also now working with four new hospitals each in France, Italy, Mexico, and Portugal.
In addition to the speed of evaluation, Durand identifies something else that may have encouraged hospitals to adopt AI during the pandemic: they are thinking about how to prepare for the inevitable staff shortages that will arise after the crisis. Traumatic events like a pandemic are often followed by an exodus of doctors and nurses. “Some doctors may want to change their way of life,” he says. “What’s coming, we don’t know.”
“Once people start using our algorithms, they never stop”
Hospitals’ new openness to AI tools hasn’t gone unnoticed. Many companies have begun offering their products for a free trial period, hoping it will lead to a longer contract.
“It's a good way for us to demonstrate the utility of AI,” says Brandon Suh, the CEO of Lunit. Prashant Warier, the CEO and cofounder of Qure.ai, echoes that sentiment. “In my experience outside of covid, once people start using our algorithms, they never stop,” he says.
Both Qure.ai’s and Lunit’s lung screening products were certified by the European Union’s health and safety agency before the crisis. In adapting the tools to covid, the companies repurposed the same functionalities that had already been approved.
Qure.ai’s qXR, for example, uses a combination of deep-learning models to detect common types of lung abnormalities. To retool it, the firm worked with a panel of experts to review the latest medical literature and determine the typical features of covid-induced pneumonia, such as opaque patches in the image that have a “ground glass” pattern and dense regions on the sides of the lungs. It then encoded that knowledge into qXR, allowing the tool to calculate the risk of infection from the number of telltale characteristics present in a scan. A preliminary validation study the firm ran on over 11,000 patient images found that the tool was able to distinguish between covid and non-covid patients with 95% accuracy.
But not all firms have been as rigorous. In the early days of the crisis, Malik exchanged emails with 36 companies and spoke with 24, all pitching him AI-based covid screening tools. “Most of them were utter junk,” he says. “They were trying to capitalize on the panic and anxiety.” The trend makes him worry: hospitals in the thick of the crisis may not have time to perform due diligence. “When you’re drowning so much,” he says, “a thirsty man will reach out for any source of water.”
Kay Firth-Butterfield, the head of AI and machine learning at the World Economic Forum, urges hospitals not to weaken their regulatory protocols or formalize long-term contracts without proper validation. “Using AI to help with this pandemic is obviously a great thing to be doing,” she says. “But the problems that come with AI don’t go away just because there is a pandemic.”
UCSD’s Longhurst also encourages hospitals to use this opportunity to partner with firms on clinical trials. “We need to have clear, hard evidence before we declare this as the standard of care,” he says. Anything less would be “a disservice to patients.”
To have more stories like this delivered directly to your inbox, sign up for our Webby-nominated AI newsletter The Algorithm. It's free.
Why Meta’s latest large language model survived only three days online
Galactica was supposed to help scientists. Instead, it mindlessly spat out biased and incorrect nonsense.
A bot that watched 70,000 hours of Minecraft could unlock AI’s next big thing
Online videos are a vast and untapped source of training data—and OpenAI says it has a new way to use it.
Responsible AI has a burnout problem
Companies say they want ethical AI. But those working in the field say that ambition comes at their expense.
Biotech labs are using AI inspired by DALL-E to invent new drugs
Two groups have announced powerful new generative models that can design new proteins on demand not seen in nature.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.