The average wait times in U.S. Pat. emergency rooms top two hours, leaving both clinicians and patients to feel the pain of an overburdened system. Many and parent has endured those hours with a distressed child, triaged out for lack of urgency only to be sent home with unneeded antibiotics for a garden-variety viral infection.
soak up, the chance to revisit old-fashioned physician house calls holds a strong appeal. What if the visit came from an intelligent machine? AI systems are already adept at recognizing patterns in medical imaging to diagnosis. New findings published February 11 in Nature Medicine show similar training for the diagnosis from the raw data in a child's medical chart.
For this study at Guangzhou Women and Children's Medical Center in southern China, a team of physicians distilled information from thousands of health records into key words linked to different diagnoses. Investigators then get these key words to the AI system so it could detect the terms in real medical charts. Once trained, the system combed the electronic health records (EHRs) or 567,498 children, parsing the real-world physician notes and highlighting important information.
It drilled down from broad to specific diagnoses from among 55 categories. So how did the robo-doc do? "I think it's pretty good," says Mustafa Bashir, an associate professor of radiology at Duke University Medical Center who was not involved in the work. "Conceptually, it's not the original, but the size of the data set and successful execution are important." The data processing, Bashir says, follows the typical steps of taking a big giant messy data set, putting it through an algorithm and yielding order from the chaos. In that sense, he says, the work is not especially novel, but "that said, their system does appear to perform well."
The practice of medicine is both art and a science. Skeptics might argue a computer that has processed a lot of patient data cannot furnish the type of qualitative judgment made by a general practitioner to diagnosis a human from a distance. In this case, though, a lot of human expertise was brought to bear before the machine training began. “This was a massive project that we started about four years ago,” says study author Kang Zhang, a professor of ophthalmology and chief of ophthalmic genetics at the University of California, San Diego. He and his colleagues began with a team of physicians reviewing 6,1
To check the system's accuracy, Zhang and his colleagues also employed old-fashioned " technology ”—human diagnostics. They compared the machine's conclusions with those in the original records – and they had another team of clinicians making diagnoses using the same data as the AI system.
The machine received good grades, agreeing with the humans about 90 percent of the time. It was especially effective at identifying neuropsychiatric conditions and upper respiratory diseases. For acute upper respiratory infection, the most common diagnosis in the huge patient group, the AI system got it right 95 percent of the time. Would 95 percent be good enough? One of the next questions that needs to be researched, Zhang says, is whether the system will miss something dire. The benchmark, he says, should be how senior physicians perform, which is also not 100 percent.
A human clinician would serve as a quality-control backup for the AI system. In fact, human and machine would probably follow a similar series of steps. Just like a doctor, the machine starts with a broad category, such as "respiratory system," and works from the top down to arrive at a diagnosis. "It mimics the human physician's decision progress," says Dongxiao Zhu, an associate professor of computer science at Wayne State University who did not take part in the study.
But Zhu sees this as "augmented intelligence" rather than "artificial intelligence ”Because the system only handled 55 diagnostic options, not the thousands of possibilities in the real world. The machine does not yet exist in the more complex aspects of a diagnosis such as accompanying conditions or disease stage, he says. How well this system can translate outside of its Chinese setting remains unclear. Bashir says although applying AI to patient information would be difficult anywhere, these authors have proved it is achievable [further] Zhu expresses additional skepticism. Pulling diagnostic key words from text notes in EHR will be "radically different" in a language like English rather than Chinese, he says. He also has all the work required for only 55 diagnoses, including the human energy of pediatricians grading 11,926 records for comparison of their conclusions with the machine's diagnoses. Given the four years the overall process required, parents probably have a long wait ahead of a computerized clinician can save them that visit to the ER.