Google DeepMind introduced AI co-clinician on May 01, 2026, a research initiative designed to examine how multimodal AI systems could support healthcare workers and patients more effectively. The project addresses growing pressure on health systems worldwide to improve outcomes, reduce costs, and expand access to care amid a projected shortage of more than 10 million health workers by 2030, according to the World Health Organization.
The new system is intended to explore a model of “triadic care,” in which an AI agent works alongside a physician and patient rather than replacing clinical judgment. DeepMind stated the goal is to build tools that can extend the reach of clinicians while keeping doctors in control of decisions. The company framed the effort as the next step in its medical AI research, following earlier systems such as MedPaLM, which focused on medical knowledge testing, and AMIE, which performed in text-based simulated consultations.
A key feature of AI co-clinician is its ability to process more than text. The system was tested with live audio and video, allowing it to observe physical cues such as gait, breathing patterns, and visible skin changes. In telemedical simulations, the model was able to guide patients through parts of a physical exam and assist with tasks such as checking inhaler technique or helping identify a shoulder injury. Those capabilities suggest that multimodal AI could eventually support remote consultations where visual and auditory observation matter.
DeepMind emphasized safety controls built into the system. AI co-clinician uses a dual-agent design in which a “Planner” continuously reviews the interaction and checks whether the “Talker” stays within clinical boundaries. The company stated this structure is meant to reduce unsafe outputs and improve reliability in medical settings, where factual accuracy and restraint are essential.
The research team evaluated the system in several ways. In one test, they adapted the NOHARM safety framework to measure both incorrect responses and failures to surface important information. In blind comparisons involving 98 primary care queries, the system recorded zero critical errors in 97 cases and was preferred over other evidence synthesis tools by physicians. DeepMind stated this suggests the model can be useful for clinicians seeking grounded, high-quality clinical information.
The study also examined how well the system handled medication-related questions using the OpenFDA RxQA benchmark, which is designed to test knowledge and reasoning about drugs and treatment. In open-ended evaluations, AI co-clinician outperformed other frontier models, indicating progress in an area that is especially important in day-to-day care planning.
In patient-facing simulations, however, human doctors still performed better overall. Working with academic physicians from Harvard and Stanford, the research team ran a randomized study involving 20 synthetic clinical scenarios and 10 physician patient-actors. Across more than 140 assessed areas, physicians outperformed the AI in detecting red flags and directing physical examinations, even though the system matched or exceeded physician performance in 68 categories, including triage. The findings suggest that the tool may be most valuable as a support system rather than a substitute for clinical expertise.
DeepMind stated the broader objective is to develop AI that can assist physicians in ways that are trustworthy, clinically grounded, and adaptable to real-world care environments. The company is continuing research collaborations across several countries, including the United States, India, Australia, New Zealand, Singapore, and the United Arab Emirates, as it works to test the system in more diverse healthcare settings.