Clinical AI Safety Review
Independent assessment of your AI tool's clinical safety before hospital deployment. We identify foreseeable failure modes, evaluate human-AI interaction risks, and produce governance-ready documentation.
We assess whether clinical AI tools are safe, credible, and ready for real-world deployment. Practising clinicians with published expertise in AI evaluation, human–AI interaction, and clinical decision-making.
Healthcare systems need defensible governance processes and independent clinical assessment of AI claims. But compliance alone doesn't answer the question that matters:
This isn't hypothetical. In our own research analysing 850 patient assessments at a major London teaching hospital, we found that among unplanned admissions, clinicians often recorded a single working diagnosis, and in the majority of these, no expression of uncertainty was documented. If clinicians already struggle to express uncertainty in routine practice, what happens when an AI system presents a confident recommendation?
Understanding these patterns – how clinicians actually think and document – is essential before deploying tools designed to influence their decisions.
It requires people who have practised at the bedside, published on AI evaluation methodology, and understand how human-AI interaction alters decision-making – a question we explored in depth in our book Thinking About Thinking.
We provide independent clinical-scientific assessment for vendors, providers, investors, and independent auditors making high-consequence decisions about healthcare AI.
Independent assessment of your AI tool's clinical safety before hospital deployment. We identify foreseeable failure modes, evaluate human-AI interaction risks, and produce governance-ready documentation.
Independent evaluation before you deploy an AI tool. We assess workflow integration risks, clinician interaction hazards, and produce the clinical safety case your governance board needs.
Independent clinical-scientific assessment of AI healthcare companies. We evaluate whether the clinical claims hold up, the evidence is robust, and the deployment risks are manageable.
We assess AI tools through the lens of clinicians making high-stakes decisions under time pressure in intensive care and cardiology — not former clinicians commenting from the sidelines.
Our work includes peer-reviewed research on AI evaluation, digital intervention design, clinician reasoning, and human–AI interaction, published in high impact journals including BMJ, Nature Medicine and npj Digital Medicine.
Between us, we combine critical care, cardiology, health economics, implementation thinking, and AI evaluation science. That matters because real deployment decisions are rarely only technical.
When we issue an opinion, we do so as practising doctors whose professional standing is attached to that judgment. That matters when the audience is a governance board, regulator, investor, or court.
Our book, Thinking About Thinking: A Prescription for Healthcare Improvement, examines how clinicians reason under uncertainty and how technologies – including AI – reshape those decisions. That analytical framework informs every assessment we deliver: understanding not just what a tool does, but how it changes the decisions clinicians make.
We operate with clear conflict-of-interest disclosure and do not assess products where there is a material financial interest. Independence is central to the value of our work.
Board-certified intensive care physician. PhD focused on AI evaluation methodology and human–AI interaction. Co-author of DECIDE-AI. 60+ publications and 5,000+ citations.
Dr Nagendran is a board-certified intensivist and clinical AI researcher. He completed his undergraduate medical degree at Emmanuel College, Cambridge and Green Templeton College, Oxford. His PhD at Imperial College London used eye-tracking and high-fidelity simulation to study how clinicians interact with AI recommendations under cognitive load – research that directly informs every safety assessment he delivers.
He is a co-author of the DECIDE-AI reporting guideline (Nature Medicine), which sets international standards for early-stage clinical AI evaluation, and his award-winning BMJ systematic review of deep learning diagnostic accuracy has been featured by Fortune and BBC World. He also peer-reviews research and protocols for journals and funders including BMJ, Nature Medicine and UKRI, and consults for public and private sector clients on clinical AI deployment.
He continues to practise intensive care medicine in a tertiary liver transplant ICU — the exact context in which high-stakes digital tools increasingly need defensible evaluation.
Cardiologist and internal medicine physician combining clinical expertise with health economic evaluation. Executive MSc from LSE. Principal investigator of the THIRST Alert RCT.
Dr Chen is a cardiologist, internal medicine physician and academic with expertise in Health Informatics and Health Economics. He completed his undergraduate medical degree at Gonville & Caius College, Cambridge and Green Templeton College, Oxford.
He has an MSc in Health Economics, management and cardiovascular outcomes from the LSE, where he specialised in the economic evaluation of healthcare technologies. He was Chief Registrar at St Bartholomew's Hospital, one of Europe’s largest cardiovascular centres of excellence, and co-created a management consulting fellowship for clinicians hosted by Deloitte UK.
He was also the principal investigator of the THIRST Alert randomised controlled trial at UCLH, which examined an EHR-embedded clinical decision tool for point-of-care trial recruitment and conduct. He has peer-reviewed research and grant proposals from institutions including the Swiss National Science Foundation, the BMJ and the European Heart Journal.
His combination of cardiology, health economics, operational leadership, and digital health trial methodology means he brings the payer and system perspective that investors and health systems need.
Chen Y, Nagendran M. Routledge, 2024. Our book on metacognition in clinical practice — why some clinicians make better decisions than others, how cognitive biases distort reasoning under pressure, and why this matters more than ever as AI enters clinical workflows.
Vasey B, Nagendran M et al. Nature Medicine, 2023. Co-authored guideline defining how early-stage AI-driven decision support systems should be evaluated in clinical settings. Adopted internationally.
Chen Y et al. Forthcoming. RCT testing an EHR-integrated decision prompt for clinical trial recruitment and conduct at UCLH — a pragmatic study of how clinicians respond to digital nudges embedded in their workflow.
Nagendran M, Chen Y et al. BMJ, 2020. Award-winning review showing that many claims of AI outperforming clinicians were based on weak comparative evidence. Top 5% Altmetric score; featured by Fortune and BBC World.
Nagendran M, Festor P et al. npj Digital Medicine, 2024. Eye-tracking study examining how physicians interact with explainable AI and the downstream consequences on their decision-making.
Festor P, Nagendran M et al. PLOS Digital Health, 2025. Physical simulation study directly measuring the consequences of AI-assisted prescribing in a high-stakes ICU environment.
A 30-minute introductory conversation to understand the product, deployment, diligence, or governance question you are facing.