Clinical AI Safety & Accountability

Independent expert judgment for AI and digital tools in healthcare

We assess whether clinical AI tools are safe, credible, and ready for real-world deployment. Practising clinicians with published expertise in AI evaluation, human–AI interaction, and clinical decision-making.

25+
Years of clinical experience
90+
Peer-reviewed publications
6,000+
Citations

AI and digital tools promise to transform clinical workflows into efficient, learning health systems

Healthcare systems need defensible governance processes and independent clinical assessment of AI claims. But compliance alone doesn't answer the question that matters:

“What happens when this tool is put in the hands of clinicians, under real-world uncertainty and pressure, with patients at the receiving end?”

This isn't hypothetical. In our own research analysing 850 patient assessments at a major London teaching hospital, we found that among unplanned admissions, clinicians often recorded a single working diagnosis, and in the majority of these, no expression of uncertainty was documented. If clinicians already struggle to express uncertainty in routine practice, what happens when an AI system presents a confident recommendation?

Understanding these patterns – how clinicians actually think and document – is essential before deploying tools designed to influence their decisions.

It requires people who have practised at the bedside, published on AI evaluation methodology, and understand how human-AI interaction alters decision-making – a question we explored in depth in our book Thinking About Thinking.

Three services. One question: will it be widely and durably adopted?

We provide independent clinical-scientific assessment for vendors, providers, investors, and independent auditors making high-consequence decisions about healthcare AI.

For AI vendors

Clinical AI Safety Review

Independent assessment of your AI tool's clinical safety before hospital deployment. We identify foreseeable failure modes, evaluate human-AI interaction risks, and produce governance-ready documentation.

You receive Clinical red-team assessment · Foreseeable hazard analysis · Severity/likelihood ratings with mitigation recommendations · Monitoring trigger specification · Governance-ready summary (DCB0129/0160 aligned)
Discuss your product →
For healthcare systems and providers

Risk, Adoption and Scale

Independent evaluation before you deploy an AI tool. We assess workflow integration risks, clinician interaction hazards, and produce the clinical safety case your governance board needs.

You receive Workflow integration risk assessment · Decision impact analysis · Alert burden and trust calibration evaluation · Clinical safety case support for governance sign-off
Discuss your deployment →
For Investors and independent auditors

Clinical AI Due Diligence

Independent clinical-scientific assessment of AI healthcare companies. We evaluate whether the clinical claims hold up, the evidence is robust, and the deployment risks are manageable.

You receive Clinical evidence appraisal · Technical architecture risk assessment · Deployment feasibility evaluation · Liability and accountability mapping · Expert witness availability
Discuss your case →

A unique blend of academic, digital, regulatory, and strategic insights, pooled together in two practising clinicians.

We practise medicine

We assess AI tools through the lens of clinicians making high-stakes decisions under time pressure in intensive care and cardiology — not former clinicians commenting from the sidelines.

We study how clinical decisions actually work

Our work includes peer-reviewed research on AI evaluation, digital intervention design, clinician reasoning, and human–AI interaction, published in high impact journals including BMJ, Nature Medicine and npj Digital Medicine.

We bring more than one lens

Between us, we combine critical care, cardiology, health economics, implementation thinking, and AI evaluation science. That matters because real deployment decisions are rarely only technical.

We sign our names

When we issue an opinion, we do so as practising doctors whose professional standing is attached to that judgment. That matters when the audience is a governance board, regulator, investor, or court.

We understand the reasoning layer

Our book, Thinking About Thinking: A Prescription for Healthcare Improvement, examines how clinicians reason under uncertainty and how technologies – including AI – reshape those decisions. That analytical framework informs every assessment we deliver: understanding not just what a tool does, but how it changes the decisions clinicians make.

We manage conflicts transparently

We operate with clear conflict-of-interest disclosure and do not assess products where there is a material financial interest. Independence is central to the value of our work.

Two clinician-scientists. One focus.

Dr Myura Nagendran
Dr Myura Nagendran
Intensivist · PhD Clinical AI

Board-certified intensive care physician. PhD focused on AI evaluation methodology and human–AI interaction. Co-author of DECIDE-AI. 60+ publications and 5,000+ citations.

Dr Nagendran is a board-certified intensivist and clinical AI researcher. He completed his undergraduate medical degree at Emmanuel College, Cambridge and Green Templeton College, Oxford. His PhD at Imperial College London used eye-tracking and high-fidelity simulation to study how clinicians interact with AI recommendations under cognitive load – research that directly informs every safety assessment he delivers.

He is a co-author of the DECIDE-AI reporting guideline (Nature Medicine), which sets international standards for early-stage clinical AI evaluation, and his award-winning BMJ systematic review of deep learning diagnostic accuracy has been featured by Fortune and BBC World. He also peer-reviews research and protocols for journals and funders including BMJ, Nature Medicine and UKRI, and consults for public and private sector clients on clinical AI deployment.

He continues to practise intensive care medicine in a tertiary liver transplant ICU — the exact context in which high-stakes digital tools increasingly need defensible evaluation.

Dr Yang Chen
Dr Yang Chen
Cardiologist · Executive MSc Health Economics

Cardiologist and internal medicine physician combining clinical expertise with health economic evaluation. Executive MSc from LSE. Principal investigator of the THIRST Alert RCT.

Dr Chen is a cardiologist, internal medicine physician and academic with expertise in Health Informatics and Health Economics. He completed his undergraduate medical degree at Gonville & Caius College, Cambridge and Green Templeton College, Oxford.

He has an MSc in Health Economics, management and cardiovascular outcomes from the LSE, where he specialised in the economic evaluation of healthcare technologies. He was Chief Registrar at St Bartholomew's Hospital, one of Europe’s largest cardiovascular centres of excellence, and co-created a management consulting fellowship for clinicians hosted by Deloitte UK.

He was also the principal investigator of the THIRST Alert randomised controlled trial at UCLH, which examined an EHR-embedded clinical decision tool for point-of-care trial recruitment and conduct. He has peer-reviewed research and grant proposals from institutions including the Swiss National Science Foundation, the BMJ and the European Heart Journal.

His combination of cardiology, health economics, operational leadership, and digital health trial methodology means he brings the payer and system perspective that investors and health systems need.

Examples of the research and writing behind the way we assess healthcare AI.

Thinking About Thinking: A Prescription for Healthcare Improvement

Chen Y, Nagendran M. Routledge, 2024. Our book on metacognition in clinical practice — why some clinicians make better decisions than others, how cognitive biases distort reasoning under pressure, and why this matters more than ever as AI enters clinical workflows.

DECIDE-AI: international reporting guideline for clinical AI evaluation

Vasey B, Nagendran M et al. Nature Medicine, 2023. Co-authored guideline defining how early-stage AI-driven decision support systems should be evaluated in clinical settings. Adopted internationally.

THIRST Alert: EHR-embedded clinical decision support RCT

Chen Y et al. Forthcoming. RCT testing an EHR-integrated decision prompt for clinical trial recruitment and conduct at UCLH — a pragmatic study of how clinicians respond to digital nudges embedded in their workflow.

Deep learning vs clinicians: a systematic review

Nagendran M, Chen Y et al. BMJ, 2020. Award-winning review showing that many claims of AI outperforming clinicians were based on weak comparative evidence. Top 5% Altmetric score; featured by Fortune and BBC World.

Eye tracking and explainable AI recommendations in intensive care

Nagendran M, Festor P et al. npj Digital Medicine, 2024. Eye-tracking study examining how physicians interact with explainable AI and the downstream consequences on their decision-making.

Safety of human–AI cooperative decision-making in intensive care

Festor P, Nagendran M et al. PLOS Digital Health, 2025. Physical simulation study directly measuring the consequences of AI-assisted prescribing in a high-stakes ICU environment.

Book an introductory call.

A 30-minute introductory conversation to understand the product, deployment, diligence, or governance question you are facing.

For vendors: early external scrutiny before a buyer, pilot, or governance process asks harder questions.
For providers: independent judgment before deployment, procurement, or board-level approval.
For investors and legal teams: clinically grounded evaluation of claims, evidence, risk, and accountability.