This case study examines a class of AI risk that is already operational, externally generated, and materially ungoverned: decision-shaped AI output produced under correct facts.
The core finding is not that AI systems hallucinate, misstate evidence, or violate explicit rules. The finding is that they assemble accurate claims into authoritative, decision-ready narratives in regulated healthcare contexts, without accountability, auditability, or enforceable role boundaries.
This case study examines a class of AI risk that is already operational, externally generated, and materially ungoverned: decision-shaped AI output produced under correct facts.
The core finding is not that AI systems hallucinate, misstate evidence, or violate explicit rules. The finding is that they assemble accurate claims into authoritative, decision-ready narratives in regulated healthcare contexts, without accountability, auditability, or enforceable role boundaries.