Why this hasn't landed yet
The finding is framed as a methods paper, not a scandal. No named hospital, no named product, no patient harmed on record. The word 'simulator' makes it sound like a precaution rather than a proof. The story requires two steps of inference to become alarming, and most coverage stops at one.
What happens next
Regulators and hospital procurement offices now have a working tool, not just a policy argument. The next move is whether the FDA or CMS folds something like this into AI device approval criteria — the pressure is already forming, given ECRI named AI the top health technology hazard for 2025 and the Federal AI Risk Management Act is pending.
The catch
AI developers whose tools fail this audit will note that the simulator was validated on a single decision aid for antidepressant selection and argue their use case is different, which is the same argument made after the Obermeyer 2019 algorithm bias finding and which bought several more years of unreformed deployment.