As well being care payers more and more depend on synthetic intelligence (“AI”) to hurry up affected person declare adjudication and prior-authorization determinations, suppliers needs to be looking out for algorithms designed to disclaim claims with minimal
human oversight. Two putative class motion lawsuits filed in opposition to Cigna and UnitedHealth in November 2023 allege that the payers’ respective AI fashions, “PxDx” and “nH Predict,” had been used rather than actual medical professionals
to wrongfully deny medically mandatory care.
The lawsuit in opposition to Cigna alleges that Cigna makes use of its PxDx (procedure-to-diagnosis) algorithm to routinely evaluation claims with out conducting a radical medical evaluation of the claims as required by legislation and contract. [1] The lawsuit alleges that over a interval of two months in 2022, for instance, Cigna used PxDx to routinely evaluation and deny over 300,000 claims, spending a mean of simply 1.2 second per declare.
[2] Regardless of insurance coverage legal guidelines and rules in lots of states requiring physicians to evaluation claims earlier than payers might deny them on the premise of medical necessity, the
lawsuit means that physicians employed by payers to evaluation insurance coverage claims (usually known as “medical administrators”) are signing off on denials with out even them.
The go well with in opposition to UnitedHealthcare is restricted to Medicare Benefit plans. It alleges that UnitedHealth’s AI mannequin, nH Predict, has a identified 90% error price and denies claims by forcing unrealistic expectations of Medicare Benefit sufferers’
restoration in post-acute care settings. The algorithm overrides the opinions of actual life treating physicians who deemed the underlying care was medically mandatory.[3]
Each lawsuits depend on solely a small share of sufferers interesting wrongfully denied claims. Certainly, the place solely about 0.2% of policyholders enchantment denied claims, the bulk will find yourself paying out-of-pocket prices or forgoing the rest of prescribed
post-acute care. [4] In Medicare Benefit instances particularly, this will imply the aged are pressured out of care services or should deplete their financial savings to proceed
receiving medically mandatory care the place an algorithm has disagreed with their treating doctor.
Not solely are these algorithms getting used to disclaim claims, in addition they are getting used to alter human conduct. The UnitedHealthcare go well with alleges that the defendants have arrange worker targets as a way to preserve expert nursing facility stays inside 1% of
the times projected by nH Predict, and that “[e]mployees who deviate from the nH Predict AI Mannequin projections are disciplined and terminated, no matter whether or not a affected person requires extra care.” [5]
Whereas this conduct is regarding, sufferers and suppliers alike have recourse by means of interesting claims and by proactively looking for to barter contracts with payers to ban these practices. Suppliers can also have already got agreements in place that
enable for suppliers to hunt recourse by means of litigation in opposition to the payer for such use of AI. Thompson Coburn’s Well being Care Group has a workforce of attorneys out there to help with these points.
[1] See Amy Snyder, et al. v. The Cigna Group, Cigna Well being and Life Ins. Co., and Cigna Well being Mgmt., Inc., 3:23-cv-01451-OAW (D. Conn. Nov. 2, 2023); see additionally Patrick Rucker, et al.,
How Cigna Saves Thousands and thousands by Having Its Docs Reject Claims With out Studying Them,
ProPublica (Mar. 25, 2023),
[3] See Property of Gene B. Lokken, et al. v. UnitedHealth Group, Inc., et al.,
0:23-cv-03514 (D. Minn. Nov. 14, 2023).
[5]See Property of Gene B. Lokken et al. v. UnitedHealth Group, Inc., et al.,
0:23-cv-03514 (D. Minn. Nov. 14, 2023).