For Hospitals On the Edge, HIT Is the Tipping Point

“No aspect of health IT entails as much uncertainty as the magnitude of its potential benefits.”

A few years into the Meaningful Use program, it seems this quote from a 2008 Congressional Budget Office report entitled “Evidence on the Costs and Benefits of Health Information Technology” may have been written with the assistance of a crystal ball.

Fast forward to 2013.

“Just from reading a week’s worth of news, it’s obvious that we don’t really know whether healthcare IT is better or worse off than before [Meaningful Use incentives],” popular blogger and health IT observer Mr. HIStalk wrote earlier this year.

So, perhaps RAND was hypnotized by Cerner funding when they created their rosy prognosis (hearken back, if you will, to 2005 and the projected $81 billion in annual healthcare savings). Maybe they were just plain wrong and the most recent RAND report stands as a tacit mea culpa.

Either way, we’re left with hypotheses that, while not incontrovertible, are gaining traction:

  1. Health IT benefits will manifest gradually over an extended timeframe.
  2. Those benefits will not quickly morph into reduced costs, if they ever do.
  3. Because of 1 and 2, investing in a hugely expensive electronic health record system is potentially risky.

How risky? Without question, massive health IT expense and the predominant proprietary IT model are threats to a hospital or health system’s financial viability, to its solvency.

We’re seeing some examples even now.

Michigan’s Henry Ford Health System recently reported a 15 percent decrease in net income as a result of uncompensated care and $36 million spent on a proprietary EHR system. According to health system CEO Nancy Schlichting, “We knew that 2012 and 2013 would not be easy years for the system because of the Epic costs.”

The Boston Globe reported that fully one-third of Massachusetts’ hospitals are on or over the financial edge. Profit margins for the state’s acute care hospitals have fallen to an average of 2.1 percent as a result of government regulation and reduced revenue from the continuing economic malaise. Health IT is certainly not the only culprit, but if EHR systems can’t pay for themselves, they are making a bad financial situation worse for some hospitals.

While it may seem that the financial problem created by expensive, proprietary health IT is simple and straightforward—health IT expenses push the budget into the red, doctors see fewer patients, revenue falls and creditors come calling—healthcare economics are unique and apparently beamed from some other dimension where up is down and black is white, so linear explanations don’t really hold.

As Steven Brill’s article in Time magazine (subscription required) on healthcare costs makes clear, there is no underlying logic to hospital fees and billing practices. The chargemaster rules and answers to no one, which enables some supposedly non-profit hospitals and health systems to earn profit margins of 20 to 30 percent, sometimes more.

This scenario cannot continue for a host of reasons that culminate in only one that truly matters: Healthcare costs regularly destroy the financial lives of individual Americans and threaten to swamp the entire economy. The chargemaster can’t continue to print money for an expensive EHR when the healthcare fabric is starting to weaken at the seams.

“The ROI-rationale for the widespread adoption of health IT in the U.S. is a macro, national one. It’s a public health calculation that’s been mired in commercial/private health arithmetic,” wrote health economist and management consultant Jane Sarasohn-Kahn in 2008. “The denominator of these wrong-headed ROI calculations has been wrong-chosen: it’s been the individual physician practice, or the hospital, or the single health plan. The denominator is the public’s health.”

Not much has changed since 2008 in how we view the potential ROI from health IT. Indeed, the public health argument for health IT is currently the most viable argument for continuing down the same path, which is just fine from a federal government perspective. But individual hospitals and health systems don’t often make macroeconomic or macro-public health decisions. They decide what’s best for that hospital / health system.

Which makes it perhaps unwise to pay $80, $150 or even $700 million dollars for a proprietary model EHR unless you think you can continue to extract 200, 300 and 400 percent margins from informationally unarmed patients and an increasingly less powerful insurance industry.

Brill’s revealing expose makes clear that some of the things health IT was supposed to eliminate (extensive and duplicate lab tests and CT scans, etc.), thereby reducing overall health system costs, are exactly what hospitals use to boost revenue.

So, what are the chances hospitals will look to increase efficiency by eliminating profitable procedures, especially when they have this shiny new $750 million EHR to pay for and may see a decrease in revenue because productivity and patient visit numbers are down?

For an answer, let’s go back to the 2008 CBO report:

“The use of information technology might lead to greater efficiency in delivering health care and to higher-quality services, but financial incentives could constrain many of those positive changes. For example EHRs could provide physicians with a useful tool for reducing the number of unnecessary or duplicated laboratory tests that they ordered, but the likelihood of such reductions could depend on factors such as whether physicians were compensated for controlling the use of laboratory testing (as in some managed care plans) or whether they derived income from ordering more tests.”

Just like that, healthcare as not a business (A mission? A calling? A social good? A human right?) runs smack into healthcare as a business, and keeping the lights on always wins out, as it must.

That said, I believe the chargemaster will be tamed because the system as is cannot stand.  Will we be able to change the incentives for healthcare providers to improve efficiencies and eliminate waste? Maybe and maybe not. It’s a huge task about which I’m not overly optimistic.

What happens if we can’t?

There will be hospital bankruptcies—the dissolution of assets to pay off debts—which, for many citizens, will create a gaping hole in the fabric of the healthcare service system. Not all states and regions will follow the path of Massachusetts, but many will.

With all the economic uncertainty associated with healthcare at this point in time, it is a truly risky investment to pay multi-millions of dollars for a proprietary health IT system that, let’s face it, may take years to generate any concomitant savings, if it ever happens.

Edmund Billings, MD, is chief medical officer of Medsphere Systems Corporation, the developer of the OpenVista electronic health record.

3 replies »

  1. Either investment in health IT eliminates “…(extensive and duplicate lab tests and CT scans, etc) thereby reducing overall health system costs…” or it does not.

    If the system properly implemented cannot pay for itself many times over then it is likely not an appropriate solution regardless of initial investment cost.

  2. To paraphrase an observation from McKinsey over a decade ago, regarding IT’s impact on productivity, “when tailored to sector-specific businesses, deployed in an appropriate sequence, and co-evolved with managerial innovation, the impact of information technology investments can be huge.” As I often tell my students, “technology by itself has no value . . . value is only found in what you do with it”. Simply purchasing and installing an expensive EMR doesn’t bring value; it is only if you change how you work and what you do to take advantage of an EMR, that will make any difference. In many (most?) cases, we have a ways to go in figuring this out.

  3. Au contraire!

    NOT investing in a electronic health record system is the beginning of the end.