GLENDALE, CA, May 21, 2019 – In a drive for standardization, the proposed U.S. Core Data for Interoperability1, mandated by the 2015 21st Century Cure Act2, calls for the identification and refinement of a basic set of clinical data to be required for all electronic health records. The American Medical Informatics Association, however, recommends that healthcare researchers and providers concentrate on sharing data as quickly as possible, deferring the creation of standards until sometime in the future.3 “This is a false choice and a distraction,” says James D’Arezzo, CEO of Condusiv Technologies. D’Arezzo, whose company is the world leader in I/O reduction and SQL database performance, adds, “What the healthcare industry really needs to focus on is enabling its heavily overburdened IT infrastructure to do the job it’s being asked to do.”
The industry’s preoccupation with interoperability and standardization, notes D’Arezzo, is perfectly understandable. Turf wars over proprietary interfaces and protocols are having a major impact on healthcare IT budgets.4 Noncompatible electronic health records contribute significantly to the fact that computerized recordkeeping consumes more than 50% of the average physician’s workday, which now stretches to more than 11 hours.5 Healthcare organizations struggling to process this tsunami of data are frustrated by the number and variety of analytics tools they are forced to use.6
Supporting all this activity, however-unnoticed and, says D’Arezzo, dangerously neglected-is the basic computational machinery itself. Data analytics requires a computer system to access multiple and often far-flung databases, pulling information together through millions of individual input-output (I/O) operations. The system’s analytic capability is dependent on the efficiency of those operations, which in turn is dependent on the efficiency of the computer’s operating environment. According to experts, the most widely used operating system, Microsoft Windows, is in many ways the least efficient. In any storage environment, from multi-cloud to a PC hard drive, Windows penalizes optimum performance due to server inefficiencies in the handoff of data to storage. This is a problem that, untreated, worsens with time. The average Windows-based system pays a 30% to 40% penalty in overall throughput capability due to I/O degradation.7
The good news, says D’Arezzo, is that I/O degradation is a software problem, one for which relatively inexpensive software solutions exist. Dealing with it does not require (and is not helped by) major investments in new computational and storage hardware. Condusiv Technologies, the world leader in this area, provides solutions that-without the need for additional investment in hardware-can, in the Windows environment, improve overall system throughput by 30% to 50% or more.
“Over the long term,” says D’Arezzo, “given the realities of the healthcare industry-and the lifesaving potential of current healthcare data collection and analytics-the current level of system inefficiency is insupportable. It’s as though you were being asked to double the volume of traffic flowing through New York City while keeping a third of the tollbooths and highway interchanges closed. In that context, the share-versus-standardize debate is like arguing about who has the right-of-way on 42nd Street. That will work itself out in time; right now, we need to get the roads unclogged.”
About Condusiv Technologies
Condusiv® Technologies is the world leader in software-only storage performance solutions for virtual and physical server environments, enabling systems to process more data in less time for faster application performance. Condusiv guarantees to solve the toughest application performance challenges with faster-than-new performance via V-locity® for virtual servers or Diskeeper® for physical servers and PCs. With over 100 million licenses sold, Condusiv solutions are used by 90% of the Fortune 1000 and almost three-quarters of the Forbes Global 100 to increase business productivity and reduce data center costs while extending the life of existing hardware. Condusiv Chief Executive Officer Jim D’Arezzo has had a long and distinguished career in high technology.
Condusiv was founded in 1981 by Craig Jensen as Executive Software. Jensen authored Diskeeper, which became the best-selling defragmentation software of all time. Over 37 years, he has taken the thought leadership in file system management and caching and transformed it into enterprise software.
1. “Draft U.S. Core Data for Interoperability (UCDI) and Proposed Expansion Process,” Office of the Coordinator for Health Information Technology, January 5, 2018.
2. “21st Century Cures Act,” U.S. Food and Drug Administration, January 2017.
3. Miliard, Mike, “Interoperability: Do we have the value proposition upside down?”, Healthcare IT News, Apri 10, 2019.
4. “Interoperability: Sharing patient information to coordinate care,” Elation Health.
5. Finnegan, Joanne, “Primary care doctors spend more than 50% of workday on EHR tasks, American Medical Association study finds,” Fierce Healthcare, September 13, 2017.
6. Donovan, Fred, “Healthcare Organizations Stymied by Health Data Analytics Tools,” HIT Infrastructure, March 14, 2019.
7. Morin, Brian, “Windows is still Windows Whether in the Cloud, on Hyperconverged or All-flash,” Condusiv blog post, June 5, 2018.
For more information, visit https://condusiv.com
Follow us on Twitter and Like Us on Facebook