An assessment of health care information and management systems society and leapfrog data on computerized provider order entry.
Most national studies examining CPOE systems have used data from one of two sources: Health Care Information and Management Systems Society (HIMSS) Analytics (see, e.g., Teufel, Kazley, and Basco 2009) or the Leapfrog Group (see, e.g., Hillman and Givens 2005). While the use of these national datasets has provided potentially valuable information that complements the single-institutional studies that dominate the literature, there has been no empirical assessment of either the HIMSS Analytics or the Leapfrog data. Thus, it is unclear whether these datasets are appropriate for rigorous health services research endeavors designed to influence either health policy or the adoption of a resource-intensive CPOE system by health organizations.
The purpose of this paper is to assess the internal consistency, level of agreement, and responsiveness of the HIMSS and Leapfrog measures of CPOE. In so doing, we will provide information regarding the appropriateness of these national datasets for the purposes of health service research. Given the intensifying interest in generalizable studies of CPOE systems in hospitals, we expect our results to be of interest to researchers, policy makers, and hospital leaders alike.
The internal consistency, level of agreement, and responsiveness of CPOE measures are important components of high-quality and generalizable research regarding CPOE use and adoption. These characteristics are similar to the well-known concepts of reliability and validity, which are difficult to assess in this case due to the lack of a gold standard measure. In our study, we focus on internal consistency, which, like reliability, refers to the stability of a measure and the degree to which the measure is free from random error (Ray et al. 2006). Level of agreement attempts to address issues of validity in the absence of a gold standard. It is the notion that a scale actually measures what it intends to measure. Responsiveness is the ability of the scale to detect changes over time (Ray et al. 2006). Each of these characteristics influences the extent to which a given measure can represent the practices or constructs under study. The following is a more in-depth overview of each of these elements.
Internal consistency is the reproducibility of a measure. A measure with good internal consistency can be replicated and avoids random errors through a smaller range of fluctuation from one measure to the next (Cherulnik 2001). Accurate consistency of a measure may be indicative of reliability if the reported measure has not changed in practice. Internal consistency can be influenced by the source reporting the measure, untruthful reporting, the consistent description of the construct, the method of data collection, and response bias.
The level of agreement is closely related to construct validity, which is the degree to which a measure provides trustworthy information about the construct it intends to measure (Cherulnik 2001; Grembowski 2001). A measure that has a high level of agreement with a gold standard measure is believed to be free of systematic error or bias, and thus valid. Validity can be influenced by a definition, individual perception of a construct, response bias, untruthful reporting, or a vague description of the construct to be measured. In instances where a gold standard does not exist (e.g., national data on CPOE), measuring the level of agreement between two different surveys is the first step in the empirical assessment of those surveys.
Lastly, responsiveness describes the sensitivity of a measure and its ability to accurately detect small but important changes in a construct over time (Grembowski 2001). Measuring responsiveness has become increasingly important because detecting small changes in management or health outcomes is needed in health services research. It has been claimed that "Responsiveness should join reliability and validity as necessary requirements for instruments designed primarily to measure change over time" (Guyatt, Walter, and Norman 1987).
The internal consistency, level of agreement, and responsiveness of CPOE measures are of great importance to researchers. Because there is national interest in increasing CPOE adoption and use, some researchers have focused on using national measures to study its impact. In doing so, they rely upon the established national data, which has not been empirically evaluated. If a CPOE measure fails to show internal consistency, it may bias the results by over- or underestimating CPOE adoption and use in a national sample. Such practice would make inference between CPOE and performance tenuous.
Likewise, if national measures of CPOE do not agree, it is possible that one or both measures are not valid. Potential reasons for disagreement may be that the two surveys are measuring something different than intended, thus biasing the results of any study using the measure. Because researchers have struggled to define and measure CPOE, it is conceivable that the measures may be flawed or may not distinguish between full CPOE use and early CPOE adoption. Other scenarios, such as considering an organization to have CPOE without a clinical decision support system as one that is using CPOE, are also possible. The challenge lies in the fact that CPOE is not well defined on a national level and may vary based on the system used and the level of adoption within the organization.
Lastly, a CPOE measure without good responsiveness would bias results of a study by failing to detect changes in CPOE from one year to the next. Because such variance may exist, researchers may ask whether changes in the reported CPOE measure for an organization are representative of current CPOE practice within that organization. Each of these criteria--internal consistency, agreement, and responsiveness--gives value to the measures themselves by ensuring they are truly measures of the construct of interest. Understanding these criteria in the context of national CPOE data will help researchers make better-informed choices about the use of such data sources.
Data sources for this study are the HIMSS Analytics data and the Leapfrog Group data. Participation in each survey is voluntary, and each responding organization chooses the individuals that complete the survey. The HIMSS Analytics data contain a wide range of information on integrated delivery systems, including detailed information on the health information technology environment. These data are collected annually at the system level, but they also contain information on a large number of individual hospitals and other health care organizations. We used HIMSS Analytics data from 2005 to 2007. We did not use data earlier than 2005 because HIMSS made a significant change in the definition of CPOE in 2004. We used the responses from two questions from the HIMSS survey; one on the presence of clinical decision support, and the other on the presence of CPOE. Responses on each question were considered positive if the application status was "Live and Operational," and negative for any other status.
The Leapfrog data contains information collected from all participating hospitals in one dataset from 2003 to 2007. Hospitals voluntarily report information annually about hospital practices (including CPOE use), performance, and outcomes to the Leapfrog Group. The Leapfrog Group allows for voluntary participation from one year to the next. Overall, the Leapfrog Group aims to use large employer purchasing power to improve health care safety, quality, and affordability through public reporting. We used Leapfrog data from 2005 through 2007 to correspond with the HIMSS data. The Leapfrog survey uses several criteria to determine whether the hospital-wide CPOE implementation in responding hospitals meets the standard they seek. These criteria include the reported percentage of prescribers using CPOE and the presence of decision support functionality. Responses to the survey were considered positive if the implementation "fully meets the standard" and negative for any other status.
We used the American Hospital Association Annual Survey of Hospitals data from 2005 to 2007 to match hospitals in each dataset and to obtain hospital characteristics for comparison. In addition to variables that describe hospital characteristics obtained from the American Hospital Association data, we classified hospitals' geographic location using the Rural-Urban Commuting Area Codes (Morrill, Cromartie, and Hart 1999). Rural--Urban Commuting Area codes, based on hospital zip codes, take into consideration commuting patterns and as such represent the state of the art in measuring rurality (Rural Health Research Center 2009).
Measuring CPOE presents several challenges. One of the main challenges is the nonstandardized definition of CPOE. For example, these two surveys use different definitions of CPOE, and HIMSS changed its definition of CPOE between the 2004 and 2005 surveys. A second challenge is defining CPOE adoption and use. There are differences between the presence of a CPOE system, what capabilities it includes, and how often or extensively clinicians use it. We do not attempt to measure CPOE adoption and use, but only whether the hospital reports that it has a hospital-wide CPOE system according to the definitions of each survey. As noted below, we calibrated the available information from both surveys to be as equivalent as possible. This approach reduces the likelihood that differences in adoption and utilization among hospitals will affect our results.
In order to allow for a fair comparison, we defined full CPOE status conservatively and dichotomously as the most complete status of CPOE reported in each of the two surveys. For the Leapfrog data, this corresponds to a survey response by the hospital that it fully meets the Leapfrog CPOE standard. Leapfrog includes in its definition of CPOE the requirement that there be some decision support capabilities in the system to reduce or prevent medication errors. The HIMSS definition of CPOE does not include this requirement, so in order to make the two measures comparable, we constructed a measure of CPOE in the HIMSS data that includes the presence of clinical decision support in the hospital. Therefore, our final measure of the presence of CPOE in the HIMSS data required the presence of both CPOE and clinical decision support with a live and operational status. Furthermore, we only consider hospitals with known CPOE status. Thus, we excluded hospitals with missing data for CPOE questions.
Samples and Analytical Approach
Given the concepts of internal consistency, level of agreement, and responsiveness, we used different samples and approaches to investigate each concept. Internal consistency is often measured through the stability of a measure from one point of time until another, as in the case of the consistency of items on a survey. To test for data internal consistency, we took advantage of the fact that when a hospital achieves "complete" CPOE status it is highly unlikely that in subsequent years their status would regress back to less than complete status. Thus, we examined the percentage of hospitals, in a given dataset (HIMSS or Leapfrog) that had CPOE in a given year and reported CPOE in the subsequent year. To do so, the HIMSS and Leapfrog datasets were each examined separately.
The concept of "level of agreement" in empirical assessments quantifies the degree to which differing measures of the same construct agree with each other. To test the level of agreement of each dataset, we restricted the sample of hospitals to organizations that responded to both the HIMSS and Leapfrog survey in any given year. For each year, we calculated the [kappa] statistic and McNemar's test. The [kappa] statistic provides a measure of the degree of association between two categorical variables, and McNemar's test is used to compare two proportions from dependent or matched samples (Rosner 2006).
We also calculated the [kappa] statistic and McNemar's test on a restricted sample containing hospitals that had reported CPOE status across all 3 years as a sensitivity analysis. We did this because no gold standard data source on CPOE exists. Specifically, we wanted to be as sure as possible that we "correctly" classified hospitals for CPOE status. Thus, we constructed a measure that used hospitals that responded as having CPOE in all 3 years as positive for CPOE status and negative otherwise for each survey. This approach maximized our ability to examine the level of agreement by utilizing hospitals whose responses over time strongly suggested they were consistently classified as having CPOE in the Leapfrog and HIMSS data, respectively.
Responsiveness is the ability of a measure to detect changes over time. A responsive measure will be consistent and accurate despite changes in the response itself. To test for data responsiveness, we examined the percent increase in complete CPOE status from 1 year to the next in HIMSS and Leapfrog data, respectively. We examined this calculation in the individual HIMSS and Leapfrog datasets as well as in the restricted dataset that represented organizations that responded to both surveys in any given year.
Lastly, we used Stata version 11 for all data management, including merging of data, sample identification, and calculation of descriptive statistics.
We present the number of hospitals with known CPOE status from the HIMSS, Leapfrog, and combined dataset by year in Table 1. Briefly, in 2007, the HIMSS data included CPOE information on 4,679 organizations. During the same year, the Leapfrog dataset included CPOE information on 2,619 hospitals. The number of hospitals with known CPOE status in both the HIMSS and Leapfrog datasets was 1,053 in 2005 and 2,355 in 2007.
The 2007 organizational characteristics of respondents in the HIMSS, Leapfrog, and the overlapping set are presented in Table 2. In most cases, respondents to the Leapfrog survey also were represented in the HIMSS dataset. As such, the hospitals presented in the combined dataset have identical characteristics to the Leapfrog set in many categories. Overall, the Leapfrog dataset included a lower proportion of small hospitals < 125 beds. All other organizational characteristics are generally similar.
In the HIMSS dataset, 238 hospitals reported complete CPOE status in 2005, of which 180 (76 percent) reported complete CPOE status in the HIMSS 2006 data; and 148 (62 percent) reported complete CPOE status in the HIMSS 2007 data (see Table 3). In 2006, 276 hospitals reported complete CPOE status in the HIMSS data of which 216 (78 percent) reported the same status in 2007.
Leapfrog data reported 69 hospitals with complete CPOE status in 2005. In the subsequent year (2006), 59 of these hospitals (86 percent) reported complete CPOE status. The same 59 hospitals (86 percent) reported complete CPOE status in 2007 as well. In 2006, 95 hospitals indicated complete CPOE status on the Leapfrog survey, of which 77 (81 percent) reported the same status in 2007.
Level of Agreement
Measures of agreement are reported in Table 4. There were 1,053 hospitals in both datasets that reported on CPOE in 2005. Of these, HIMSS and Leapfrog agreed on 15 hospitals as having CPOE and 902 hospitals as not having CPOE. For 2005, [kappa] = 0.14, p<.001; and McNemar's [chi square] = 79.53, p<.001. These indicate that there is significant disagreement in the likelihood of CPOE classification between the Leapfrog and HIMSS surveys.
In 2006, there were 1,823 hospitals in both datasets that reported on CPOE. Of these, HIMSS and Leapfrog agreed on 48 hospitals as having CPOE and on 1,549 hospitals as not having CPOE. For 2006, [kappa] = 0.25, p< .001; and McNemar's [chi square] = 110.46, p< .001. Similarly, there is significant disagreement between Leapfrog and HIMSS on CPOE status.
In 2007, there were 2,355 hospitals in both datasets that reported on CPOE. Of these, HIMSS and Leapfrog agreed on 89 hospitals as having CPOE and on 1,869 hospitals as not having CPOE. For 2007, [kappa] = 0.25, p< .001; and McNemar's [chi square] = 269.34, p< .001. Again, there is significant disagreement in the identification of CPOE status between the two data sources.
In the sensitivity analysis using the sample restricted to hospitals the reported having CPOE in all 3 years for each survey, there were 1,346 hospitals. Of these, HIMSS and Leapfrog agreed on 7 hospitals as having CPOE and on 1,238 hospitals as not having CPOE. For the sample in this sensitivity analysis, [kappa] = 0.10, p < .001; and McNemar's [chi square] = 47.14, p <.001; suggesting again that there is significant disagreement between the two datasets with respect to CPOE status.
When considering the 2005 HIMSS full sample only, 2,045 hospitals reported not having CPOE, of which 97 (4.7 percent) reported complete CPOE status in 2006 (see Table 5). In 2006, 3,208 hospitals in the HIMSS data reported not having CPOE of which 312 (9.7 percent) reported full CPOE status in 2007.
When considering the 2005 Leapfrog full sample only, 1,991 hospitals did not have CPOE, of which 36 (1.8 percent) reported full CPOE status in 2006. In 2006, 2,253 hospitals in the Leapfrog data reported not having CPOE, of which 18 (0.8 percent) changed their status to full CPOE in 2007.
We conducted the same analyses on the overlapping sample of hospitals that report on CPOE in all 3 years (see Table 5), but we did not use the restricted measure of CPOE status that we used in the level of agreement sensitivity analysis. From 2005 to 2006, the HIMSS data reported that 5.4 percent of hospitals switched their status from not having CPOE to complete CPOE implementation. From 2006 to 2007, the HIMSS data further found that 12.5 percent of hospitals switched their status from no CPOE to complete CPOE implementation. During the same period, the Leapfrog data (examining the same subsample of hospitals) reported a 2005-2006 increase in CPOE adopters of 1.7 percent; and a 2006-2007 increase in CPOE adopters of 3 percent.
Many national studies that examine CPOE use in hospitals have relied on either HIMSS or Leapfrog data. These datasets have been used to estimate hospital CPOE adoption rates (Ford et al. 2008), examine hospital characteristics associated with adoption (Teufel, Kazley, and Basco 2009), and correlate CPOE adoption with medication-related quality outcomes (Jha et al. 2008; Yu et al. 2009b; Kazely and Diana 2011). No previous study has empirically examined the HIMSS or Leapfrog datasets, despite the fact that findings from previous studies can influence managerial and policy decisions. In this study, we sought to examine the internal consistency, level of agreement, and responsiveness of the CPOE variables from these datasets.
Overall, respondents to both the Leapfrog and HIMSS datasets have similar organizational characteristics and in many cases overlap. However, our findings indicate minimal agreement between the two datasets regarding which hospitals have adopted CPOE, but adequate consistency within a given dataset from year to year. Both surveys tend to inaccurately estimate overall changes in true hospital CPOE status over time (i.e., responsiveness), but in different ways. Compared with the Leapfrog data, the HIMSS data tend to overestimate increases in adoption over time. Likewise, relative to the HIMSS data, the Leapfrog data have more downward trending estimates for year over year increases in CPOE status. Our findings suggest serious limitations on the use of either dataset for many health services research purposes. However, given the properties of each survey instrument, it would appear that each data source has differing strengths.
The [kappa] statistic compares the expected level of agreement that would be present by chance to the level of agreement present in the data. A [kappa] of 0 indicates agreement levels expected purely by chance, and a [kappa] of 1 indicates perfect agreement. The [kappa] statistics for all years is high enough to reject the hypothesis that the level of agreement between these two surveys is occurring only by chance. However, the [kappa] statistics are consistently low enough to indicate there is only marginal agreement between the two surveys. Rosner (2006) suggests that a [kappa] < 0.4 indicates marginal reproducibility, and the t, statistics indentified in our study range from 0.1 to 0.3. Further, based on our sensitivity analysis, it appears that the level of agreement may not be improving over time. McNemar's test was used to test the probability that the two surveys were equivalent in the proportions of hospitals identified as having CPOE. In each individual year and in the restricted sample, McNemar's test leads us to reject that hypothesis and conclude that the two surveys are indentifying different proportions of hospitals that have CPOE.
The HIMSS dataset shows consistently higher estimates of hospitals with CPOE than does the Leapfrog dataset, while the two datasets are much closer in estimates of those without CPOE. Our analysis is not able to identify the cause of this difference, but it does suggest that either HIMSS data overstate or Leapfrog data understate the true number of hospitals with CPOE. The Leapfrog criteria for reporting fully implemented CPOE is more restrictive than the HIMSS criteria, which is one likely source of the disagreement. Without a consensus definition of what constitutes full CPOE implementation, it may be impossible to determine which is more accurate. However, it seems reasonable to suggest that researchers wishing to avoid the possibility of having high false positives (regarding CPOE status) may be better served utilizing the Leapfrog data. Conversely, if a given researcher's concern is potential false negatives, the HIMSS data may be more appropriate to use in that instance. These qualities of the two surveys offer the researcher the opportunity to conduct sensitivity analyses using both datasets if such an approach is relevant to their study.
Once a hospital reported full CPOE status, the Leapfrog data and the HIMSS data (less so) were able to reidentify the same hospital as reporting full CPOE status in a subsequent year with adequate consistency. The Leapfrog consistency rate (81-86 percent) and the HIMSS consistency rate (62-78 percent) are either within, or approaching, the generally expectable rate of 80-90 percent (Hennekens and Buring 1987). Because CPOE adoption is a large undertaking requiring significant financial and human resources to implement, one would expect that sudden downgrades from the status of full CPOE would not take place regularly. Nevertheless, a number of hospitals from each dataset reported less than full CPOE status in an immediately subsequent year. This may be the result of legitimate cases where an implementation failure occurs and full CPOE status regresses (see, e.g., the experience of Cedar Sinai Medical Center; Connolly 2005); or it may reflect variability in how the question is perceived from year to year by a given individual or his/her successor who fills out the survey.
The differences noted in responses by the same organization to an almost identical question (full CPOE status) on the HIMSS and Leapfrog surveys may have something to do with how, and why, these data are collected. First, as noted above, the mission of the Leapfrog and HIMSS organizations are very different. Thus, certain responses on the Leapfrog survey may result in benefits to the organization (e.g., being able to participate in providing care for Leapfrog employer members' health plan beneficiaries) while no such direct incentives exist for specific responses on the HIMSS survey. This difference may result in less exaggeration of CPOE status on the Leapfrog dataset because organizations would be subject to the loss of benefits, sanctions, or public embarrassment in the event of a Leapfrog audit. Similarly, because Leapfrog offers hospitals partial "credit" for different stages of CPOE implementation, hospitals may be further dissuaded from overestimating their current CPOE status, and they may find benefit from more slowly reporting progress. In contrast, fewer forces may influence the accuracy of responses on the HIMSS survey.
Second, differences in responses by the same hospital to the Leapfrog and HIMSS surveys may be a function of market forces. In regions where competition among Leapfrog-participating hospitals is relatively high, responding hospitals may be more prone to err on the side of "meeting the CPOE standard" and thus garner access to profitable patient populations. Hospitals in less competitive markets may not have such pressures to assure their financial well-being. Lastly, while both surveys are voluntary, it is not known who ultimately in the organization provides the response to either survey or whether this person changes from one year to the next.
Our analysis has several limitations worth noting. First, we only utilized 3 years of HIMSS and Leapfrog data. Most results either showed obvious yearly trends (in one direction or the other) or were relatively consistent from year to year. Nevertheless, we recognize that using only 3 years worth of data may be a limiting factor. Second, we examined only variables related to CPOE for each of the datasets. Thus, our findings may not be applicable to other variables collected in either the HIMSS or Leapfrog data. Future studies should address the validity of other variables in these datasets. Third, some aspects of our analysis required that we compare only those hospitals that participated in both surveys in a given year. The loss of some hospitals that report to one survey but not the other could biases our findings. Lastly, we carefully calibrated the questions regarding CPOE status from both surveys so that we examined "full CPOE" in both datasets. Despite our best efforts, we recognize that the two surveys ultimately word their questions differently, which may explain some of the differences in responses we identified.
We recommend that future research in this area examine two key issues. First, we need a clear definition of what constitutes a CPOE system. It may seem obvious, but without a consensus on this question, validating the presence of CPOE will remain highly problematic. Second, a validation of any measure of CPOE--or other health information technologies such as electronic medical records--will require comparison of survey responses with the direct assessment of the presence of CPOE through standard survey validation approaches. Without a gold standard of CPOE use, it is not possible to fully validate the data.
In conclusion, the disagreement between the HIMSS and Leapfrog datasets regarding hospital CPOE status creates a challenge for researchers, practitioners, and policy makers who wish to understand CPOE from a national perspective. Future research is encouraged to overcome limitations in each data source. Currently, notwithstanding our analyses, the HIMSS and Leapfrog surveys remain the most widely available secondary sources of information on CPOE. Importantly, the AHA has recently begun to collect data on CPOE among hospitals in their newly developed Hospital EHR Adoption Database, a supplemental to the AHA Annual Survey (Jha et al. 2009). While this dataset promises to provide additional valuable information about CPOE (and other health information technology), to our knowledge, it has not undergone a similar assessment as described herein. Both the Leapfrog and HIMSS datasets, as well as the new AHA data, will continue to be among the scant choices for health services researchers interested in this topic for the foreseeable future. In the absence of a true gold standard measure for CPOE, it is our hope that all organizations collecting data used by health services researchers consider strategies to validate their data collection efforts.
Joint Acknowledgment/Disclosure Statement. There is no form of financial or material support, other contributors, or disclosures. The authors wish to thank HIMSS Analytics and The Leapfrog Group for the use of their data.
Bates, D. W., and A. A. Gawande. 2003. "Improving Safety with Information Technology." New England Journal of medicine 348 (25): 2526-34.
Bates, D. W., L. L. Leape, D.J. Cullen, N. Laird, L. A. Peterson, J. Teich, E. Burdick, M. Hickey, S. Kleefield, B. Shea, M. V. Vliet, and D. L. Seger. 1998. "Effect of Computerized Physician Order Entry and a Team Intervention on Prevention of Serious Medication Errors." The Journal of the American Medical Association 280 (15): 1311-6.
Bates, D. W., J. M. Teich, J. Lee, D. Seger, G.J. Kuperman, N. Ma'Luf, D. Boyle, and L. L. Leape. 1999. "The Impact of Computerized Physician Order Entry on Medication Error Prevention." Journal of the American Medical Informatics Association 6 (4): 313-21.
Bobb, A., K. Gleason, M. Husch, J. Feinglass, P. R. Yarnold, and G. A. Noskin. 2004. "The Epidemiology of Prescribing Errors: The Potential Impact of Computerized Prescriber Order Entry." Archives of Internal Medicine 164 (7): 785-92.
Chaudhry, B.,J. Wang, S. Wu, M. Maglione, W. Mojica, E. Roth, S. C. Morton, and P. G. Shekelle. 2006. "Systematic Review: Impact of Health Information Technology on Quality, Efficiency, and Costs of Medical Care." Annals of Internal Medicine 144 (10): 742-52.
Cherulnik, P. D. 2001. Methods for Behavioral Research: A Systematic Approach. Thousand Oaks, CA: Sage Publications Inc.
Connolly, C. 2005. "Cedars-Sinai doctors cling to pen and paper." Washington Post.
Cutler, D. M., N. E. Feldman, and J. R. Horwitz. 2005. "U.S. Adoption of Computerized Physician Order Entry Systems." Health Affairs 24 (6): 1654-63.
Ford, E. W., A. S. McAlearney, M. T. Phillips, N. Menachemi, and B. Rudolph. 2008. "Predicting Computerized Physician Order Entry System Adoption in US Hospitals: Can the Federal Mandate Be Met?" International Journal of Medical Informatics 77 (8): 539-45.
Ford, E. W., and J. C. Short. 2008. "The Impact of Health System Membership on Patient Safety Initiatives." Health Care Management Review 33 (1): 13-20.
Grembowski, D. 2001. The Practice of Health Program Evaluation. Thousand Oaks, CA: Sage Publications Inc.
Guyatt, G., S. Walter, and G. Norman. 1987. "Measuring Change over Time: Assessing the Usefullness of Evaluative Instruments." Journal of Chronic Diseases 40 (2): 171-8.
Han, Y. Y., J. A. Carcillo, S. T. Venkataraman, R. S. B. Clark, R. S. Watson, T. C. Nguyen, H. Bayir, and R. A. Orr. 2005. "Unexpected Increased Mortality after Implementation of a Commercially Sold Computerized Physician Order Entry System." Pediatrics 116 (6): 1506-12.
Hennekens, C., and J. Buring. 1987. Epidemiology in Medicine. Boston, MA: Little Brown and Company.
Hillman, J. M., and R. S. Givens. 2005. "Hospital Implementation of Computerized Provider Order Entry Systems: Results from the 2003 Leapfrog Group Quality and Safety Survey." Journal of Healthcare Information Management 19 (4): 38-48.
Jha, A. K., C. M. DesRoches, E. G. Campbell, K. Donelan, S. R. Rao, T. G. Ferris, A. E. Shields, S. Rosenbaum, and D. Blumenthal. 2009. "Use of Electronic Health Records in U.S. Hospitals." New England Journal of Medicine 360 (16): 1628-38.
Jha, A. K., E.J. Orav, A. B. Ridgway, J. Zheng, and A. M. Epstein. 2008. "Does the Leapfrog Program Help Identify High-Quality Hospitals?" Joint Commission Journal on Quality and Patient Safety 34 (6): 318-25.
Kaushal, R., K. G. Shojania, and D. W. Bates. 2003. "Effects of Computerized Physician Order Entry and Clinical Decision Support Systems on Medication Safety. A Systematic Review." Archives of Internal Medicine 163 (12): 1409-16.
Kazley, A. S., and M. L. Diana. 2011. "Hospital CPOE Adoption and Quality: An Examination of the United States." Health Care Management Review 36 (1): 86-94.
Koppel, R., J. P. Metlay, A. Cohen, B. Abaluck, A. R. Localio, S. E. Kimmel, and B. L. Strom. 2005. "Role of Computerized Physician Order Entry Systems in Facilitating Medication Errors." Journal of the American Medical Association 293 (10): 1197-203.
Morrill, R., J. Cromartie, and L. Hart. 1999. "Metropolitan, Urban, and Rural Commuting Areas: Toward a Better Depiction of the U.S. Settlement System." Urban Geography 20 (8): 727-48.
Poon, E. G., D. Blumenthal, T. Jaggi, M. M. Honour, D. W. Bates, and R. Kaushal. 2004. "Overcoming Barriers to Adopting and Implementing Computerized Physician Order Entry Systems in U.S. Hospitals." Health Affairs 23 (4): 184-90.
Ray, M. N., T. K. Houston, F. B. Yu, N. Menachemi, R. S. Maisiak, J.J. Allison, and E. S. Berner. 2006. "Development and Testing of a Scale to Assess Physician Attitudes about Handheld Computers with Decision Support." Journal of the American Medical Informatics Association 13 (5): 567-72.
Reckmann, M. H., J. I. Westbrook, Y. Koh, C. Lo, and R. Day. 2009. "Does Computerized Physician Order Entry Reduce Prescribing Errors for Hospital Inpatients? A Systematic Review." Journal of the American Medical Informatics Association 16 (5): 613-23.
Rosner, B. 2006. Fundamentals of Biostatistics. Belmont, CA: Thomson Higher Education.
Rural Health Reserach Center. 2009 "UW RHRC Rural Urban Communting Area Codes--RUCA" [accessed on December 7, 2009]. Available at: http:// depts.washington.edu/uwruca/index.php
Teufel, R.J., A. S. Kazley, and W. T. Basco. 2009. "Early Adopters of Computerized Physician Order Entry in Hospitals that Care for Children: A Picture of US Health Care Shortly after the Institute of Medicine Reports on Quality." Clinical Pediatrics 48 (4): 389-396.
Yu, F., M. Salas, Y. I. Kim, and N. Menachemi. 2009a. "The Relationship between Computerized Physician Order Entry and Pediatric Adverse Drug Events: A Nested Matched Case-Control Study." Pharmacoepidemiology and Drug Safety 18 (8): 751-5.
Yu, F. B., N. Menachemi, E. S. Berner, J. J. Allison, N. W. Weissman, and T. K. Houston. 2009b. "Full Implementation of Computerized Physician Order Entry and Medication-Related Quality Outcomes: A Study of 3364 Hospitals." American Journal of Medical Quality 24 (4): 278-86.
Additional supporting information may be found in the online version of this article:
Appendix SAI: Author Matrix.
Please note: Wiley-Blackwell is not responsible for the content or functionality of any supporting materials supplied by the authors. Any queries (other than missing material) should be directed to the corresponding author for the article.
Address correspondence to Mark L. Diana, M.B.A., M.S.I.S., Ph.D., Assistant Professor, Department of Health Systems Management, Tulane University, 1440 Canal Street, Suite 1900, New Orleans, LA 70112; e-mail: email@example.com. Abby Swanson Kazley, Ph.D., Assistant Professor, is with the Department of Health Professions, Medical University of South Carolina, Charleston, SC. Nir Menachemi, Ph.D., M.P.H., Associate Professor and Director, Doctor of Public Health Program, Health Care Organization and Policy, UAB School of Public Health, Birmingham, AL.
Table 1: Number of Hospital Respondents to the HIMSS and Leapfrog Surveys by Year No. of Hospitals in No. of Hospitals in No. of Individual HIMSS Data with Leapfrog Data with Hospitals Appearing Available Available in BOTH HIMSS and Information on CPOE Information on CPOE Leapfrog Datasets Adoption Adoption that Report CPOE Results 2005 2,357 2,060 1,053 2006 3,734 2,348 1,823 2007 4,679 2,619 2,355 CPOE, computerized provider order entry; HIMSS, Health Information Management Systems Society. Table 2: Organizational Characteristics of Hospitals in Each Sample (2007) Hospitals in Hospitals in Hospital with the HIMSS Data the Leapfrog CPOE Data with Reported Data with Reported from CPOE Reported CPOE Both Data Information Information Sources Sample size 4,679 2,619 2,355 Mean bed size (SD) 165 (182) 204 (200) 206 (201) Size category Small (< 125 beds) 2,619 (56%) 1,028 (39%) 1,028 (44%) Med (126-399 beds) 1,628 (35%) 1,023 (39%) 1,023 (43%) Large (400+ beds) 432 (9%) 314 (12%) 304 (13%) Unknown (missing) 0 (0%) 264 (10%) 0 (0%) Tax status For profit 780 (17%) 376 (14%) 376 (16%) Not for profit 3,899 (83%) 1,979 (76%) 1,979 (84%) Unknown (missing) 0 (0%) 264 (10%) 0 (0%) System affiliated Yes 2,563 (55%) 1,396 (53%) 1,396 (59%) No 2,116 (45%) 959 (37%) 959 (41%) Unknown (missing) 0 (0%) 264 (10%) 0 (0%) Geographic location Urban 2,411 (52%) 1,533 (59%) 1,533 (65%) Rural 2,266 (48%) 820 (31%) 820 (35%) Unknown 2 (0%) 266 (10%) 2 (0%) Organizational type General med/surg 4,438 (95%) 2,285 (87%) 2,285 (97%) Specialty 71 (1.5%) 35 (1.5%) 3.5 (1.5%) Children's 72 (1.5%) 3:5 (1.5%) 35 (1.5%) Acute long-term care 98 (2%) 0 (0%) 0 (0%) Unknown (missing) 0 (0%) 264 (10%) 0 (0%) CPOE, computerized provider order entry; HIMSS, Health Information Management Systems Society. Table 3: Consistency of CPOE Measure in HIMSS and Leapfrog Data over Time (A) (B) (C) Number of Percent in Percent in Hospitals with Column (A) That Column (A) That CPOE in 2005 Had CPOE in Had CPOE in 2006 2007 HIMSS 238 76% (180) 62% (148) Leapfrog 69 86% (59) 86% (59) (D) (E) Percent in Number of Column (D) That Hospitals with Had CPOE in CPOE in 2006 2007 HIMSS 276 78% (216) Leapfrog 95 81% (77) CPOE, computerized provider order entry; HIMSS, Health Information Management Systems Society. Table 4: Agreement between HIMSS and Leapfrog Data Regarding CPOE Status by Year (2005-2007) Leapfrog 2005 HIMSS 2005 CPOE No CPOE Total CPOE 15 120 135 No CPOE 16 902 918 Total 31 1,022 1,053 [kappa] = 0.14, p <.001; McNemar's [chi square] (1) = 79.53, p < .001. Leapfrog 2006 HIMSS 2006 CPOE No CPOE Total CPOE 48 192 240 No CPOE 34 1,549 1,583 Total 82 1,741 1,823 [kappa] = 0.25, p <.001; McNemar's [chi square] (1) = 110.46, p < .001. Leapfrog 2007 HIMSS 2007 CPOE No CPOE Total CPOE 89 362 451 No CPOE 35 1,869 1,904 Total 124 2,231 2,355 [kappa] = 0.25, p <.001; McNemar's [chi square] (1) = 269.34, p < .001. Leapfrog 2005-2007 * HIMSS 2005-2007 CPOE No CPOE Total CPOE 7 85 92 No CPOE 16 1,238 1,254 Total 23 1,323 1,346 Note. [kappa] =0.10, p <.001; McNemar's [chi square] (1)= 47.14, p <.001. * Data restricted to those hospitals that reported the presence of CPOE in all 3 years. CPOE, computerized provider order entry; HIMSS, Health Information Management Systems Society. Table 5: Changes in CPOE Status over Time in the HIMSS and Leap-frog Surveys (A) (B) (C) No. of Hospitals Percent in No. of without CPOE Column (A) That Hospitals in 2005 Had Adopted without CPOE in CPOE as of 2006 2006 Sample not restricted to overlapping hospitals HIMSS data 2,045 out of 2,287 4.7% (n = 97) 3,208 out of 3,638 Leapfrog data 1,991 out 1.8% (n = 36) 2,253 out of 2,348 of 2,060 Sample restricted to overlapping * hospitals HIMSS data 1,208 out of 1,372 5.4% (n=65) 1,865 out of 2,151 Leapfrog data 1,000 out 1.7% (n= 17) 1,716 out of 1,796 of 1,031 (A) (D) No. of Hospitals Percent in without CPOE Column (C) That in 2005 Had Adopted CPOE as of 2007 Sample not restricted to overlapping hospitals HIMSS data 2,045 out of 2,287 9.7% (n = 312) Leapfrog data 1,991 out 0.8% (n= 18) of 2,060 Sample restricted to overlapping * hospitals HIMSS data 1,208 out of 1,372 12.5% (n=234) Leapfrog data 1,000 out 3% (n= 52) of 1,031 * Overlapping hospitals include those responding to both the HIMSS and Leapfrog survey in a given year. CPOE, computerized provider order entry; HIMSS, Health Information Management Systems Society.
|Printer friendly Cite/link Email Feedback|
|Title Annotation:||Methods Articles|
|Author:||Diana, Mark L.; Kazley, Abby Swanson; Menachemi, Nir|
|Publication:||Health Services Research|
|Date:||Oct 1, 2011|
|Previous Article:||Empirical validation of patient versus population preferences in calculating QALYs.|
|Next Article:||Mapping physician networks with self-reported and administrative data.|