About the Authors:
Anat Tchetchik
* E-mail: [email protected]
Affiliation: Department of Business and Management, Guilford Glazer Faculty of Management, Ben Gurion University of the Negev, Beer Sheva, Israel
Amir Grinstein
Affiliations Department of Business and Management, Guilford Glazer Faculty of Management, Ben Gurion University of the Negev, Beer Sheva, Israel, Faculty of Economics and Business Administration, VU University Amsterdam, Amsterdam, The Netherlands, D’Amore-McKim School of Business, Northeastern University, Boston, Massachusetts, United States of America
Eran Manes
Affiliations Department of Business and Management, Guilford Glazer Faculty of Management, Ben Gurion University of the Negev, Beer Sheva, Israel, The School of Industrial Management, Jerusalem College of Technology, Jerusalem, Israel
Daniel Shapira
Affiliation: Department of Business and Management, Guilford Glazer Faculty of Management, Ben Gurion University of the Negev, Beer Sheva, Israel
Ronen Durst
Affiliation: Cardiology Division, Hadassah Hebrew University Medical Center, Jerusalem, Israel
Introduction
The question when and to what extent academic research can benefit practice is a topic of great interest and debate among policy makers and academic researchers [1–4]. Whereas there are recent calls and policies to encourage universities to play an even more active role in the commercialization of academic knowledge and overall to be more engaged with industry [1], [5–7], in reality the gap between theory or research and capability of transferring its results into practice is repeatedly evidenced [3], [8–9]. Further, within the broad topic of the link between academic research and practice a host of open questions remain, for example, what type of academic research can be most effectively translated into better practice? One key reason is that although of clear and enormous importance, the impact of academic research on practice cannot often be easily quantified. Unlike private enterprises that can measure the marginal effect of their expenditure in R&D on economic profits, most publicly funded research is made by non-profit organizations and most efforts of academics are aimed at a single key output: research publications. That is, the accepted measures of academic research impact on society are based on the research output itself rather than its impact beyond “the ivory tower.”
Improvements in practice may potentially be driven from a research strategy that either boosts high-volume research productivity or promotes high-quality research productivity [10]. While the debate among advocators of these two strategies is not in itself novel, most of the literature has been focusing on the extent by which these competing strategies correspond with scientific excellence, while overlooking their linkage with practice. [11]. On the one hand, proponents of the high-quality strategy argue that greater quantity inevitably comes at the expanse of quality [10–12]. This has been demonstrated, for example, in physics and in biomedical research [13–14]. Within the biomedical research domain, it has been noted by a former editor of the New England Journal of Medicine that researchers should be judged primarily by the quality and number of their real scientific contributions, not by the quantity of papers they publish, [15–16]. On the other hand, it is argued that creativity and breakthroughs in science might be the result of numerous trials-and-errors, and that researchers ought to work on a large set of experiments that may result in a large set of minor findings and publications before they eventually establish high-quality research finding. According to this view, therefore, quantity is a necessary (yet insufficient) condition for quality [10], [17–19]. Some empirical findings indeed lend support to this view [10, 20].
This work aims at offering an empirical contribution to the study of the research-practice nexus. Attempts to study this link were limited to the R&D literature and employed indicators such as ‘engagement with industry’ and patenting as a quantifiable proxy for the contribution of R&D to practice [7], [21–25]. Yet, the important linkage between bio-medical research and medical practice has not yet been studied. We attempt at filling this gap by studying physicians in university hospitals. This group represents a highly relevant test case because it engages in biomedical academic research while also providing medical care of measurable quality. Furthermore, biomedical academic research is one of the most heavily funded research areas by governments worldwide. In the United States, for example, at the National Institutes of Health (NIH) in 2012 a budget request for $31.987 billion was aimed at providing funds for more than 40,000 competitive research grants and more than 325,000 research personnel at over 3,000 research institutions and small businesses across the country [26]. Hospital physicians constitute a central group that benefits from these funds: in that year about 60% of the funds awarded by NIH were aimed at medical schools and independent hospitals [27]. Underpinning this policy is the expectation that biomedical academic research makes a valuable contribution to medical practice. Specifically, physicians who engage in biomedical research are expected to improve their clinical capacities because of their exposure to up-to-date medical knowledge and advanced technologies and applications [28]. As physicians’ resource constraints make it difficult to achieve both volume and quality [29–30], there is likely to be a trade-off between the two research strategies, and it is also not at all clear which of the two strategies is of more relevance for healthcare improvement [31–37]. For this reason, and in an effort to maximize the added value of academic research to practice–in the context of biomedical research–the following question arises: what is the type of research productivity strategy that is expected to yield a higher benefit to medical practice? We believe this research question is both novel and important.
While the contribution of biomedical research to medical practice is potentially substantial, any attempt to quantify it runs into both theoretical and methodological difficulties. Nevertheless, such attempts are important if we accept that the justification for generous support of research should be based on measurable performance. For that purpose, we draw a distinction between direct and indirect contributions of biomedical research, of which the former, as defined next, is both quantifiable and measurable. Publishing a scholarly article that reports the findings of a clinical trial or epidemiologic study may have two main effects, among others: it may help streamline existing treatment and care methods, improve diagnostic procedures and so forth. This is a direct effect as it has a straightforward and immediate impact on the quality of healthcare. The greater the extent to which the study employed high-quality techniques and state of the art methodologies, the bigger is the expected contribution of the study to the quality of care. The second, indirect, effect by which the study may affect the existing body of knowledge is by triggering and inspiring future related studies. In that respect, the overall scientific quality of the article may be less important compared to the speed with which it becomes accessible to the relevant scientific community.
In this study we aimed to examine only the direct differential influence of the high-quality versus the high-volume research productivity on the quality of medical care. Capitalizing on the fact that physicians in university hospitals engage in biomedical academic research while also providing medical care of measurable quality, we aimed at measuring the synergy effect between both activities, thus focusing entirely on the direct contribution of research productivity to the quality of medical care while ignoring indirect channels of potential influence. To this end we collected secondary data on departments across three specializations (Cardiology, Oncology and Orthopedics) in 50 U.S.-based university hospitals served by 4,330 physicians. After controlling for covariates known to influence quality of medical care (e.g., hospital’s income per physician, size, and patients’ characteristics), we examined the correlation between the volume and quality of publications of physicians in each department and two measures of the department’s quality of medical care as reported in U.S. News & World Report’s “Best Hospitals” quality of care ranking.
Theoretical reasoning and hypothesis
The idea that investment in R&D is a major determinant of output and long-term growth is a well-established paradigm in the literature on R&D and economic growth, where it enjoys wide theoretical and empirical support [38]. Not only biomedical research but academic research in general confers clear and important long-term benefits on society. The advance of science and academic research continues to play a major role in the progress of humanity by raising the stock of knowledge and improving existing techniques and methods in a multiplicity of areas, not the least of them life-saving treatments and medications. Yet, measurement of this overall effect faces obvious obstacles. However, focusing on the direct effect allows us to identify potential synergy between research and (medical) practice. The underlying rationale is straightforward: by keeping up with latest advances in related fields and literatures, by attending conferences and maintaining contacts with peers, and by reading (and writing) papers that utilize state-of-the-art techniques and methods in the field, physicians who engage in biomedical research are supposed to greatly improve their capacities in whatever realm of practical medicine they occupy. Based on the generally accepted arguments that compared to low-quality research, high-quality research involves exposure to more advanced techniques, more sophisticated literatures and more cutting-edge methods, and that there is a trade-off between volume and quality [10], [15], [39] we suggest the following hypothesis:
Hypothesis: High-quality research productivity at the department level is associated with greater direct contribution to quality of medical care than high-volume research productivity.
It is important to note that our results bear little if any relevance to other potential channels through which research may contribute to the quality of health care, such as the indirect effect mentioned above.
Material and Methods
Sample
Our original intention was to collect data on all physicians whose names appeared on the websites of the sampled hospital departments. We subsequently decided, however, that in the case of departments with more than 100 physicians a sub-sample of approximately 40 randomly selected physicians would suffice. Our sample included 50 U.S.-based university hospitals engaged in biomedical research (S1 Table). Using a publicly available list of all U.S.-based university hospitals as the sampling frame (N = 133) [40] and a stratified sampling approach, we identified 50 geographically heterogeneous hospitals. Within each hospital we identified three central specializations: Cardiology, Oncology, and Orthopedics. Overall, therefore, our data comprised 150 department-level observations.
Biomedical academic research: volume versus quality
The two key independent variables were the volume and quality of biomedical academic research conducted by hospital physicians in each of the three specializations at the department level. These data were not readily available and had to be retrieved through a detailed bibliometric database search for each physician (N = 4,330). The individual-level data were then aggregated to yield the department-level data. A detailed description of the construction of the two variables follows.
To estimate the research volume of sampled physicians, PubMed was used as the basis for calculation of the annual average number of publications during the period between 2002 and 2011. To ensure proper identification of the relevant physician, we typed the physician’s full name (surname, first name, middle initial) in the “author field”. Such a search has the disadvantage of retrieving only publications since 2002, whereas a less limiting search (surname, first and middle initials) retrieves all available records. Nevertheless, by searching as described we obtained a more certain identification of the physician.
To estimate the quality of research by a particular physician, we employed two measures. Our main indicator for research quality is a physician’s average citation per publication and our secondary indicator for research quality is the H-index suggested by Hirsch [41]. Average citations per publication is a relatively clean measure of research impact, though its main weakness is its vulnerability to single peaks. The H-index complements our primary research quality measure as it is considered a consistent and systematic indicator of research impact and quality according to developing scientometric literature [42–43]. Yet, by construction the H-index is highly correlated with the quantity of publications, and it suffers from non-linearity in the sense that, at high levels of H-index, a marginal increase is becoming very difficult to achieve (the number of citations required to obtain the H+1 index is constantly increasing at the margin). We hence choose the average citations per publication to be our main quality indicator and the H-index as the secondary one, for robustness purposes. We obtained both average citations and H-index indicators of each physician from the Web of Science, an online academic citation index that covers over 12,000 journals worldwide. Here too, we searched by typing the physician’s surname, first name and middle initial into the “author field,” enabling us to retrieve an average number of citations per item and a personal H-index for publications since 2006. This search procedure might seem limiting. However since 2006, most journals in the three chosen specializations specify authors’ surname and first name, and hence the probability for missed papers is negligible and nonsystematic. Publications not categorized as “medical research” were excluded from the retrieval process. Physicians who had common names, raising the possibility that publications by more than one physician would be retrieved (e.g., publications listed under both ophthalmology and orthopedics), were removed from the sample. At last, both the quality measures we employed reflect the quality of recent research activity carried out by physicians while also maintaining enough variation to conduct statistical analysis (CV = 0.61 and 0.56 for the average citations and H-index, respectively). Overall, the names of 30 physicians were removed from the sample (about 0.7%), yielding an effective sample of 4,330 physicians.
Quality of medical care
We based our key dependent variable—Index of Hospital Quality (IHQ)—on the 2011–2012 U.S. News & World Report’s “Best Hospitals” ranking, applied at the department level for each of the three specializations. This ranking system is viewed by many physicians and public policy makers as one of the most sophisticated, accurate, and influential means of assessing hospital quality [44–46]. It is based on a comprehensive quality assessment theoretical paradigm, reflecting performance in three intertwined dimensions of healthcare: structure, process, and outcomes [47]. The IHQ is expressed as a composite score between 0 and 100 (with the top-ranked hospitals receiving the highest score), and it incorporates dimensions such as survival rates, success in keeping patients safe, and reputation among physicians. As for the other dimensions of the ranking, they are based on data from multiple sources such as the annual survey database of the American Hospital Association (for a more comprehensive review of the development and use of the IHQ see the “Best Hospitals” ranking’s website [48]). As a separate proxy for quality of care, we also used a sub-dimension of the IHQ index—namely, survival rate. Survival rate refers to hospital’s success at keeping patients alive. It is judged by comparing the number of Medicare inpatients with certain conditions who died within 30 days of admission in the recent three years with the number expected to die given the severity of illness. This measure of hospital quality is composed of four tiers, where the highest survival rate is represented by tier 1 and the lowest by tier 4.
Finally, we collected several relevant covariates known to influence the quality of medical care in hospitals. These are often used in the literature [36], [44] and are described and summarized in Table 1.
[Figure omitted. See PDF.]
Table 1. Descriptive statistics of variables employed in the regression analyses.
https://doi.org/10.1371/journal.pone.0129259.t001
Econometric concerns
Some econometric concerns worth attention. First, note that IHQ, the dependent variable is a complex construct based on both objective data, namely; structure (which refers to resources that relate directly to patient care), outcome (risk-adjusted mortality rates) and the process component of care delivery (which has to do with patient safety) on the one hand, and subjective data, namely “reputation” which is based upon a survey among 200 physicians, on the other hand. As IHQ factors in a reputation component, it may raise concern over endogeneity as reputations might be related to research quality, which is our main independent variable. Note however that for the three specializations chosen in this study, IHQ ranking depends heavily on objective data (i.e. structure, outcome, and process). Moreover, in the survey where physicians were asked to list the hospitals they consider to be the best in their specialization, there was no reference to physicians’ research performance in the process that determines the reputation score. Also, U.S. News conducts a statistical adjustment to keep a small number of hospitals with very high reputational scores from swamping the rest of the field in the final rankings.
Of more concern is the possibility of self-selection; as the assignment of physicians to hospitals cannot be assumed of random nature, a suspicion arises that better researchers may be drawn to hospitals that are in better positions to provide research funding. If such matching exists than the positive correlation between quality of healthcare and various measures of research quantity and quality could just be an artifact of the match between skill and wealth. This problem can be best dealt with by including hospital’s revenue per physician (with a 1-year lag to avoid simultaneity) in the regressions. Given the ceteris paribus nature of a regression analysis, this allows one to measure the impact of the independent variables (volume and quality) on the dependent variables (IHQ) holding the level of hospital wealth constant.
Thirdly, a problem of endogeneity may arise if the quality of medical care and the volume and quality of research are determined jointly. This concern can be dismissed, however, because physicians’ research volume and quality as employed here are measures that accumulate over several years. Moreover, accounting for the fact that research outputs published on any given year are usually the result of distant past efforts, simultaneity should not be present.
Lastly, since by construction, high values of the H-index cannot be achieved if the number of publications is small, we were not surprised to find strong positive correlation between the H-index and the average number of publications (Spearman’s rank correlation coefficient is 0.72, p<0.001). Yet, despite this correlation, in most regressions collinearity did not hinder significance tests, hence we felt safe to use the H-index in its original values. That said, in order to grant more credence to our results we also obtained a “clean” measure of quality from the original H-index, by means of running an auxiliary regression, where we regressed the H-index on the number of publications. The residuals from this regression are, by construction, uncorrelated with the number of publications, and are therefore a cleaner measure of quality. However, as the results obtained with the residuals are similar to those obtained by using the original H-index, we do not report them here.
Results and Discussion
Fig 1 below presents the observed correlations between our key dependent and independent variables, for each specialization. The figure demonstrate that in general there exists a positive correlation between each of the quality/quantity research measures and IHQ. However, the “noise” that is apparent in Fig 1 and the multiple outliers, warrants a more rigorous statistical approach that controls for the possible intervention of other variables. Another notable feature that arises from Fig 1 is the different nature through which IHO is associated with all research quality and quantity measures in the three specializations. This calls for a statistical analysis that employs interaction terms, rather than estimating an average effects.
[Figure omitted. See PDF.]
Fig 1. Scatter plots between IHQ and research output measures by specialization.
Research output measures include: average citations per publication, H-index, average number of publications per year.
https://doi.org/10.1371/journal.pone.0129259.g001
We now turn to our primary statistical analyses. We employed two models that examined the effects of volume and quality of physicians’ research at the department level on their department’s quality of clinical care—one model for each of our dependent variables, IHQ and survival rate. For the first dependent variable IHQ which is a continuous variable, we capitalized on the panel structure of our dataset and used a fixed-effects model with clustered errors where the effects are the specializations (Hausman specification test suggests rejecting the null hypothesis, i.e., that a random effects model is preferred due to greater efficiency).
The empirical equation estimated is:
The dependent variable is the IHQ of specialization i in hospital j. The variables for-Profit, Staffed beds, Length of stay, Median income, Hospital income per physician, and Median age, are control variables at the hospital level. The variable Physiciansij, represents the number of physicians in specialization i in hospital j. The interaction terms: Research volume*Cardiologyj, Research quality*Cardiologyj, Research volume*Oncologyj, and Research quality*Oncology j aimed at capturing the differential effects of research quality and volume on the specializations Cardiology and Oncology, respectively, relative to the omitted specialization, Orthopedic. Finally, αk, k = 1,..,13 is a vector of parameters representing the impact of the independent variables on the dependent variable, βi is a vector of fixed-effects for the three specializations and ɛij is the error term.
Table 2 presents the results of two fixed effects regressions. For both regressions the dependent variable is IHQ, whereas Eq.(1) provides the results of the regression that employs the average citation as a proxy for research quality, and Eq.(2) provides the results of the regression that employs our secondary research quality measure, H-index. All other variables remain the same across the two regressions. For ease of presentation and interpretation only the calculated elasticities are provided followed by their standard error.
[Figure omitted. See PDF.]
Table 2. Calculated Elasticities based on a Fixed-effects regression with clustered errors, Dependent variable: index of hospital quality (IHQ).
https://doi.org/10.1371/journal.pone.0129259.t002
As noted, in both regressions the coefficients of the non-interacted terms: Avg. publications, Avg. citations and H-index refer to the effect of these variables on the IHQ of the omitted specialization (Orthopedics in this case). The coefficients of the interacted terms (Average publications*cardiology, Average publications*Oncology, Average Citations*cardiology, Average Citations*Oncology, H-index*cardiology and H-index*Oncology) refer to the effect of Avg. publication, Avg. citation and H-index in the specializations Oncology and Cardiology on IHQ relative to their effect in the omitted specialization-Orthopedics. The findings suggest that volume of research has a significant positive effect on clinical excellence, in particular for the specializations Cardiology and Oncology. For example, ceteris paribus, an increase of 10% in research volume (Avg. publications) in Oncology is correlated with an average increase of 0.92% and 1.2% in the IHQ according to Eq.(1) and Eq.(2) respectively, whereas in Cardiology, the same increase in research volume in is correlated with an average increase of 0.53% and 0.6% in the IHQ according to Eq.(1) and Eq.(2) respectively. However, the results show that holding research volume constant, a rise in research quality has a greater effect on clinical performance for all specializations. For example, a rise of 10% in the Avg. citations in Orthopedics correlates with 1.58% rise in IHQ, 1.78% in Cardiology, and 1.58% in Oncology. For the H-index regression (right column), an increase of 10% in Orthopedics’ H-index correlates with 3.25% increase in the IHQ, 2.65% increase for Cardiology, and 2.93% increase for Oncology.
It is noteworthy that while the distributions of our quality indicators—Avg. citations and H-index—are right skewed for many departments, we still adhere to the use of the average statistic for the following reason: it appears that in majority of cases there is a large portion of new/non-tenured researchers with very low quality figures, and a small portion of senior researchers with high quality figures. We therefore believe, based on conversations with senior physicians/researchers that these relatively high quality researchers are the “tone givers” and the quality benchmarks, and therefore the average better represents the department’s quality of research. Still, for robustness we also conducted the same analysis using medians instead of averages. The results appear in Table 3, and as can be seen, they remain qualitatively similar: while research volume has positive and significant effect in most cases, quality is associated with greater impact on the dependent variable IHQ.
[Figure omitted. See PDF.]
Table 3. Calculated Elasticities based on a Fixed-effects regression with clustered errors, dependent variable: index of hospital quality (IHQ).
https://doi.org/10.1371/journal.pone.0129259.t003
Note that, as in Table 2, the coefficients of the interacted terms (volume interacted with Cardiology and Oncology and quality interacted with both specialties) are measured relative to the coefficients of the un-interacted terms for volume and quality. These latter terms reflect the impact of research volume and quality, respectively, on IHQ at the omitted specialization (Orthopedics)
We now move on to test the model with the second dependent variable: survival rate, which is an ordered categorical variable ranked from 1 to 4. For this purpose we used an ordered logistic regression.
Because our data included only four observations at the lowest survival rate ranking (i.e., tier 4), they were merged into category 3. For simplicity of exposition and interpretation, we reversed the order of the variable. Also, we included the covariate clinical services, which was not included in the previous model because it is used to construct the IHQ variable. Due to space limits this regression was ran only for our main quality measure; average citations, for the secondary measure, due to high collinearity between the H-index and avg. papers, this regression did not yield significant results. Table 4 reports the calculated elasticities, i.e., the change in the probability to be ranked top tier (highest level of survival rate) as each predictor increases by 1% (or from 0 to 1 in the case of a dummy variable).
[Figure omitted. See PDF.]
Table 4. Generalized ordered logistics regression with clustered errors.
Dependent variable: survival rate (reversed).
https://doi.org/10.1371/journal.pone.0129259.t004
The results in Table 4 are in line with those in Table 2; research volume has positive and significant effect on the dependent variable, in particular in Oncology. More specifically, a 10% increase in the number of publications correlates with 1.89% increase in the probability to be ranked in the top tier. Yet, research quality as measured by average citations has a greater effect. For Orthopedics and Cardiology, an increase of 10% in the average citations increases the probability to be ranked top tier by 2.67% on average, and by 2.9% in Oncology.
As a final robustness check, we also run the same regressions after eliminating from the sample physicians who did not publish at all. We assume that these physicians have graduated from medical school not long before the time the survey was conducted and hence they have not yet realized their potential as researchers (S2 Table).
Discussion
Understanding the relationship between research and practice is a persistent and difficult problem for policy makers and scholars who work in professional schools, such as business, engineering, medicine, education, and public administration. Per the latter, although one of professional schools’ key missions is developing knowledge that can be translated into skills that advance the practice of the professions, this mission often remains an elusive ideal [3].
We try to contribute to the debate over the link between academic research and practice by shedding light on the dilemma whether the volume or the quality of academic research provides a more valid measure of research contribution to practice [10], [15], [17]. This dilemma is relevant for many academic disciplines and for society at large. Although the debate remains unsolved in many cases, the fact that physicians at university hospitals produce both research output and medical care of measurable quality presents a unique opportunity to measure the differential effect of high-quality versus high-volume biomedical academic research on medical practice, thereby offering a partial answer to the debate.
In this study, we empirically demonstrated that high-quality research is significantly correlated with quality of medical care, as measured by IHQ and survival rate. While we also show that high-volume research may also correlates with quality of medical care, this effect is significantly weaker than the correlation of high-quality research with quality of medical care. For any given level of research volume an increase of 1% in high-quality research contributes more to quality of medical care than an increase in 1% of high-volume research. Our results provide empirical evidence for the claim that in most cases high-quality research, is associated with more value to practice [14–16], [51]. Our interpretation of the results is as follows: owing to the synergy between biomedical research and medical practice, and based on the assumption that high-quality research uses more meticulous standards, more advanced methods and more cutting-edge techniques, it should produce better results in terms of practical outcome than high-volume, often low-quality research. Moreover, we maintain that the actual positive association between the quality of biomedical research and the quality of medical care is even stronger than the apparent results because in principle it is reasonable to assume that those institutions that conduct high-quality research often attract the most severe medical cases.
Little information is available on the contribution of biomedical academic research to the quality of medical care. Previous research has focused primarily on the relationship between hospital teaching status and health care quality. Most of these studies suggested that clinical care outcomes in teaching hospitals are better than in non-teaching hospitals [52], although these findings are also under debate [53–54]. To the best of our knowledge, the association between biomedical research and hospitals’ quality of care has been examined in only one empirical study [36] which focused on one category (cardiovascular) and adopted a basic analytical approach (Spearman's rank correlation tests). The authors of that study reported a moderate link between physicians’ research productivity (e.g., numbers of publications and citations) and hospital-level risk-adjusted mortality ratio for congestive heart failure and acute myocardial infarction, suggesting that increased research productivity may lead to better healthcare. This result points to one theoretically relevant linkage between biomedical research and practice, perhaps the most intuitive one. However, bibliometric research on the more general relationship between academic research and university/academic unit ranking shows that the reality is more complex; A variety of research publication strategies exist, and different strategies may yield different consequences [55]. Our study indeed demonstrates the different consequences of high-volume versus high-quality research productivity strategies.
Limitation of our study
Several caveats of our research should be mentioned. The first is related to the fact that we are able to measure only the direct effect between research and practice. Although research volume by itself has relatively smaller effect on quality of medical care, it may have an indirect effect. More specifically, in reality research volume and research quality are correlated and it could be argued that a combination of some level of volume with an emphasis on quality research is optimal [19], [35]. Unfortunately, in most cases resource constraints force scientists (and physicians) to choose one publication strategy over the other.
A second caveat of this research, which opens the door for future related studies, has to do with the measurement of quality. Quality of research is a complex concept, the measuring of which entails more sophisticated methods than a simple scalar measures such as the H-index. Although we used different measures of research quality, the scatter plot matrix (S1 Fig) demonstrates a positive correlation between the two measures (especially in Oncology). Future research may therefore consider the use of other research quality measures. For example, the percentage or number of papers published in the top 10% journals of the field may be a meaningful marker for quality. While the debate on how to measure research quality rests outside the scope of this study, we still feel that obtaining similar results while using two different quality measures renders more consistency to our methodological approach.
Finally, due to the cross sectional nature of our data, the existence of causal relationship between biomedical research quality and quality of medical care cannot be inferred. The latter requires a panel data with time series.
Conclusions
The management of research—from selection of ideas and projects to choice of the average time spent on a research project—is clearly incentive-dependent. Our results imply that incentives should be directed at scientists, physicians, or academic departments that yield research publications of high impact. This view finds some empirical support in the work of Harzing [39], who demonstrates (although not only in the biomedical research area) that the Australian government’s performance evaluations of academic institutions showed a sharp decline in research impact during the period in which evaluation emphasized volume rather than quality of publications. Our study also corresponds with Azoulay, Zivin, and Manso [56] who showed that incentive schemes that reward long-term success, tolerate early failure and give its appointees greater freedom to experiment and take more risky and innovative scientific projects, are more conductive to scientific breakthroughs.
Our research may have far-reaching policy implications for the creation of incentive schemes aimed at scientists, the public funding of academic research, and policy making at the government and regulatory levels. More specifically, it conveys implications for physicians engaged in biomedical academic research, for the public funding of biomedical research, and for health-policy making at government and hospital levels. An important policy implication of this research concerns the need to initiate an adequate publicly funded system for evaluation of quality and impact standards, according to which the performance of academic departments and individual scientists will be assessed and financially supported. Some countries, notably the UK and Australia, already employ such systems.
In the specific context of biomedical research, many hospital physicians are under great pressure to publish (or perish) and have to juggle between patients, research, teaching, and administrative work. This reality often drives them to compromise on the quality of biomedical research in favor of volume of publications. The vast majority of biomedical research papers are indeed published in low-impact journals [31]. Moreover, the pressure to publish sometimes even leads to scientific fraud [32, 37]. Our findings may help persuade health policymakers, hospital managers, and physicians to consider a shift in the current approach by designing and providing the necessary incentives for physicians to use their scarce time in the most valuable way, that is to say by carrying out high-quality research, even at the expense of lower volume.
Supporting Information
[Figure omitted. See PDF.]
S1 Fig. Correlation between H-index and Avg. Citations (standardized) by specialization.
Standardized scores for Average citations and Average H-index were calculated at the department level in hospital j for each specialization i as follows:
https://doi.org/10.1371/journal.pone.0129259.s001
(TIF)
S1 Table. List of university hospitals studied.
https://doi.org/10.1371/journal.pone.0129259.s002
(DOCX)
S2 Table. Calculated Elasticities based on a Fixed-effects regression with clustered errors, dependent variable: index of hospital quality (IHQ).
Quality and quantity indicators calculated after omitting newly recruited physicians’ data.
https://doi.org/10.1371/journal.pone.0129259.s003
(DOCX)
Acknowledgments
We are grateful to the editor Peter van den Besselaar, to Frank Miedema and two anonymous referees for their insightful comments and suggestions, which helped improving this manuscript significantly. We are also indebted to Eshel Ben-Jacob for his encouragement and for very helpful comments that greatly improved the methodological section of the paper and to Ariel Halevi for valuable comments and insights.
Author Contributions
Conceived and designed the experiments: AG EM DS RD. Performed the experiments: AG AT EM DS. Analyzed the data: AT EM. Contributed reagents/materials/analysis tools: AT. Wrote the paper: AG EM AT DS RD.
Citation: Tchetchik A, Grinstein A, Manes E, Shapira D, Durst R (2015) From Research to Practice: Which Research Strategy Contributes More to Clinical Excellence? Comparing High-Volume versus High-Quality Biomedical Research. PLoS ONE 10(6): e0129259. https://doi.org/10.1371/journal.pone.0129259
1. Zinner DE, Bolcic-Jankovic D, Clarridge B, Blumenthal D, Campbell EG. Participation of academic scientists in relationships with industry. Health Affairs. 2009; 28(6):1814–1825. pmid:19887423
2. Knights D, Scarbrough H. In search of relevance: Perspectives on the contribution of academic-practitioner networks. Organization Studies. 2010; 31(9–10): 1287–1309.
3. Van De Ven AH, Johnson PE. Knowledge for theory and practice. Academy of Management Review. 2006; 31(4): 802–821.
4. Bonham AC, Solomon MZ. Moving comparative effectiveness research into practice: Implementation science and the role of academic medicine. Health Affairs. 2010; 29(10):1901–1905. pmid:20921492
5. Siegel DS, Wright M, Lockett A. The rise of entrepreneurial activity at universities: organizational and societal implications. Industrial and Corporate Change. 2007; 16(4): 489–504.
6. Bercovitz J, Feldman M. Entrepreneurial universities and technology transfer: A conceptual framework for understanding knowledge-based economic development. Journal of Technology Transfer 2006; 31(1):175–188.
7. Perkmann M, King Z, Pavelin S. Engaging excellence? Effects of faculty quality on university engagement with industry. Research Policy. 2011; 40: 539–552.
8. Byron JP, McMillen JC, Proctor EK, Carpenter CR, Griffey RT, Bunger AC,et al. A compilation of strategies for implementing clinical innovations in health and mental health. Med Care Res Rev. 2012; 69(2): 123–157. pmid:22203646
9. Landry R, Saihi M, Amara N, Ouimet M. Evidence on how academics manage their portfolio of knowledge transfer activities. Research Policy. 2010; 39: 1387–1403. pmid:20339908
10. Adams J. Benchmarking international research. Nature. 1998; 396(6712): 615–618. pmid:9872303
11. Cole S, Cole JR. Scientific output and recognition: A study in the operation of the reward system in science. American Sociological Review. 1967, 32: 377–390. pmid:6046811
12. Feist GJ. Quantity, quality, and depth of research as influences on scientific eminence: Is quantity most important? Creativity Research Journal. 1997; 10(4): 325–335.
13. Cole JR, Cole S. Social stratification in science. 1973; Chicago: University of Chicago Press.
14. Groneberg-Kloft B, Fischer TC, Quarcoo D, Scutaru C. New quality and quantity indices in science (NewQIS): The study protocol of an international project. J Occup Med Toxicol. 2009; 4: 16. pmid:19555514
15. Angell M, Relman AS. Redundant Publication. New England Journal of Medicine. 1989; 321: 1212–1214.
16. Relman AS. Publishing biomedical research: Roles and responsibilities. Hastings Center Report. 1990; 20(3): 23–27. pmid:2376495
17. Simonton DK. Scientific genius: A psychology of science. 1988, Cambridge, England: Cambridge University Press.
18. Simonton DK. Chance-configuration theory of scientific creativity. In Gholson B, Shadish WR, Neimeyer RA, Houts AC (Eds.), Psychology of science: Contributions to meta-science, Cambridge, England: Cambridge University Press. 1989. pp. 170–213.
19. Simonton DK. Creativity in science: Chance, logic, genius, and zeitgeist. 2004, Cambridge University Press.
20. Lawani SM. Some bibliometric correlates of quality in scientific research. Scientometrics. 1986; 9(1–2): 13–25.
21. Calderini M, Franzoni C, Vezzulli A. "If star scientists do not patent: The effect of productivity, basicness and impact on the decision to patent in the academic world." Research Policy. 2007; 36(3): 303–319.
22. Von Proff S. Patent and publication productivity of German professors–a life cycle view. International journal of technology transfer and commercialization. 2011; 10(3): 392–410.
23. Azoulay P, Ding W, Stuart T. The determinants of faculty patenting behavior: Demographics or opportunities? Journal of economic behavior & organization. 2007; 63(4), 599–623.
24. Crespi G, D’Este P, Fontana R, Geuna A. The impact of academic patenting on university research and its transfer. Research policy. 2011; 40(1): 55–68.
25. D’Este P, Tang P, Mahdi S, Neely A, Sánchez-Barrioluengo M. The pursuit of academic excellence and business engagement: is it irreconcilable? Scientometrics. 2013; 95(2): 481–502.
26. Collins FS. Department of Health and Human Services, National Institutes of Health Fiscal Year 2012 Budget Request, Witness appearing before the Senate Subcommittee on Labor–HHS–Education Appropriations, May 11, 2011.
27. Department of Health and Human Services, National Institutes of Health. 2012. http://report.nih.gov/award/organizations.cfm?ot=&fy=2012&state=&ic=&fm=&orgid Accessed 2012 Mar 13.
28. Gourevitch MN, Jay MR, Goldfrank LR, Mendelsohn AL, Dreyer BP, Foltin GL,et al. Training physician investigators in medicine and public health research. Am J Public Health. 2012; 102(7): e39–e45. pmid:22594745
29. Danielle O. Neuron overload and the juggling physician. The Lancet. 2010; 376(9755): 1820–1821. pmid:21128338
30. Manabe YC, Shevin TJ, Thomas D, Quinn TC, Ronald A, Coutinho A, et al. Resurrecting the triple threat: Academic social responsibility in the context of global health research. Clin Infect Dis. 2009; 48(10): 1420–1422. pmid:19368501
31. Altman DG. The scandal of poor medical research. BMJ. 1994; 283–308.
32. Altman DG. Poor-quality medical journals: What can journal do? JAMA. 2001; 287(21): 2765–2767.
33. Altman LK. For science’s gatekeepers, a credibility gap. The New York Times, May 2, 2006.
34. Colquhoun D. Publish-or-perish: Peer review and the corruption of science, The Guardian, September 5, 2011. http://www.guardian.co.uk/science/2011/sep/05/publish-perish-peer-review-science. Accessed 2013 Jan 25.
35. Matheson A. Five steps for structural reform in clinical cancer research. Am J Public Health. 2010; 100(4): 596–603. pmid:20167889
36. Pons J, Sais C, Illa C, Me´ndez R, Sunen E, Casas M,et al. Is there an association between the quality of hospitals’ research and their quality of care? J Health Serv Res Policy. 2010; 15(4): 204–209. pmid:20466753
37. Wager E. Authors, ghosts, damned lies, and statisticians. PLoS Med. 2007; 4(1): e34. pmid:17227138
38. Romer P. Endogenous technological change. Journal of Political Economy. 1990; 98: S71–S102.
39. Harzing AWK. Australian research output in economics & business: High volume, low impact? Australian Journal of Management. 2005; 30(2): 183–200.
40. University hospitals in the United States from Wikipedia, the free encyclopedia. http://en.wikipedia.org/wiki/Category:University_hospitals_in_the_United_States. Accessed 2013 Jan 25.
41. Hirsch JE. An index to quantify an individual’s scientific research output. Proc Natl Acad Sci USA. 2005; 102(46): 16569–16572. pmid:16275915
42. Ball P. Index aims for fair ranking of scientists. Nature. 2005; 436: 900. pmid:16107806
43. van Raan AFJ. Comparison of the Hirsch-index with standard bibliometric indicators and with peer judgment for 147 chemistry research groups. Scientometrics. 2006; 67(3): 491–502.
44. Chen J, Radford MJ, Wang Y, Marciniak TA, Krumholz HM. Do “America’s best hospitals” perform better for acute myocardial infarction? N Engl J Med. 1999; 340: 286–292. pmid:9920954
45. Green J, Wintfeld N, Krasner M, Wells C. In search of America’s best hospitals:: The promise and reality of quality assessment. JAMA 1997; 4:1152–1155.
46. Rosenthal GE, Chren MM, Lasek RJ, Landefeld CS. The annual guide to “America’s best hospitals”. Evidence of influence among health care leaders. J Gen Intern Med. 1996; 11: 366–369. pmid:8803744
47. Donabedian A. Basic approaches to assessment: structure, process and outcome. In: The definition of quality and approaches to its assessment: explorations in quality assessment and monitoring (vol 1). Donabedian A, ed. Ann Arbor, MI: Health Administration Press, 1980 pp. 77–128.
48. Comarow A. Best Hospitals 2012–13: How They Were Ranked, July 16, 2012. http://health.usnews.com/health-news/best-hospitals/articles/2012/07/16/best-hospitals-2012-13-how-they-were-ranked. Accessed 2013 Jan 25.
49. Williams R. Generalized ordered logit/partial proportional-odds models for ordinal dependent variables. Stata Journal. 2006, 6: 58–82.
50. Fu V. Estimating generalized ordered logit models. Stata Technical Bulletin. 1998; 44: 27–30. Stata Press: College Station, TX
51. Bronwyn HH, Jaffe A, Trajtenberg M. Market value and patent citations: A first look. NBER Working Paper No. 7741 2000.
52. Kupersmith J. Quality of care in teaching hospitals: A literature review. Acad Med. 2005; 80: 458–466 pmid:15851459
53. Peterson ED, Roe MT, Mulgund J, DeLong ER, Lytle BL, Brindis RG. Association between hospital process performance and outcomes among patients with acute coronary syndromes. JAMA. 2006; 295: 1912–1920 pmid:16639050
54. Reid RO, Friedberg MW, Adams JL, McGlynn EA, Mehrotra A. Associations between physician characteristics and quality of care. Arch Intern Med. 2010; 170(16); 1442–1449. pmid:20837830
55. Liner GH, Sewell E. Research requirements for promotion and tenure at PhD granting departments of economics. Applied Economics Letters. 2009; 16: 765–768.
56. Azoulay P, Graff Zivin JS, Manso G. Incentives and creativity: Evidence from the academic life sciences. RAND Journal of Economics. 2011; 42(3): 527–554.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2015 Tchetchik et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The question when and to what extent academic research can benefit society is of great interest to policy-makers and the academic community. Physicians in university hospitals represent a highly relevant test-group for studying the link between research and practice because they engage in biomedical academic research while also providing medical care of measurable quality. Physicians’ research contribution to medical practice can be driven by either high-volume or high-quality research productivity, as often pursuing one productivity strategy excludes the other. To empirically examine the differential contribution to medical practice of the two strategies, we collected secondary data on departments across three specializations (Cardiology, Oncology and Orthopedics) in 50 U.S.-based university hospitals served by 4,330 physicians. Data on volume and quality of biomedical research at each department was correlated with publicly available ratings of departments’ quality of care, demonstrating that high-quality research has significantly greater contribution to quality of care than high-volume research.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer