Abstract- Delphi as a consensus development technique enables anonymous, systematic refinement of expert opinion with the aim of arriving at a combined or consensual position. In this study, we determined the criteria and their weights for Iranian Medical Schools' ranking through a Delphi process. An expert committee devised 13 proposed criteria with 32 indicators with their weights, which were arranged hierarchically in the form of a tree diagram. We used the Delphi technique to reach a consensus on these criteria and weights among the deans of 38 public Iranian medical schools. For this purpose, we devised and sent a questionnaire to schools and asked them to suggest or correct the criteria and their weights. We repeated this process in two rounds till all the schools reached an acceptable consensus on them. All schools reached a consensus on the set of 13 criteria and 30 indicators and their weights in three main contexts of education, research and facilities, and equipment which were used for Medical Schools' ranking. Using Delphi technique for devising the criteria and their weights in evaluation processes such as ranking makes their results more acceptable among universities.
© 2016 Tehran University of Medical Sciences. All rights reserved.
Acta Med Iran, 2016;54(6):401-406.
Keywords: Delphi technique; Consensus development methods; Ranking; Criteria
Introduction
In group decision making, the group is often dominated by the ideas of a few members. Consensus techniques are used to eliminate this problem and make sure that all the ideas are taken into effect equally. One of these techniques is the Delphi process that can be deployed both for consensus measurement and consensus development (1,2).
Delphi provides the opportunity for systematic refinement of experts' opinion anonymously to reach to a consensual position (2). It is widely used in health research, education, nursing practice, clinical medicine, prioritization and qualitative research (1-5).
In Delphi technique information and opinions of participants are gathered without their physical assembling to overcome geographical or time limitations. Instead, media like mail, Fax, or email are used. This reliable technique facilitates problem solving and decision-making (1,6) and is a structured process held in several rounds for collecting and resending the opinions to a group of experts till a reasonable level of consensus is reached (7,8).
Delphi participants should be aware of its aim to be able to answer appropriately or keep their interest. This is considered to be the most important issue in a Delphi exercise. On the other hand, there is no need for high degree of expertise (2). However, the Delphi method has got some limitations too. The most important one is about its accuracy (9). The critiques declare that Delphi is suitable as the last solution for extremely complex problems for which there are no other better models (9). Also, some critiques express that in Delphi method the researcher makes participants believe that the predetermined outcome is their own. In this case, better defense would be provided for convincing a larger community (10).
On the other hand, transparency has become the main component of the recent education systems. There are a wide variety of tools available for increasing transparency in higher education. A well-known example is university ranking, which is useful for enhancing the quality of education (11).
Poorly-designed ranking systems may result in deceptive information so that an increase in an institution's rank may not indicate an increase in its quality (12). In spite of the criticisms, evidence indicates that rankings have an influence on academic decision-making and the structure of the educational organizations (13). Universities are supposed to take good advantage of the ranking as a tool that shows their educational and research strategic goals (13).
There is no agreement on the definition of quality in higher education because quality has different meanings from the perspective of various stakeholders. So university ranking systems for a measure of quality should consider the needs of all stakeholders and their participation (11).
Ongoing comparison of a university's operations, and performance against others provides the chance of identifying "best" practices and the university finds a benchmark for setting its own goals (14).
In Iran, the Ministry of Health and Medical Education (MOHME) as the central body of medical education is responsible for the evaluation of universities of medical sciences. It has adopted many short- and long-term strategies to establish efficient modern systems of evaluation and ranking. The product of the ranking process is useful both for the ministry as a guide for planning and for each university as a tool to recognize its position amongst all others. MOHME performed the first Medical Schools ranking in 2003 (1) and repeated the process for three times in two years intervals. Delphi technique was used to devise a set of criteria and their weights in this ranking process, which is going to be presented in this article.
Materials and Methods
To rank the medical schools nationwide, we constituted a medical education expert committee, which devised a set of 13 criteria and 32 indicators in three main contexts of education, research and facilities, and equipment, and arranged them hierarchically in the form of a tree diagram.
The summary of mathematical procedures for calculating the scores of main branches of the diagram and consequently the trunk of the tree is as follows: Schools' scores in the end-branch indicators were calculated based on the data gathered from the schools according to the devised guidelines. Then they were standardized, i.e. converted into a 0 to 100 scale before being multiplied by their weights. The weight of each criterion is a predetermined constant for each branch, which is applied to all the schools unanimously. The total weight of all the criteria is 100%. Finally, the resultant weighted scores for the similar criteria (pertaining to the same node on the tree diagram) were summed up to derive their parent branch's score. These steps are likewise repeated until the total scores of highest levels of the diagram were obtained. Eventually, the total scores of the schools were re-standardized, and schools were stratified based on their total individual scores. Furthermore, they were also ranked in each indicator (1).
To devise above mentioned criteria and indicators, medical education expert committee suggested the criteria and their raw weights using the Focus Group Technique. In the next step, we performed the Delphi technique among the deans of all 38 public Iranian medical schools, as representatives of their schools' educational system. We invited them to participate in a workshop in which we introduced the project, explained all the indicators and criteria, and the methods of weighting, especially the Delphi technique. Then we gave them the first round Delphi questionnaire in which each level of the tree diagram was shown on a separate page, and the expert committee's proposed weight for each criterion was mentioned in a box. There was an empty box for schools' suggested weights and a blank space for their comments. On the next page, we prepared the schools with the detailed definitions of the same criteria. We designed these two pages for every level of the ranking tree diagram. According to our Delphi instruction, the deans had to constitute independently a committee in their schools, comprising the vice-chancellors of education, research, student affairs, and any other experts to discuss and express their comments on the importance of the criteria and finally suggest the new weights in prepared empty boxes with their related comments.
They returned the questionnaires by post in 20 days. We had emphasized that "no response" from schools at each round meant complete agreement with the proposed weights. However, the process of responding was encouraged by calling the schools' deans within 10 days after the workshop.
We calculated the means of the schools' proposed weights after discarding the lowest and the highest 5% (trimmed mean), and then organized the comments and sent them for the second round of Delphi to be reviewed by schools. So they could change their responses based on the mentioned weights and comments, to reach more favorable consensus. In the second round, an acceptable consensus was reached.
MOHME has traditionally categorized the country's universities according to their size into three types: 1 (large), 2 (medium) and 3 (small). We compared the suggested weights by these three types of significant differences. We used Independent T-test for this purpose.
Results
In the first round the response rate was 92% (35 schools out of 38) and in the second round the response rate was 89%; i.e. of the 38 school participating in the survey, 28 (74%) schools completed the questionnaire; and 6 (15%) schools agreed with all criteria and weights without completing the questionnaire.
Finally, schools reached a consensus on a tree diagram including a set of 13 criteria and 30 indicators with their weights, in three main contexts of education, research and facilities, and equipment. In this two rounded Delphi, four indicators were omitted, and two were added and the trimmed means for final weights were in the range of first rounds' weights +/- 2%.
The final criteria and indicators are as follows:
Education
1- National University Entrance Examination (NUEE) score (school's admission): In Iran medical students are admitted through a national entrance examination. This indicator was determined by admitted students' average score at NUEE.
2- Ratios:
2-1- Faculty: class size in basic sciences and clinical settings, student/faculty ratios, full and associate professors to all faculty members' proportion, and board membership.
2-2- Educational facilities: Students/facilities ratios for laboratory equipment, and educational beds.
3- Educational activities:
3-1- Credits and courses: consist of core curricular credits, basic sciences' laboratory programs, extracurricular clinical training, extracurricular courses (computer and English language courses), and the frequency of delivering credits and courses, and summer courses.
3-2- Formative evaluation: including assessment of teaching references, exams' references and procedures.
4- Clinical training:
4-1- Inpatient setting: assessing scholarly activities such as teaching rounds, morning reports, journal clubs, and mortality and morbidity reports.
4-2- Outpatient setting: assessing clinics' training programs.
5- Administrative affairs:
5-1- Compliance with regulations: includes assessing prerequisites, conditional and dismissal regulations.
5-2- Orientation: new student orientation program, regulations and programs' guide book, and academic faculty advisor.
6- Students' output:
6-1- Graduation rate: the percentage of students who succeed in graduation within the assigned period of study.
6-2- Comprehensive examinations: including basic sciences and pre-internship comprehensive examinations that are held by MOHME.
6-3- Acceptance as resident: the percentage of graduates who are accepted into residency programs within two years after graduation.
Research
7- Research grants: the total grant allocated to researchers.
8- Research activities:
8-1- Research projects and dissertations: including assessment of research activities, dissertations, and research related workshops.
8-2- Research centers: assessing human resources, equipment, research activities, output and publication in 3 types of research centers i.e. MOHME approved, the university approved and student centers.
9- Publications:
9-1- Journals and bulletins
9-2- Faculty books
9-3- Faculty articles: including articles published in ISI, Pubmed and Scopus journals, Iranian indexed journals and other journals.
10- Congresses and seminars:
10-1- National: number of events and their continuous medical education credits.
10-2- International: number of events and their continuous medical education credits.
Facilities and equipment
11- Education and research:
11-1- Library: including reference books, journals, and physical spaces.
11-2- Audiovisual resources
11-3- Computer resources
11-4- Educational facilities: including facilities in basic sciences laboratories and clinical wards, and paraclinical equipment.
12- Accommodation & Welfare:
12-1- Dormitory and pavilion: including the ratio of students/beds, and facilities.
12-2- Health services for students: including clinics for students and insurance.
12-3- Students' nutrition: including menus and restaurant facilities,
12-4- Sports facilities: including saloons and their equipment, and sports teams.
12-5- Faculty welfare: including clubs and housing facilities.
12-6- Special services: including all other facilities for faculty members and students.
13- Educational Spaces:
13-1- School: including classes and spaces.
13-2- Hospitals: including wards, operation rooms, and clinics.
Four indicators were omitted through Delphi process: "granting facilities for faculty members", "spin-off companies", "recruitment process" and "special exams". On the other hand, two criteria of "health services for the students" and "formative evaluation" were added.
Table 1 demonstrates the final criteria and indicators designed for ranking medical schools, and their related weights.
Type 2 and 3 universities had suggested lower weights for the criteria relating to faculty ratios, including faculty per student ratio, class size and full and associate professors to all faculty members' proportion (P.value=0.002). In addition, type 1 universities, tended to increase the weights of criteria related to clinical training (P=0.01). This trend was obvious in the research center criteria as well (P=0.001). These universities had given higher weight to "publications" criteria (P=0.03). Another significant difference was seen in "board membership" criterion, and Type 1 universities had suggested significantly higher weight for this criterion (P=0.000). No significant differences were observed in other criteria among these three types of universities.
Discussion
Nowadays, rankings provide the chance of comparing medical schools regarding academic, education, research, or other special (e.g. social) missions (15-17). The ranking process has a competitive nature, which makes it somehow unacceptable for some schools. So as a national project, the cooperation of schools in every defined step of the process, e.g., determining the weights of criteria as one of the most important steps is critical (13).
We had no better choice than the Delphi method for consensus development on determining the criteria and their weights; because:
- There were 38 school deans nationwide who had to make expert committees within their schools. So they constituted a large number of experts who could not be gathered physically for consensus development.
- As a national project, we had geographical limitations for gathering experts in different meetings.
We made our Delphi process more effective by letting the participants be informed of other schools' comments and explanations about the proposed weights. Universities' proposed weights showed the bias of their status. There are fewer full and associate professors in small universities; so, type 3 universities suggested lower weight for this criterion. In addition, the lack of research centers and a low number of academic publications in small universities has led them to decrease the weights of these criteria too. On the contrary, large universities suggested more weights to these criteria, as well as "board membership" and "clinical training" criteria. During this two-round Delphi technique, 4 indicators were omitted, and 2 were added to the tree diagram. The main reasons for these changes were:
1- The main objective of this ranking was comparing Medical Schools in training medical students. The schools omitted the criteria that they believed were not relevant to this purpose, such as "spin-off companies," "granting facilities for faculty members" and "faculty recruitment process."
2- Schools omitted the criteria which were not uniform among all schools, such as "special exams".
They also added "formative evaluations" and "health services for students" to the diagram.
By performing the Delphi technique, we reached to the weights on which all the schools had a consensus. Starting the first round of the Delphi with experts' proposed weights facilitated the participation of the schools. We assumed that providing proposed weights in the first round and universities' comments for their peers in the second round were the main reasons for achieving consensus in only two rounds. On the other hand, this participating process caused the project to be feasible and made the results more reliable and suitable for further planning in the MOHME, motivated the schools to get involved and cooperate actively in the project, and encouraged them to make advances based on each criterion.
Acknowledgement
This work was granted by Iran's Ministry of Health and Medical Education. The authors thank Prof. Hasan Ghazizadeh Hashemi, for his support and insightful comments and suggestions. The authors wish to express their appreciation to all members of expert committees and deans of medical schools, for their participation in the study.
References
1. Mohammadi A, Mojtahedzadeh R. Medical schools of Iran, ranking and database. Tehran: Ministry of Health and Medical Education, 2004.
2. Bowels N. The Delphi technique. RCNi 1999;13:32-6
3. Rhee JS, Weaver EM, Park SS, Baker SR, Hilger PA, Kriet JD, et al. Clinical consensus statement: Diagnosis and management of nasal valve compromise. Otolaryngol Head Neck Surg 2010;143:48-59.
4. Akins RB, Tolsol H, Cole BR. Stability of response characteristics of a Delphi panel: application bootstrap data expansion. BMC Med Res Methodol 2005;5:37.
5. Mdege ND, Raistrick D, Johnson G. Medical specialists' views on the impact of reducing alcohol consumption on the prognosis of, and risk of, hospital admission due to specific medical conditions: results from a Delphi survey. J Eval Clin Pract 2014;20:100-10.
6. Alahlafi A, Burge S. What should undergraduate medical students know about psoriasis? Involving patients in curriculum development: modified Delphi technique. BMJ 2005;330:633-6.
7. Ludwig L, Starr S. Library as place: results of Delphi study. J Med Libr Assoc 2005;93:315-26.
8. De Villiers MR, De Villiers PJ, Kent AP. The Delphi technique in health sciences education research. Med Teach 2005;27:639-43.
9. Linstone HA. Eight basic pitfalls: a checklist. (Accessed April 4, 2016, at www.is.njit.edu/pubs/delphibook/ch8.pdf).
10. Stuter LM. Lynn's education and research network: about consensus and facilitation; 2005. (Accessed April 4, 2016, at http://www.learn-usa.com/transformation_process/%7 Econsensus.htm).
11. Thakur M. The impact of ranking systems on higher education and its stakeholders. J Institut Res 2007;13:83-96.
12. De Witte K, Hudrlikova L. What about excellence in teaching? A benevolent ranking of universities. Scientometrics 2013;96:337-64.
13. Bengoetxea E, Buela-Casal G. The new multidimensional and user-driven higher education ranking concept of the European Union. Int J Clin Health Psychol 2013;13:67-73.
14. Lerner AL. A strategic planning primer for higher education; 1999. (Accessed April 12, 2016, at http://grapevine.laspositascollege.edu/academicservices/d ocuments/InstitutionalEffectiveness/CSUStrategicPlannin gReport.pdf).
15. Rankings and Accountability in Higher Education, Uses and Misuses. (Accessed April 15, 2016, at http://unesdoc.unesco.org/images/0022/002207/220789e. pdf).
16. Prathap G. Ranking of Indian medical colleges for their research performance during. Ann Lib Inform Stud 2011;58:203-10.
17. Mullan F, Chen C, Petterson S. Kolsky G, Spagnola M. The Social Mission of Medical Education: Ranking the Schools. Ann Intern Med 2010;152:804-11.
Rita Mojtahedzadeh1, Aeen Mohammadi1, Noushin Kohan2, Mitra Gharib1, and Mitra Zolfaghari3
1 Department of E-learning in Medical Education, Virtual School, Center of Excellence for E-learning in Medical Education,
Tehran University of Medical Sciences, Tehran, Iran
2 Department of Medical Education, School of Medicine, Tehran University of Medical Sciences, Tehran, Iran
3 Department of E-learning in Medical Education, Virtual School, Nursing and Midwifery Care Research Center,
Tehran University of Medical Sciences, Tehran, Iran
Received: 20 Jan. 2015; Accepted: 28 Nov. 2015
Corresponding Author: A. Mohammadi
Department of E-learning in Medical Education, Virtual School, Tehran University of Medical Sciences, Tehran, Iran
Tel: +98 21 88978088, Fax: +98 21 88969594, E-mail address: [email protected]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright Tehran University of Medical Sciences Publications 2016