Purpose This study evaluated the Dr Lee Jong-wook Fellowship Program’s impact on Tanzania’s health workforce, focusing on relevance, effectiveness, efficiency, impact, and sustainability in addressing healthcare gaps.
Methods A mixed-methods research design was employed. Data were collected from 97 out of 140 alumni through an online survey, 35 in-depth interviews, and one focus group discussion. The study was conducted from November to December 2023 and included alumni from 2009 to 2022. Measurement instruments included structured questionnaires for quantitative data and semi-structured guides for qualitative data. Quantitative analysis involved descriptive and inferential statistics (Spearman’s rank correlation, non-parametric tests) using Python ver. 3.11.0 and Stata ver. 14.0. Thematic analysis was employed to analyze qualitative data using NVivo ver. 12.0.
Results Findings indicated high relevance (mean=91.6, standard deviation [SD]=8.6), effectiveness (mean=86.1, SD=11.2), efficiency (mean=82.7, SD=10.2), and impact (mean=87.7, SD=9.9), with improved skills, confidence, and institutional service quality. However, sustainability had a lower score (mean=58.0, SD=11.1), reflecting challenges in follow-up support and resource allocation. Effectiveness strongly correlated with impact (ρ=0.746, P<0.001). The qualitative findings revealed that participants valued tailored training but highlighted barriers, such as language challenges and insufficient practical components. Alumni-led initiatives contributed to knowledge sharing, but limited resources constrained sustainability.
Conclusion The Fellowship Program enhanced Tanzania’s health workforce capacity, but it requires localized curricula and strengthened alumni networks for sustainability. These findings provide actionable insights for improving similar programs globally, confirming the hypothesis that tailored training positively influences workforce and institutional outcomes.
Purpose To generate Cronbach’s alpha and further mixed methods construct validity evidence for the Blended Learning Usability Evaluation–Questionnaire (BLUE-Q).
Methods Forty interprofessional clinicians completed the BLUE-Q after finishing a 3-month long blended learning professional development program in Ontario, Canada. Reliability was assessed with Cronbach’s α for each of the 3 sections of the BLUE-Q and for all quantitative items together. Construct validity was evaluated through the Grand-Guillaume-Perrenoud et al. framework, which consists of 3 elements: congruence, convergence, and credibility. To compare quantitative and qualitative results, descriptive statistics, including means and standard deviations for each Likert scale item of the BLUE-Q were calculated.
Results Cronbach’s α was 0.95 for the pedagogical usability section, 0.85 for the synchronous modality section, 0.93 for the asynchronous modality section, and 0.96 for all quantitative items together. Mean ratings (with standard deviations) were 4.77 (0.506) for pedagogy, 4.64 (0.654) for synchronous learning, and 4.75 (0.536) for asynchronous learning. Of the 239 qualitative comments received, 178 were identified as substantive, of which 88% were considered congruent and 79% were considered convergent with the high means. Among all congruent responses, 69% were considered confirming statements and 31% were considered clarifying statements, suggesting appropriate credibility. Analysis of the clarifying statements assisted in identifying 5 categories of suggestions for program improvement.
Conclusion The BLUE-Q demonstrates high reliability and appropriate construct validity in the context of a blended learning program with interprofessional clinicians, making it a valuable tool for comprehensive program evaluation, quality improvement, and evaluative research in health professions education.
Purpose The primary aim of this study is to validate the Blended Learning Usability Evaluation–Questionnaire (BLUE-Q) for use in the field of health professions education through a Bayesian approach. As Bayesian questionnaire validation remains elusive, a secondary aim of this article is to serve as a simplified tutorial for engaging in such validation practices in health professions education.
Methods A total of 10 health education-based experts in blended learning were recruited to participate in a 30-minute interviewer-administered survey. On a 5-point Likert scale, experts rated how well they perceived each item of the BLUE-Q to reflect its underlying usability domain (i.e., effectiveness, efficiency, satisfaction, accessibility, organization, and learner experience). Ratings were descriptively analyzed and converted into beta prior distributions. Participants were also given the option to provide qualitative comments for each item.
Results After reviewing the computed expert prior distributions, 31 quantitative items were identified as having a probability of “low endorsement” and were thus removed from the questionnaire. Additionally, qualitative comments were used to revise the phrasing and order of items to ensure clarity and logical flow. The BLUE-Q’s final version comprises 23 Likert-scale items and 6 open-ended items.
Conclusion Questionnaire validation can generally be a complex, time-consuming, and costly process, inhibiting many from engaging in proper validation practices. In this study, we demonstrate that a Bayesian questionnaire validation approach can be a simple, resource-efficient, yet rigorous solution to validating a tool for content and item-domain correlation through the elicitation of domain expert endorsement ratings.
Purpose The Dr. LEE Jong-wook Fellowship Program, established by the Korea Foundation for International Healthcare (KOFIH), aims to strengthen healthcare capacity in partner countries. The aim of the study was to develop new performance evaluation indicators for the program to better assess long-term educational impact across various courses and professional roles.
Methods A 3-stage process was employed. First, a literature review of established evaluation models (Kirkpatrick’s 4 levels, context/input/process/product evaluation model, Organization for Economic Cooperation and Development Assistance Committee criteria) was conducted to devise evaluation criteria. Second, these criteria were validated via a 2-round Delphi survey with 18 experts in training projects from May 2021 to June 2021. Third, the relative importance of the evaluation criteria was determined using the analytic hierarchy process (AHP), calculating weights and ensuring consistency through the consistency index and consistency ratio (CR), with CR values below 0.1 indicating acceptable consistency.
Results The literature review led to a combined evaluation model, resulting in 4 evaluation areas, 20 items, and 92 indicators. The Delphi surveys confirmed the validity of these indicators, with content validity ratio values exceeding 0.444. The AHP analysis assigned weights to each indicator, and CR values below 0.1 indicated consistency. The final set of evaluation indicators was confirmed through a workshop with KOFIH and adopted as the new evaluation tool.
Conclusion The developed evaluation framework provides a comprehensive tool for assessing the long-term outcomes of the Dr. LEE Jong-wook Fellowship Program. It enhances evaluation capabilities and supports improvements in the training program’s effectiveness and international healthcare collaboration.
Purpose This study aimed to identify the effects of a 12-week interprofessional simulation program, operated between February 2020 and January 2021, on the patient safety competencies of healthcare professionals in Switzerland.
Methods The simulation training was based on 2 scenarios of hospitalized patients with septic shock and respiratory failure, and trainees were expected to demonstrate patient safety competencies. A single-group before and after study was conducted after the intervention—simulation program, using a measurement tool (the Health Professional Education in Patient Safety Survey) to measure the perceived competencies of physicians, nurses, and nursing assistants. Out of 57 participants, 37 answered the questionnaire surveys 4 times: 48 hours before the training, followed by post-surveys at 24 hours, 6 weeks, and 12 weeks after the training. The linear mixed effect model was applied for the analysis.
Results Four components out of 6 perceived patient safety competencies improved at 6 weeks but returned to a similar level before training at 12 weeks. Competencies of “communicating effectively,” “managing safety risks,” “understanding human and environmental factors that influence patient safety,” and “recognize and respond to remove immediate risks of harm” are statistically significant both overall and in the comparison between before the training and 6 weeks after the training.
Conclusion Interprofessional simulation programs contributed to developing some areas of patient safety competencies of healthcare professionals, but only for a limited time. Interprofessional simulation programs should be repeated and combined with other forms of support, including case discussions and debriefings, to ensure lasting effects.
Citations
Citations to this article as recorded by
Interprofessional education interventions for healthcare professionals to improve patient safety: a scoping review Yan Jiang, Yan Cai, Xue Zhang, Cong Wang Medical Education Online.2024;[Epub] CrossRef
In addition to online questionnaires, many medical schools use supplemental evaluation tools such as focus groups to evaluate their courses. Although some benefits of using focus groups in program evaluation have been described, it is unknown whether these inperson data collection methods provide sufficient additional information beyond online evaluations to justify them. In this study, we analyze recommendations gathered from student evaluation team (SET) focus group meetings and analyzed whether these items were captured in open-ended comments within the online evaluations. Our results indicate that online evaluations captured only 49% of the recommendations identified via SETs. Surveys to course directors identified that 74% of the recommendations exclusively identified via the SETs were implemented within their courses. Our results indicate that SET meetings provided information not easily captured in online evaluations and that these recommendations resulted in actual course changes.
Citations
Citations to this article as recorded by
Assessing the Utility of Oral and Maxillofacial Surgery Posters as Educational Aids in Dental Education for Undergraduate Students: Is it Useless or Helpful? Seyed Mohammad Ali Seyedi, Navid Kazemian, Omid Alizadeh, Zeinab Mohammadi, Maryam Jamali, Reza Shahakbari, Sahand Samieirad WORLD JOURNAL OF PLASTIC SURGERY.2024; 13(1): 57. CrossRef
Grupos focais como ferramenta de pesquisa qualitativa na fisioterapia: implicações e expectativas Dartel Ferrari de Lima, Adelar Aparecido Sampaio Revista Pesquisa Qualitativa.2023; 11(27): 361. CrossRef
Educational attainment for at-risk high school students: closing the gap Karen Miner-Romanoff SN Social Sciences.2023;[Epub] CrossRef
Student evaluations of teaching and the development of a comprehensive measure of teaching effectiveness for medical schools Constantina Constantinou, Marjo Wijnen-Meijer BMC Medical Education.2022;[Epub] CrossRef
National Security Law Education in Hong Kong: Qualitative Evaluation Based on the Perspective of the Students Daniel T. L. Shek, Xiaoqin Zhu, Diya Dou, Xiang Li International Journal of Environmental Research and Public Health.2022; 20(1): 553. CrossRef
Mentoring as a transformative experience Wendy A. Hall, Sarah Liva Mentoring & Tutoring: Partnership in Learning.2021; 29(1): 6. CrossRef
Purpose Evaluating educational programs can improve the quality of education. The present study evaluated the undergraduate occupational health program at the Semnan University of Medical Sciences in Semnan, Iran, with a focus on the associations between alumni perceptions of the learning environment and the outcomes of the occupational health program. Methods: A cross-sectional questionnaire survey was conducted among alumni of the undergraduate occupational health program. We asked alumni to rate their perceptions of the items using a 4-point Likert scale. The associations between alumni perceptions of the educational program and curriculum, faculty, institutional resources, and learning outcomes were modeled and described using structural equation modeling procedures. Results: A descriptive analysis of alumni perceptions indicated low evaluations for the administrative system, practical and research-based courses, and the number of faculty members. We found that a structural model of the evaluation variables of curriculum, faculty qualifications, and institutional resources significantly predicted undergraduate educational outcomes. The curriculum had direct and indirect effects on learning outcomes, mediated by faculty. Conclusion: The findings of our study highlight the usefulness of the structural equation modeling approach for examining links between variables related to the learning process and learning outcomes. Surveys of alumni can provide data for reassessing the learning environment in the light of the professional competencies needed for occupational health graduates.
Citations
Citations to this article as recorded by
Integrated-Based Curriculum of Pharmaceutical Dosage Forms (ICPDF): What Factors Affect the Learning Outcome Attainment? Anis Yohana Chaerunisaa, Akhmad Habibi, Muhaimin Muhaimin, Mailizar Mailizar, Tommy Tanu Wijaya, Ahmad Samed Al-Adwan International Journal of Environmental Research and Public Health.2023; 20(5): 4272. CrossRef
Purpose Medical professionals from Korea and Laos have been working together to develop a continuing professional development training program covering the major clinical fields of primary care. This study aimed to evaluate the effectiveness of the program from 2013 to 2014 using the Kirkpatrick model. Methods: A questionnaire was used to evaluate the reaction of the trainees, and the trainers assessed the level of trainees’ performance at the beginning and the end of each clinical section. The transfer (behavioral change) of the trainees was evaluated through the review of medical records written by the trainees before and after the training program. Results: The trainees were satisfied with the training program, for which the average score was 4.48 out of 5.0. The average score of the trainees’ performance at the beginning was 2.39 out of 5.0, and rose to 3.88 at the end of each section. The average score of the medical records written before the training was 2.92 out of 5.0, and it rose to 3.34 after the training. The number of patient visits to the district hospitals increased. Conclusion: The continuing professional development training program, which was planned and implemented with the full engagement and responsibility of Lao health professionals, proved to be effective.
Citations
Citations to this article as recorded by
Practicalities and dichotomies of education policy and practice of higher education in the Golden Triangle Area (Southeast Asia): Implications for international development Shine Wanna Aung, Than Than Aye Policy Futures in Education.2024; 22(7): 1421. CrossRef
Evaluation of cost-effectiveness of single-credit traffic safety course based on Kirkpatrick model: a case study of Iran Mina Golestani, Homayoun Sadeghi-bazargani, Sepideh Harzand-Jadidi, Hamid Soori BMC Medical Education.2024;[Epub] CrossRef
Effectiveness of E-learning on “Sexual Health” among students of Shahid Beheshti University of Medical Sciences based on the Kirkpatrick model Zohreh Sadat Mirmoghtadaie, Zahra Mahbadi, Zinat Mahbadi Journal of Education and Health Promotion.2024;[Epub] CrossRef
The benefits and limitations of establishing the PA profession globally Arden R. Turkewitz, Jane P. Sallen, Rachel M. Smith, Kandi Pitchford, Kimberly Lay, Scott Smalley JAAPA.2024;[Epub] CrossRef
Transforming the “SEAD”: Evaluation of a Virtual Surgical Exploration and Discovery Program and its Effects on Career Decision-Making Kameela Miriam Alibhai, Patricia Burhunduli, Christopher Tarzi, Kush Patel, Christine Seabrook, Tim Brandys Journal of Surgical Education.2023; 80(2): 256. CrossRef
Evaluation of the effectiveness of a training programme for nurses regarding augmentative and alternative communication with intubated patients using Kirkpatrick's model: A pilot study Marzieh Momennasab, Fatemeh Mohammadi, Fereshteh DehghanRad, Azita Jaberi Nursing Open.2023; 10(5): 2895. CrossRef
Outcome Evaluation of a Transnational Postgraduate Capacity-Building Program Using the Objective Structured Clinical Examination Kye-Yeung Park, Hoon-Ki Park, Jwa-Seop Shin, Taejong Kim, Youngjoo Jung, Min Young Seo, Ketsomsouk Bouphavanh, Sourideth Sengchanh, Ketmany Inthachack Evaluation Review.2023; 47(4): 680. CrossRef
Developing a capacity building training model for public health managers of low and middle income countries Kritika Upadhyay, Sonu Goel, Preethi John, Sara Rubinelli PLOS ONE.2023; 18(4): e0272793. CrossRef
Implementation and evaluation of crowdsourcing in global health education Huanle Cai, Huiqiong Zheng, Jinghua Li, Chun Hao, Jing Gu, Jing Liao, Yuantao Hao Global Health Research and Policy.2022;[Epub] CrossRef
An Evaluation of the Surgical Foundations Curriculum: A National Study Ekaterina Kouzmina, Stephen Mann, Timothy Chaplin, Boris Zevin Journal of Surgical Education.2021; 78(3): 914. CrossRef
Surgical data strengthening in Ethiopia: results of a Kirkpatrick framework evaluation of a data quality intervention Sehrish Bari, Joseph Incorvia, Katherine R. Iverson, Abebe Bekele, Kaya Garringer, Olivia Ahearn, Laura Drown, Amanu Aragaw Emiru, Daniel Burssa, Samson Workineh, Ephrem Daniel Sheferaw, John G. Meara, Andualem Beyene Global Health Action.2021;[Epub] CrossRef
Evaluation of a Neonatal Resuscitation Training Programme for Healthcare Professionals in Zanzibar, Tanzania: A Pre-post Intervention Study Xiang Ding, Li Wang, Mwinyi I. Msellem, Yaojia Hu, Jun Qiu, Shiying Liu, Mi Zhang, Lihui Zhu, Jos M. Latour Frontiers in Pediatrics.2021;[Epub] CrossRef
Portfolios with Evidence of Reflective Practice Required by Regulatory Bodies: An Integrative Review Marco Zaccagnini, Patricia A. Miller Physiotherapy Canada.2021;[Epub] CrossRef
Evaluation of a training program on primary eye care for an Accredited Social Health Activist (ASHA) in an urban district Pallavi Shukla, Praveen Vashist, SurajSingh Senjam, Vivek Gupta Indian Journal of Ophthalmology.2020; 68(2): 356. CrossRef
Micro-feedback skills workshop impacts perceptions and practices of doctoral faculty Najma Baseer, James Degnan, Mandy Moffat, Usman Mahboob BMC Medical Education.2020;[Epub] CrossRef
Residents working with Médecins Sans Frontières: training and pilot evaluation Alba Ripoll-Gallardo, Luca Ragazzoni, Ettore Mazzanti, Grazia Meneghetti, Jeffrey Michael Franc, Alessandro Costa, Francesco della Corte Scandinavian Journal of Trauma, Resuscitation and Emergency Medicine.2020;[Epub] CrossRef
Medical education in Laos Timothy Alan Wittick, Ketsomsouk Bouphavanh, Vannyda Namvongsa, Amphay Khounthep, Amy Gray Medical Teacher.2019; 41(8): 877. CrossRef
Evaluation of the effectiveness of a first aid health volunteers’ training programme using Kirkpatrick’s model: A pilot study Fatemeh Vizeshfar, Marzieh Momennasab, Shahrzad Yektatalab, Mohamad Taghi Iman Health Education Journal.2018; 77(2): 190. CrossRef
Evaluation of a consulting training course for international development assistance for health Pan Gao, Hao Xiang, Suyang Liu, Yisi Liu, Shengjie Dong, Feifei Liu, Wenyuan Yu, Xiangyu Li, Li Guan, Yuanyuan Chu, Zongfu Mao, Shu Chen, Shenglan Tang BMC Medical Education.2018;[Epub] CrossRef
Empowering the Filipino Physician through Continuing Professional Development in the Philippines: Gearing towards ASEAN Harmonization and Globalization Maria Minerva P Calimag Journal of Medicine, University of Santo Tomas.2018; 2(1): 121. CrossRef
Purpose The tutorial group effectiveness instrument was developed to provide objective information on the effectiveness of small groups. Student perception of small group effectiveness during the PBL process has not been previously studied in Xavier University School of Medicine. Hence the present study was carried out.
Methods The study was conducted among the second and third semester undergraduate medical students during the last week of September 2013, in Xavier University School of Medicine, Aruba, Kingdom of the Netherlands. Students were informed about the objectives of the study and invited to participate after obtaining written, informed consent. Demographic information like gender, age, nationality and whether the respondent had been exposed to PBL before joining the institution were noted. Student perception about small group effectiveness was studied by noting their degree of agreement with a set of 19 statements using a Likert type scale.
Results Thirty four of the 37 (91.9%) second and third semester medical students participated in the study. The mean cognitive score was 3.76 while the mean motivational and demotivational scores were 3.65 and 2.51 respectively. The median cognitive category score was 27 (maximum score 35) while the motivation score was 26 (maximum score 35) and the demotivational score was 12 (maximum being 25). There was no significant difference in scores according to respondents’ demographic characteristics.
Conclusion Student perception about small group effectiveness was positive. Since most medical schools all over the world already have or are introducing PBL as a learning modality, Tutorial Group Effectiveness Instrument can provide valuable information about small group functioning during PBL sessions.
Citations
Citations to this article as recorded by
Relationship of Prior Knowledge and Scenario Quality With the Effectiveness of Problem-based Learning Discussion among Medical Students of Universitas Malikussaleh, Aceh, Indonesia Mulyati Sri Rahayu, Sri Wahyuni, Yuziani Yuziani Malaysian Journal of Medicine and Health Sciences.2023; 19(4): 15. CrossRef
Should the PBL tutor be present? A cross-sectional study of group effectiveness in synchronous and asynchronous settings Samuel Edelbring, Siw Alehagen, Evalotte Mörelius, AnnaKarin Johansson, Patrik Rytterström BMC Medical Education.2020;[Epub] CrossRef
Initiating small group learning in a Caribbean medical school P. Ravi Shankar Journal of Educational Evaluation for Health Professions.2015; 12: 10. CrossRef
Aprendizagem Baseada em Problemas na Graduação Médica – Uma Revisão da Literatura Atual Luciana Brosina de Leon, Fernanda de Quadros Onófrio Revista Brasileira de Educação Médica.2015; 39(4): 614. CrossRef
Assessing the Effectiveness of Problem-Based Learning of Preventive Medicine Education in China Xiaojie Ding, Liping Zhao, Haiyan Chu, Na Tong, Chunhui Ni, Zhibin Hu, Zhengdong Zhang, Meilin Wang Scientific Reports.2014;[Epub] CrossRef
Medical school administrators, educators, and other key personnel must often make difficult choices regarding the creation, retention, modification, or termination of the various programs that take place at their institutions. Program evaluation is a data-driven strategy to aide decision-makers in determining the most appropriate outcome for programs within their purview. The purpose of this brief article is to describe one program evaluation model, the utilization-focused approach. In particular, we address the focus of this model, the personal factor, the role of the evaluator, and the evaluation process. Based on the flexibility of this model as well as its focus on stakeholder involvement, we encourage readers to consider the utilization-focused approach when evaluating programs.
Citations
Citations to this article as recorded by
Opportunities for Pedagogical Change in Turkish Medical Education Revealed in the Wake of the COVID-19 Pandemic Umit Kartoglu, Sevgi Turan, Alp Ergör, Dilek Aslan, Gülriz Erişgen, Duygu Fındık, Özlem Kayım Yıldız, Thomas C. Reeves Teaching and Learning in Medicine.2024; 36(4): 488. CrossRef
Teaching and Facilitation Course for Family as Faculty: Preparing Families to be Faculty Partners in Healthcare Education Clara Ho, Ami Goulden, Darlene Hubley, Keith Adamson, Jean Hammond, Adrienne Zarem Clinical Social Work Journal.2024; 52(1): 23. CrossRef
Evaluating competency-based medical education: a systematized review of current practices Nouf Sulaiman Alharbi BMC Medical Education.2024;[Epub] CrossRef
Gathering Trainee Feedback to Improve Programs With Low Annual ACGME Survey Content Area Compliance: A Pilot Study Mara M. Hoffert, Leslie Pfeiffer, Molly Hepke, Wendy Brink, Jennifer Newman, Karla D. Passalacqua, Kimberly Baker-Genaw Academic Medicine.2024; 99(4): 419. CrossRef
Strategies to foster stakeholder engagement in residency coaching: a CFIR-Informed qualitative study across diverse stakeholder groups Marzena Sasnal, Rachel M. Jensen, Uyen T. Mai, Carl A. Gold, Aussama K. Nassar, James R. Korndorffer, Arden M. Morris, Rebecca K. Miller-Kuhlmann Medical Education Online.2024;[Epub] CrossRef
Teste de Progresso: a percepção do discente de Medicina Marlene Moraes Rosa Chinelato, Jose Eduardo Martinez, Gisele Regina de Azevedo Revista Brasileira de Educação Médica.2022;[Epub] CrossRef
Pandemi Döneminde Tıp Eğitimini Sürdürmek: Giresun Üniversitesi Tıp Fakültesi Deneyimi Hülya AKAN, Berkan ŞAHİN, Murat USTA, Özkan ÖZAY, Hakan YÜZÜAK, Ural OĞUZ Tıp Eğitimi Dünyası.2021; 20(60-1): 54. CrossRef
Ongoing Value and Practice Improvement Outcomes from Pediatric Palliative Care Education: The Quality of Care Collaborative Australia Penelope J Slater, Caroline J Osborne, Anthony R Herbert Advances in Medical Education and Practice.2021; Volume 12: 1189. CrossRef
Faculty Feedback Program Evaluation in CIMS Multan, Pakistan Ambreen Shabbir, Hina Raja, Anjum A Qadri, Muhammad Hisaan Anjum Qadri Cureus.2020;[Epub] CrossRef
A Guide to Evaluation of Quality Improvement and Patient Safety Educational Programs: Lessons From the VA Chief Resident in Quality and Safety Program Rebecca L. Butcher, Kathleen L. Carluzzo, Bradley V. Watts, Karen E. Schifferdecker American Journal of Medical Quality.2019; 34(3): 251. CrossRef
Design and Content Validation of Three Setting-Specific Assessment Tools for Advanced Pharmacy Practice Experiences Eric H. Gilliam, Jason M. Brunner, Wesley Nuffer, Toral C. Patel, Megan E. Thompson American Journal of Pharmaceutical Education.2019; 83(9): 7067. CrossRef
Is it a match? a novel method of evaluating medical school success Leslie L. Chang, Alisa Nagler, Mariah Rudd, Colleen O’Connor Grochowski, Edward G. Buckley, Saumil M. Chudgar, Deborah L. Engle Medical Education Online.2018; 23(1): 1432231. CrossRef
Evaluation of medical ethics doctoral program; a utilization-focused approach Leila Afshar, Seyed Ziaedin Tabei, Mohammad Hosseinzade International Journal of Ethics Education.2018; 3(1): 89. CrossRef
How we conduct ongoing programmatic evaluation of our medical education curriculum Kelly Karpa, Catherine S. Abendroth Medical Teacher.2012; 34(10): 783. CrossRef