The Karma system is currently undergoing maintenance (Monday, January 29, 2018).
The maintenance period has been extended to 8PM EST.

Karma Credits will not be available for redeeming during maintenance.
Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 28.09.15 in Vol 17, No 9 (2015): September

This paper is in the following e-collection/theme issue:

    Original Paper

    Moving Knowledge Acquisition From the Lecture Hall to the Student Home: A Prospective Intervention Study

    1University Medical Centre Göttingen, Clinic for Cardiology and Pneumology, Georg-August University Göttingen, Göttingen, Germany

    2Health Behaviour Research Centre, Department of Epidemiology & Public Health, University College London, London, United Kingdom

    3Department of Clinical, Educational and Health Psychology, University College London, London, United Kingdom

    4University Medical Centre Göttingen, Clinic for Haematology and Oncology, Georg-August University Göttingen, Göttingen, Germany

    5University Medical Centre Hamburg-Eppendorf, Department of Legal Medicine, Hamburg, Germany

    Corresponding Author:

    Tobias Raupach, MD, MME

    University Medical Centre Göttingen

    Clinic for Cardiology and Pneumology

    Georg-August University Göttingen

    Robert-Koch-Strasse 40

    Göttingen, D-37075

    Germany

    Phone: 49 55139 ext 8922

    Fax:49 5513920900

    Email:


    ABSTRACT

    Background: Podcasts are popular with medical students, but the impact of podcast use on learning outcomes in undergraduate medical education has not been studied in detail.

    Objective: Our aim was to assess the impact of podcasts accompanied by quiz questions and lecture attendance on short- and medium-term knowledge retention.

    Methods: Students enrolled for a cardio-respiratory teaching module were asked to prepare for 10 specific lectures by watching podcasts and submitting answers to related quiz questions before attending live lectures. Performance on the same questions was assessed in a surprise test and a retention test.

    Results: Watching podcasts and submitting answers to quiz questions (versus no podcast/quiz use) was associated with significantly better test performance in all items in the surprise test and 7 items in the retention test. Lecture attendance (versus no attendance) was associated with higher test performance in 3 items and 1 item, respectively. In a linear regression analysis adjusted for age, gender, and overall performance levels, both podcast/quiz use and lecture attendance were significant predictors of student performance. However, the variance explained by podcast/quiz use was greater than the variance explained by lecture attendance in the surprise test (38.7% vs 2.2%) and retention test (19.1% vs 4.0%).

    Conclusions: When used in conjunction with quiz questions, podcasts have the potential to foster knowledge acquisition and retention over and above the effect of live lectures.

    J Med Internet Res 2015;17(9):e223

    doi:10.2196/jmir.3814

    KEYWORDS



    Introduction

    In recent years, there has been increasing interest in the use of podcasts as means of knowledge transmission. In broad terms, podcasts can generally be described as audio and/or video files that can be played back on various electronic devices including tablets and smartphones. In fact, the word “podcast,” first used in 2004, is a portmanteau created from the name of one particular device (the iPod) and the word “broadcast.” There is no uniform consensus as to what format or content is required for an electronic source to be called a “podcast.” As a consequence, anything from straight-forward recordings of lectures or conference presentations to complex animated films can be referred to as podcasts. However, some authors have used the term “vodcast” to describe online material containing videos [1] and “enhanced podcast” for audio material supplemented with still images [2]. One common feature of all these formats is that they can be used in an asynchronous manner (ie, at any time, independent of lecture hours).

    Within 10 years of their invention, technologies to capture lectures and make them available to students have been embraced by medical teachers involved in both undergraduate and continuing medical education. At the same time, both massive open online courses [3] and scholarly journals [4] now offer a wide range of options to view or listen to material online. User satisfaction is generally high [5], but there is a paucity of data linking podcast use to actual learning outcome. This is in contrast with recent calls for medical school lectures to be moved to online platforms altogether so that classroom time may be used for more efficient teaching activities [6,7]. The underlying assumption is that students viewing course material in preparation of a lecture will retain the content. To our knowledge, this hypothesis has not been tested so far.

    A PubMed search combining “medical education” with the terms “podcasts,” “lecture video,” “online lecture,” or “streaming lecture” (search date March 6, 2014) yielded 357 unique citations, and 6 additional articles were identified from reference lists and by contacting experts in the field. Only 78 out of these 363 papers had a specific focus on podcasts. Only 55 of these presented original data, and about half of these (n=27) were related to undergraduate medical education. While half of these (n=13) just reported usage patterns and student satisfaction with podcasts, only 14 original articles assessed the association between podcast use and learning outcome (6 randomized controlled trials, 7 prospective studies, and 1 retrospective analysis). Notably, none of these studies addressed podcast use for preparatory purposes. Instead, podcasts were used to either completely replace or supplement live lectures. In summary, there is currently no scientific data on the effectiveness of using podcasts to stimulate student learning prior to attending a lecture.

    The aims of this study were to assess the impact of preparatory podcast use in conjunction with quiz questions versus lecture attendance on short-term and medium-term knowledge retention, and identify significant predictors of short-term and medium-term knowledge retention.

    It was hypothesized that students engaging with the material presented in podcasts and submitting answers to quiz questions would retain significantly more knowledge than students not using podcasts. With regard to the second study aim, it was hypothesized that podcast/quiz use would be at least as effective in promoting short-term and medium-term knowledge retention as lecture attendance.

    This study did not address any specific psychological framework underlying a potential effect of podcast/quiz use. Instead, it focused on effects elicited by one particular teaching intervention (ie, podcasts and quiz questions) in a “real-world” educational setting.


    Methods

    Study Design

    This study was conducted at Göttingen Medical School. Like most German medical schools, it offers a 6-year undergraduate curriculum comprising 2 preclinical years and 3 clinical years, followed by a practice year. This prospective trial included a cohort of fourth-year medical students who were enrolled in a 6-week cardio-respiratory module in winter term 2013/14. In the preceding summer term, all 37 lectures held during the 6-week module had been recorded using Camtasia Studio 7 (TechSmith). The resulting videos featured the presentation slides used and the lecturer’s voice (duration: 35-45 minutes; format: MP4). Following the summer term 2013, the material was reviewed, and the best 10 lecture recordings with regard to sound and image quality were selected to be used in this study. Lecturers were asked to identify key aspects with particular relevance for general internal medicine and to draft free-text questions addressing that content (Table 1).

    In winter term 2013/14, students enrolled in our module were provided with online access to the 10 selected videos for a period of 7 days before the respective live lectures. A free-text quiz question was linked to each podcast, and students were invited to submit their answers via email until the night before the live lecture. On the day of the lecture, the principal investigator (TR) revealed the correct answer to the entire class and also projected a 1-minute clip from the podcast containing the answer. He then raffled a book voucher (€20) among all students who had submitted a correct answer.

    As part of an e-learning session in the final week of the module, students were invited to complete the same 10 quiz questions that had been provided with the podcasts (surprise test). In this session, students were also asked to indicate which lectures they had attended. In order to assess long-term retention, students were invited to answer the same 10 questions again 2 months later in an unannounced retention test. The study outline is summarized in Figure 1.

    Table 1. Key aspects covered in podcast lectures and student performance in quizzes as well as in the surprise and retention tests.
    View this table
    Figure 1. Study outline (numbers in boxes correspond to the 10 lectures used for this study, and vertical dotted lines indicate the date on which live lectures were held). For each lecture, podcasts were available over a period of 7 days leading up to the live lecture. During this time, students were invited to submit their quiz answers.
    View this figure

    Student Enrollment and Data Collection

    Both the surprise test and the retention test were timed to coincide with scheduled e-learning activity in our institution’s computer facilities. Both tests were unannounced in order to avoid confounding by specific preparation, and a time limit of 15 minutes was set for the completion of the 10 quiz questions in both tests. At the beginning of each e-learning session, the study rationale was explained and students were asked to provide written consent to have their data analyzed for study purposes. A total of 10 book vouchers (€20) were raffled among all participants at both the surprise test and the retention test, regardless of test performance.

    Questionnaires were created with EvaSys (Electric Paper). In the surprise test, students were asked to provide their age and gender and to indicate whether they had attended each of the 10 lectures and which podcasts they had watched. The number of quiz answers submitted during the module was derived from the emails sent to the module’s administrative staff. The potential impact of recall bias and/or podcast use without answering quiz questions was assessed by comparing the number of students indicating they had watched a particular podcast with the number of students who had submitted an answer to the corresponding quiz question. In order to adjust the analyses for student performance levels, the percent score achieved by each student in the summative end-of-module examination was also obtained (Figure 1). This examination consisted of 25 multiple choice questions addressing factual knowledge on cardiology and pneumology but specifically excluding the content covered by quiz questions as the latter focused on more complex aspects while multiple choice questions were designed to assess basic factual knowledge.

    Marking of Quiz Answers

    The marking procedure was identical for all three time points (during the module—only students who had submitted an answer via email; surprise test and retention test—all students entering data and consenting to have their data analyzed). After agreeing on corrects answers, 2 raters (TR and CG) independently marked all answers as correct (1) or incorrect (0). Inconsistencies were resolved by discussion. In addition to marking each single question, a sum score (0-10) was calculated, reflecting student attainment in quizzes throughout the module, in the surprise test, and in the retention test.

    Statistical Analysis

    Unique student identifier codes were used to merge data collected in the surprise and retention tests as well as examination results and data on podcast/quiz use. Data analysis was performed using SPSS Statistics 21 (IBM Corporation). Inter-rater agreement of the marking procedure was assessed by calculating kappa and internal consistency of both tests was assessed by calculating Cronbach alpha.

    In order to assess the impact of podcast/quiz use and lecture attendance on short-term and medium-term knowledge retention, the percentage of students providing a correct answer to each question in the surprise and retention tests was calculated. Proportions of students who had/had not submitted a correct quiz answer during the module and those who had/had not attended the corresponding lecture were compared by chi-square tests. Multivariate logistic regression models were run for each of the 10 items with the answer in the surprise/retention test as the dependent variable and controlling for age, gender, and percent score in the end-of-module examination. The comparison between podcast/quiz use and no podcast/quiz use was also adjusted for lecture attendance. Likewise, podcast/quiz use was adjusted for when assessing the impact of lecture attendance on test performance.

    A multivariate linear regression analysis was run to identify significant predictors of short-term and medium-term knowledge retention. The dependent variable was the sum score in the surprise/retention test. Age, gender, and percent score in the end-of-module examination as well as the number of submitted correct quiz answers and the number of lectures attended during the module were entered as independent variables.

    Group comparisons were performed using chi-square tests (dichotomous variables) and t tests (continuous variables). Results of descriptive analyses are presented as percentages and mean with standard deviation (SD), as appropriate. Results of linear regression analyses are reported as unadjusted and adjusted beta values (95% confidence interval) and as the amount of variance explained. Significance levels were set to .05.

    Ethical Approval

    The local Institutional Review Board (application number 13/12/13) waived ethical approval as the study protocol was not deemed to represent biomedical or epidemiological research. Study participation was voluntary, and all participants signed an informed consent form before entering the study.


    Results

    Response Rate and Participant Characteristics

    Of 130 students enrolled in the module, 126 gave written consent to have their data analyzed for this study. Only students with complete data in both the surprise and the retention tests and the end-of-module examination were included in the final analysis. A total of 101 students attended both tests, but 3 of these did not take the end-of-module examination and another 31 failed to provide complete information on lecture attendance. Thus, complete data of 67 students (24.2 years [SD 2.9]; 39 female) were available. Of these, 34 had submitted at least one correct answer during the module (mean 3.5 [SD 2.6]). On average, students had attended 7.8 (SD 2.3) live lectures. The percentage of students who recalled watching the podcast among those who had submitted a quiz answer was over 80% for all items, suggesting podcast use was not hugely underreported. On the contrary, the proportion of students who had not submitted a quiz answer among those who recalled watching the podcast ranged from 20% to 50%.

    Inter-rater Agreement, Item Characteristics, and Results of the Surprise and Retention Tests

    Inter-rater agreement for quiz questions and the surprise and retention tests were acceptable (kappa values were .86, .80, and .90, respectively). Cronbach alpha was .68 and .65 in the surprise and retention tests, respectively. The mean number of correct answers in these tests were 2.9 (SD 2.3) and 2.6 (SD 2.0), respectively. As shown in Table 1, performance in all test items was low to moderate. For example, one-third of students were aware that ruling out bronchial obstruction in a baseline lung function test is a prerequisite for bronchial hyper-reactivity testing and only 1 in 5 students displayed adequate knowledge on how to interpret clinical trial reports.

    Impact of Podcast Use and Lecture Attendance on Test Performance

    Figure 2 presents student performance in the surprise and retention tests as a function of podcast use and lecture attendance.

    In the surprise test, podcast use was associated with significantly better knowledge on all test items while such associations with lecture attendance were observed for only 3 items. Similarly, podcast use enhanced knowledge in the retention test for 7 items while there was no such effect of lecture attendance for 9 of 10 items. Adjusting for age, gender, and examination performance attenuated the associations, but the pattern of results was unchanged.

    In a sensitivity analysis, percentages of correct answers in the surprise and the retention tests were calculated separately for students who (Group 1) had submitted a quiz answer and recalled having watched the corresponding podcast and students who (Group 2) recalled having watched the podcast but had not submitted a quiz answer. With one exception (Item 2), test performance in the second group was similar to the performance of students who had neither submitted a correct answer nor watched the podcast. The proportion of correct answers in Group 2 was less than half of that observed in Group 1 and was 0% for 3 items in the surprise test and 4 items in the retention test (data not shown).

    Figure 2. Student performance in the surprise and retention tests (columns represent the percentage of students providing a correct answer; error bars represent standard errors). * P<.05 for direct comparison (chi-square test); § P<.05 in a logistic regression adjusted for age, gender, exam performance, and lecture attendance; # P<.05 in a logistic regression adjusted for age, gender, exam performance, and podcast use.
    View this figure

    Predictors of Short-Term and Medium-Term Knowledge Retention

    Results of the linear regression analyses are reported in Table 2. Podcast/quiz use and lecture attendance were significant predictors of student performance in both tests. However, the variance explained by podcast/quiz use was greater than the variance explained by lecture attendance in the surprise test (38.7% vs 2.2%) and retention test (19.1% vs 4.0%).

    Table 2. Predictors of student performance in the surprise and retention tests (R2, variance explained).
    View this table

    Discussion

    Principal Findings

    This is the first study to examine the impact of podcast use in conjunction with quizzes prior to lecture attendance on knowledge acquisition and retention in undergraduate medical students. Students who engaged with the material before the lecture displayed improved short-term and medium-term retention, regardless of whether they also attended the lecture. The impact of lecture attendance on knowledge retention was considerably weaker despite the correct answers and the decisive part of the podcast being presented to all students in the lecture hall. The most likely explanation for our finding is that—just like interaction during a live lecture [8]—the questions provided with preparation podcasts stirred student alertness, thus facilitating learning [9]. This notion is supported by the results of the sensitivity analysis indicating that watching podcasts without submitting an answer to the corresponding quiz question did not result in improved short-term or medium-term retention. It might be hypothesized that a similar effect could have been observed for lecture attendance if students had been asked to pay attention to a specific detail during the lecture and submit the answer to a related question afterward. However, according to the rationale outlined earlier, one potential use of podcasts could be to partially move the process of knowledge acquisition from the lecture hall to the preparation phase, thereby enabling teachers and learners to explore new and better ways to spend classroom time [10].

    Research Context

    There has been some debate about the usefulness of podcasts in medical education. While some authors regard them as “toys” [11] and have called for more research into their actual effectiveness, others have argued that students can benefit from exploring novel technologies even in the absence of randomized controlled trials demonstrating their effectiveness [12].

    The 14 published reports on the impact of podcast use on learning outcomes in undergraduate medical students vary considerably with regard to study design and outcome measure used. One retrospective analysis detected a small effect of podcast availability on national licensing examination scores that coincided with a national trend for better examination scores [13]. While 2 of the 7 prospective trials found no effect of supplemental podcasts on test scores [14,15], others did find an effect [16,17]. However, some of these effects were either assessed at a very early follow-up (ie, 5 days [18]) or confined to specific student populations, for example, non-native speakers [19]. In one study, students viewing more lectures were even found to score lower in a consecutive examination [20]. Of the 6 randomized trials published so far, 3 [21-23] found a significant effect of podcast use on student examination performance, whereas the other 3 did not [24-26].

    In our study, podcasts were used neither to replace nor supplement lectures but as a preparatory tool. In this regard, our results provide some suggestions on how this technology might be used to improve learning outcome [27] (as opposed to assessing whether it should be used at all [28]). When combined with quiz questions, the provision of podcasts led to a more favorable learning outcome than lecture attendance itself, and this effect was sustained and robust in the adjusted analysis. Given the relatively low uptake observed in our study and previous studies [20], one potential practical implication of our findings could be making the completion of a “preparatory podcast/quiz task” a requirement for course attendance.

    Strengths and Limitations

    Whereas many previous outcome studies assessed the association between podcast use and overall examination scores, the surprise and retention tests we used were created specifically for this study, and we made every effort to align test questions to the content taught in podcasts and lectures. Inter-rater agreement and internal consistency of the surprise and retention tests were acceptable, but mean scores in both tests were surprisingly low. One potential explanation for this is that these tests were formative in nature, and students might not have made full efforts to achieve a maximum number of correct answers. However, this should apply to all students (regardless of podcast/quiz use and lecture attendance), and using summative examinations would have had a confounding effect likely to mask any real effect of podcast/quiz use on knowledge levels [29,30]. Another explanation for the low overall scores observed in this study is that quiz questions were related to complex clinical content that—despite being highly relevant for medical practice—is not usually being covered in undergraduate medical textbooks. Moreover, students at our university are not used to open-ended questions as most end-of-course examinations still consist of multiple choice questions. The small amount of variance in surprise and retention test scores explained by the summative multiple choice examination (9.1% and 6.2%, respectively) can be taken as evidence of discriminant validity in that the study-related tests featuring open-ended questions assessed different types of knowledge than the multiple choice questions presented in the end-of-module examination.

    One particular strength of our study was the ability to disentangle the effects of podcast/quiz use and lecture attendance in the adjusted analyses presented in Table 2. These data suggest that following podcast use and submitting a correct answer, attending the live lecture had only limited additional benefit in terms of learning outcome. We cannot rule out the possibility that students prepared for lectures with material other than podcasts and/or quiz questions. However, given the marked performance differences between podcast/quiz users and nonusers, any effect of additional preparation would be either confined to podcast/quiz users or too small to detect in students not using podcasts/quizzes.

    We excluded a large number of students due to missing information on lecture attendance. This led to a student sample favoring slightly younger (24.2 [SD 2.9] vs 25.4 [SD 2.9]; P=.019) and slightly higher-performing students (end-of-module exam scores: 78.6% [SD 11.3] vs 72.8% [SD 13.8]; P=.012). The impact of these variables on our results within the final study sample was accounted for by adjusting our analyses accordingly. In addition to selection bias, recall bias is another potential threat to the validity of our findings. However, a great majority of students who had submitted a quiz answer also recalled having watched the corresponding podcast, rendering underreporting of podcast use unlikely. It might be hypothesized that lectures are in fact effective in helping students to acquire and retain knowledge. In order to artificially increase the effect of podcast/quiz use over that of lecture attendance, podcasts users would have had to systematically underreport lecture attendance. However, this was not the case as students submitting at least one correct quiz answer indicated to have attended significantly more lectures than students not submitting any answer: 8.7 (SD 2.0) versus 7.1 (SD 2.5); P=.006. In addition, there was a positive correlation between lecture attendance and podcast use (r=.252; P=.039), hence the need to control for lecture attendance in the analysis of podcast effectiveness and vice versa.

    An alternative approach to addressing our research question would have been to conduct a randomized controlled trial. Although this would have yielded higher internal validity, we doubt that we would have been able to restrict podcast use to a specific student group. The aim of this trial was not to test learning processes induced by the availability of podcasts and quiz questions. Instead, this study assessed the effect of one particular teaching intervention in the “real world” of undergraduate medical education. With regard to generalizability, our findings will need to be replicated in other settings. At the same time, there is no reason to believe that using podcasts supplemented with quiz questions as tools to stimulate student learning would be completely ineffective if implemented in a different medical school.

    Conclusions

    When used in conjunction with quiz questions, lecture podcasts have the potential to foster knowledge acquisition and retention over and above the effect of live lectures. Our findings might help pave the way to move knowledge acquisition from the lecture hall to the preparatory phase, thereby freeing up valuable lecture time for more effective learner-teacher interactions.

    Acknowledgments

    The authors would like to thank all medical students who agreed to take part in this study. This study was funded by a grant from the Faculty of Medicine, Georg-August-University Göttingen.

    Conflicts of Interest

    None declared.

    References

    1. Kurien G, Biron VL, Campbell C, Cote DW, Ansari K. Can a multisensory teaching approach impart the necessary knowledge, skills, and confidence in final year medical students to manage epistaxis? J Otolaryngol Head Neck Surg 2013;42:51 [FREE Full text] [CrossRef] [Medline]
    2. Thapa MM, Richardson ML. Dissemination of radiological information using enhanced podcasts. Acad Radiol 2010 Mar;17(3):387-391. [CrossRef] [Medline]
    3. Liyanagunawardena TR, Williams SA. Massive open online courses on health and medicine: review. J Med Internet Res 2014;16(8):e191 [FREE Full text] [CrossRef] [Medline]
    4. Björk B. A study of innovative features in scholarly open access journals. J Med Internet Res 2011;13(4):e115 [FREE Full text] [CrossRef] [Medline]
    5. Childers RE, Dattalo M, Christmas C. Podcast pearls in residency training. Ann Intern Med 2014 Jan 7;160(1):70. [CrossRef] [Medline]
    6. Kalmey J. Stop wasting classroom time: embrace the podcast and use the "lecture" to enhance learning. Acad Med 2013 Aug;88(8):1054. [CrossRef] [Medline]
    7. Prober CG, Heath C. Lecture halls without lectures--a proposal for medical education. N Engl J Med 2012 May 3;366(18):1657-1659. [CrossRef] [Medline]
    8. Nelson C, Hartling L, Campbell S, Oswald AE. The effects of audience response systems on learning outcomes in health professions education. A BEME systematic review: BEME Guide No. 21. Med Teach 2012;34(6):e386-e405. [CrossRef] [Medline]
    9. Deslauriers L, Schelew E, Wieman C. Improved learning in a large-enrollment physics class. Science 2011 May 13;332(6031):862-864 [FREE Full text] [CrossRef] [Medline]
    10. Kanter S. To be there or not to be there: is attendance really the question? Acad Med 2012 Jun;87(6):679. [CrossRef] [Medline]
    11. Zanussi L, Paget M, Tworek J, McLaughlin K. Podcasting in medical education: can we turn this toy into an effective learning tool? Adv Health Sci Educ Theory Pract 2012 Oct;17(4):597-600. [CrossRef] [Medline]
    12. White J, Sharma N. Podcasting: a technology, not a toy. Adv Health Sci Educ Theory Pract 2012 Oct;17(4):601-603. [CrossRef] [Medline]
    13. Bridge P, Jackson M, Robinson L. The effectiveness of streaming video on medical student learning: a case study. Med Educ Online 2009;14:11 [FREE Full text] [CrossRef] [Medline]
    14. Beale E, Tarwater PM, Lee VH. A retrospective look at replacing face-to-face embryology instruction with online lectures in a human anatomy course. Anat Sci Educ 2014;7(3):234-241. [CrossRef] [Medline]
    15. Bacro Thierry RH, Gebregziabher M, Fitzharris TP. Evaluation of a lecture recording system in a medical curriculum. Anat Sci Educ 2010;3(6):300-308. [CrossRef] [Medline]
    16. Evans DJR. Using embryology screencasts: a useful addition to the student learning experience? Anat Sci Educ 2011;4(2):57-63. [CrossRef] [Medline]
    17. Burnette K, Ramundo M, Stevenson M, Beeson MS. Evaluation of a web-based asynchronous pediatric emergency medicine learning tool for residents and medical students. Acad Emerg Med 2009 Dec;16 Suppl 2:S46-S50 [FREE Full text] [CrossRef] [Medline]
    18. O'Neill E, Power A, Stevens N, Humphreys H. Effectiveness of podcasts as an adjunct learning strategy in teaching clinical microbiology among medical students. J Hosp Infect 2010 May;75(1):83-84. [CrossRef] [Medline]
    19. Shaw GP, Molnar D. Non-native English language speakers benefit most from the use of lecture capture in medical school. Biochem Mol Biol Educ 2011;39(6):416-420. [CrossRef] [Medline]
    20. McNulty J, Hoyt A, Gruener G, Chandrasekhar A, Espiritu B, Price R, et al. An analysis of lecture video utilization in undergraduate medical education: associations with performance in the courses. BMC Med Educ 2009;9:6 [FREE Full text] [CrossRef] [Medline]
    21. Bhatti I, Jones K, Richardson L, Foreman D, Lund J, Tierney G. E-learning vs lecture: which is the best approach to surgical teaching? Colorectal Dis 2011 Apr;13(4):459-462. [CrossRef] [Medline]
    22. Mehrpour SR, Aghamirsalim M, Motamedi SMK, Ardeshir LF, Sorbi R. A supplemental video teaching tool enhances splinting skills. Clin Orthop Relat Res 2013 Feb;471(2):649-654 [FREE Full text] [CrossRef] [Medline]
    23. Hansen M, Oosthuizen G, Windsor J, Doherty I, Greig S, McHardy K, et al. Enhancement of medical interns' levels of clinical skills competence and self-confidence levels via video iPods: pilot randomized controlled trial. J Med Internet Res 2011;13(1):e29 [FREE Full text] [CrossRef] [Medline]
    24. Spickard A, Alrajeh N, Cordray D, Gigante J. Learning about screening using an online or live lecture: does it matter? J Gen Intern Med 2002 Jul;17(7):540-545 [FREE Full text] [Medline]
    25. Schreiber B, Fukuta J, Gordon F. Live lecture versus video podcast in undergraduate medical education: A randomised controlled trial. BMC Med Educ 2010;10:68 [FREE Full text] [CrossRef] [Medline]
    26. Spickard A, Smithers J, Cordray D, Gigante J, Wofford JL. A randomised trial of an online lecture with and without audio. Med Educ 2004 Jul;38(7):787-790. [CrossRef] [Medline]
    27. Cook DA. The failure of e-learning research to inform educational practice, and what we can do about it. Med Teach 2009 Feb;31(2):158-162. [CrossRef] [Medline]
    28. Cook D, Levinson A, Garside S, Dupras DM, Erwin PJ, Montori VM. Internet-based learning in the health professions: a meta-analysis. JAMA 2008 Sep 10;300(10):1181-1196. [CrossRef] [Medline]
    29. Raupach T, Brown J, Anders S, Hasenfuss G, Harendza S. Summative assessments are more powerful drivers of student learning than resource intensive teaching formats. BMC Med 2013;11:61 [FREE Full text] [CrossRef] [Medline]
    30. Raupach T, Hanneforth N, Anders S, Pukrop T, Th J ten Cate O, Harendza S. Impact of teaching and assessment format on electrocardiogram interpretation skills. Med Educ 2010 Jul;44(7):731-740. [CrossRef] [Medline]

    Edited by G Eysenbach; submitted 28.08.14; peer-reviewed by M Fischer, E Santoro; comments to author 17.02.15; revised version received 18.07.15; accepted 02.09.15; published 28.09.15

    ©Tobias Raupach, Clemens Grefe, Jamie Brown, Katharina Meyer, Nikolai Schuelper, Sven Anders. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 28.09.2015.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.