Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Trends in progress test performance of medical students at a university in Peru

Abstract

Background

Progress testing is a longitudinal assessment method used to monitor the acquisition and retention of knowledge throughout medical training. While progress tests (PTs) have been widely adopted internationally through collaborative networks of medical schools, in Peru, their implementation has been primarily institutional. This study aimed to evaluate longitudinal trends in PT scores at a Peruvian medical school.

Methods

We conducted a longitudinal analysis using data from PTs administered annually between 2017 and 2024. The PT assessed students’ knowledge based on the subjects completed at the time of testing. Scores ranged from 0 to 250 and were converted to a 20-point scale. Independent variables included number of PTs taken (1–7), year of entry into medical school (entry cohort; 2017–2024), year of test administration (2017–2024), and sex. Generalized estimating equations (GEE) were used to assess score trends over time, applying an identity link function with a Gaussian distribution and robust standard errors clustered by student ID.

Results

We included 1,899 test scores from 669 medical students. The mean score across all tests was 9.19 (standard deviation = 2.34). No consistent upward trend in PT scores was observed over the study period; scores decreased by 0.088 points per additional year (CI95% CI: −0.147 to −0.029, p = 0.003). Students who completed five PTs scored significantly higher than those who took four (β = 1.40; 95% CI: 0.79 to 2.01). When stratified by entry cohort, no sustained improvement in scores was observed within cohorts over time.

Conclusion

Over an eight-year period of administering a progress test at a Peruvian medical school, student performance remained stable, with an average of approximately 50% of questions answered correctly per test. Longitudinal analysis did not reveal a sustained increase in scores as students advanced through the curriculum. This pattern may be explained by the PT design, which assesses only the content covered by students at the time of each administration, in contrast to other PTs that measure end-of-curriculum knowledge across all cohorts. Nevertheless, an increase in median scores was observed during the transition from basic science to clinical subjects.

Introduction

In medical schools, progress tests (PTs) are longitudinal assessments designed to track students’ learning trajectories over time [1,2]. These evaluations provide individualized feedback on knowledge gaps based on each student’s progression through the curriculum. PTs emerged from the need to comprehensively assess learning outcomes and evaluate the effectiveness of problem-based curricula. As a result, medical schools began adopting this strategy in the late 1970s [1]. The first implementations started in the United States (University of Missouri–Kansas City), the Netherlands (Maastricht University), and Canada (McMaster University) [3,4]. In subsequent decades, the approach spread across Europe. In South America, Brazilian universities began applying PTs in the late 1990s [5], while adoption in the Middle East began in 2012 [6].

At the international level, progress tests are primarily developed by consortia of medical schools [710]. However, some institutions have implemented them as independent initiatives. In general, PTs allow for the assessment of students at multiple points throughout their educational journey. They may be administered three to four times per year and across all years of study, although their mandatory nature varies by institution [11]. Each test is generally cross-sectional in nature, as all students are expected to take the test simultaneously on the same day [12].

PTs are grounded in the principle that students are assessed based on the level of knowledge expected at the completion of their medical education. However, some variations of PTs adjust the expected knowledge level according to the student’s stage in the curriculum at the time of testing [13]. The scores obtained from PTs provide valuable data for evaluating the effectiveness of the curriculum in promoting progressive learning, as well as for identifying opportunities for curricular improvement [14]. Longitudinal studies conducted among medical students have consistently demonstrated a progressive increase in PT scores over the course of their education [1,10,1517]. Furthermore, longitudinal data from PTs have been used to assess the impact of the COVID-19 pandemic on academic performance, although findings in this area have been contradictory [18,19].

In Peru, there is no PT developed by a national association of medical schools that systematically promotes its implementation across medical training institutions. As a result, there are no standardized data available to objectively compare medical schools or to evaluate students’ knowledge progression during and at the end of their training, regardless of each university’s curricular design. To date, PTs in Peru have been implemented as isolated institutional initiatives [20,21]. As such, longitudinal analysis of student cohorts and the identification of factors associated with academic progression remain key areas for investigation. In this context, our primary objective was to analyze trends in progress test scores among students at a medical school in Peru.

Materials and methods

Study design and setting

We conducted a longitudinal study using secondary data from the School of Medicine at the University of Piura, located in Lima, Peru. This medical school has implemented PTs since its inception in 2017. We analyzed data collected between 2017 and 2024.

The PT used in this study is called “Annual Case-Based Medical Examination.” It is administered once a year during the second week of December, at the end of the academic year. The design, management, and administration of the test are entirely performed by the School of Medicine itself. Student participation is voluntary. The test consists of a written examination with 250 multiple-choice questions, each with a single correct answer. From 2017 to 2021, each question included four distractors and one correct option; starting in 2022, three distractors were used. The PT is organized into three booklets, each containing questions based on four to six narrative clinical cases. These cases are drawn from case reports published in peer-reviewed medical journals. The case descriptions, including relevant images and tables, are presented in full. The formulation of questions derived from these cases is supervised by course coordinators.

A distinctive feature of the progress test at this university is that it evaluates students based on their current stage of training, rather than on the level of knowledge expected at the completion of the medical program. Accordingly, the test content is limited to the subjects completed by the end of the academic year in which the assessment is administered. For instance, a third-year student is assessed on content from subject completed during the first, second, and third years. The number of questions per subject is proportional to the number of academic credits assigned to each course [21] (Fig 1).

thumbnail
Fig 1. Curriculum of the School of Medicine at the University of Piura, showing the courses and their corresponding academic credits (in parentheses) in relation to the annual administration of the progress test (PT).

PT1 refers to the progress test administered at the end of the first year, and so on.

https://doi.org/10.1371/journal.pone.0330029.g001

The PT was administered virtually in 2020 and 2021 due to social restrictions imposed by the COVID-19 pandemic. In all other years, the test was conducted in person on the university campus using printed questionnaires.

Participants

For the longitudinal analysis, we included students who completed six years of study prior to the medical internship (entry cohorts of 2017, 2018, and 2019) and participated in three or more PTs between 2017 and 2024. Students who discontinued their studies during the observation period were excluded from the analysis.

A multiple-group cross-sectional analysis was conducted to compare progress test scores between students in the preclinical phase (first to third year, focused on basic sciences) and those in the clinical phase (fourth to sixth year, focused on clinical sciences). This analysis included students who completed progress tests administered between 2019 and 2022. In both analyses, all students who met the inclusion criteria were included.

Variables

The primary variable of interest was the PT score, which ranges from 0 to 250 points, with one point awarded for each correct answer. The raw score was subsequently converted to a 20-point scale, where a score of 10 corresponds to answering 50% of the questions correctly. Incorrect answers did not reduce points. The number of PT taken was considered a discrete variable, with values ranging from one to seven. As participation in the test is voluntary, not all students completed the six potential tests; some students took up to seven tests due to repeating academic years. Additional variables included the year of entry (cohorts from 2017 to 2024), the year of test administration (2017–2024), and sex (male/female). All data were provided by the Assessment Unit of the School of Medicine on 03/02/2025.

Data analysis

A descriptive analysis was conducted to examine participants by sex, entry cohort, and the number of PTs taken, using both absolute and relative frequencies. The median and interquartile range of the number of PTs were calculated for each sex and entry cohort. PTs scores across all years were summarized using the mean and standard deviation, while the distribution of scores was presented using a histogram. The histogram intervals were set to a width of 1. Furthermore, the distribution of scores for each year of test administration was illustrated using dot plots. Pearson correlation coefficients were calculated to assess the relationships between PTs scores across different years. The strength of the correlations was interpreted as follows: 0 to 0.10 (negligible), 0.11 to 0.39 (weak), 0.40 to 0.69 (moderate), 0.70 to 0.89 (strong), and 0.90 to 1.00 (very strong) [22].

The comparison of PTs scores between entry cohorts and years of test administration was performed using non-parametric tests: the Mann-Whitney U test for comparisons between two groups, and the Kruskal-Wallis test for comparisons among three or more groups. Post-hoc multiple comparisons were conducted using the Dunn test.

To assess the trend in PTs scores, we applied a generalized estimating equations (GEE) model with an identity link function, assuming a Gaussian distribution for the outcome. The standard error was adjusted for clustering by a unique numeric identifier, with an exchangeable correlation structure. The model included the year of test administration and was adjusted for the number of PTs taken and sex. Subsequently, using the margins command, we estimated the predictive margins for PTs scores for each year of administration, accounting for the estimated coefficients and adjusted standard errors from the GEE model. These estimates were calculated separately by entry cohort and the number of PTs taken.

We used the same generalized estimating equations (GEE) model to assess the trend in scores within the 2017, 2018, and 2019 entry cohorts, including the year of test administration and the number of PTs taken as variables. The GEE analysis was performed in STATA version 16 using the xtgee command. Graphs were generated using GraphPad Prism 10.4.2. A statistical significance level of 5% was considered.

Ethical considerations

The study protocol was exempted from review by the Institutional Ethics Committee of the University of Piura, as the data were collected during routine academic activities and the analyzed database was fully anonymized. Prior to this exemption, the committee verified that obtaining informed consent was not feasible due to the nature of the secondary database analysis.

Results

A total of 1,899 progress test scores from 669 medical students were included in the analysis. Of these, 369 (55.2%) were female. The first cohort (2017) included 51 students, while the most recent cohorts (2023 and 2024) each comprised 124 students. Among the participants, 7.6% (n = 51) did not take any PTs. Furthermore, 27.1% (n = 181) of students completed between four and five PTs from 2017 to 2024. Older entry cohorts tended to have a higher median number of progress tests completed (Table 1).

thumbnail
Table 1. Characteristics of medical students and the number of progress tests taken from 2017 to 2024.

https://doi.org/10.1371/journal.pone.0330029.t001

Overall analysis of scores

The mean of the 1,899 PTs scores was 9.19, with a standard deviation of 2.34 (Fig 2A). The median score was 9.20, with an interquartile range (IQR) of 7.72 to 10.72. The minimum score recorded was 0.56, and the maximum was 16.72. In the PT administered in 2018, the median score was 11 (IQR = 9.42–12.16), while the lowest median score was observed in 2023, at 8.10 (IQR = 6.84–9.44). The PT administered in 2020 and 2021 were conducted virtually, with median scores of 10.24 and 10.48, respectively (Fig 2B).

thumbnail
Fig 2. (A) Histogram of the 1,899 progress test scores taken between 2017 and 2024.

(B) Scatter plot of progress test scores by year of administration. The colors of the points reflect the scores of students from a particular entry cohort. In the 2017 plot, only one color is used, in the 2018 plot, two colors (representing two cohorts), and so on. The red central lines represent the median, and the bars represent the 1st and 3rd quartiles. The dashed black horizontal line indicates half of the score on the 20-point scale.

https://doi.org/10.1371/journal.pone.0330029.g002

PT scores from a given year demonstrated moderate to strong positive linear correlations with scores from the subsequent five years of test administrations. However, the 2017 PT scores did not show a significant correlation with those from the 2023 and 2024 administrations. Similarly, the 2018 PT scores were not significantly correlated with those from 2024 (Fig 3).

thumbnail
Fig 3. Heat map of the Pearson correlation coefficients between progress test scores across test administrations from 2017 to 2024.

Light blue: 0.40 a 0.69 (moderate correlation), dark blue: 0.70 a 0.89 (strong correlation). P values > 0.05 were considered non-significant correlations. n is the number of valid data pairs used to assess the correlation.

https://doi.org/10.1371/journal.pone.0330029.g003

In the PT administered in 2019, no significant differences were observed in the median scores between the entry cohorts in 2017–2019 (H = 1.785, p = 0.41). In the PT of 2020, the 2019 entry cohort undertook clinical science subjects; however, no differences were found in their scores compared to the cohorts taking basic science subjects. In contrast, in the 2021 and 2022 PTs, students who transitioned to clinical subjects had significantly higher median scores than those in basic science subjects (Table 2).

thumbnail
Table 2. Comparison of progress test scores from 2019 to 2022 among cohorts that completed the first three years of study, based on entry cohort and transition to clinical subjects.

https://doi.org/10.1371/journal.pone.0330029.t002

Longitudinal analysis of scores

A total of 167 students were part of the 2017, 2018, and 2019 entry cohorts, of whom 12 completed three or fewer tests. The longitudinal analysis included data from 155 students, 81 (52.3%) of whom were male. By entry cohort, 45 (29.0%) were from 2017, 55 (35.5%) from 2018, and 55 (35.5%) from 2019. Regarding the number of tests taken, 18 (11.6%) students completed four tests, 66 (42.6%) completed five, 61 (39.4%) completed six, and 10 (6.5%) completed seven tests.

In 2018, 2020, and 2021, the median scores were above 10, while in the other years, the median was below half of the 20-point scale. In 2018, 2020, and 2017, the highest scores exceeded 15 (16.72, 16.0, and 15.28, respectively). Regular students may voluntarily take up to six PTs during their medical training. However, PT scores from students in the 2017 cohort were still observed in the 2023 and 2024 administrations, and from the 2018 cohort in the 2024 PT (Fig 4A). The average scores of the PTs were higher for students who completed five tests compared to those who completed four or seven tests (Fig 4B).

thumbnail
Fig 4. (A) Scatter plot of progress test scores by year of administration.

The entry cohort of 2017 is represented by blue points, the 2018 entry cohort by yellow points, and the 2019 entry cohort by green points. (B) Scatter plot of average progress test scores grouped by the number of tests taken. The bars above the points represent p-values from Dunn’s multiple comparisons tests. The dashed black horizontal line indicates half of the maximum score on a 20-point scale. The red central lines correspond to the median, and the bars represent the 1st and 3rd quartiles.

https://doi.org/10.1371/journal.pone.0330029.g004

No clear trend was observed in the distribution of PTs scores across any of the entry cohorts. In the 2017 cohort, significant differences were found in the median scores (H statistic = 31.5, p < 0.001), primarily due to the higher median scores in 2021 (Fig 5A). In the 2018 cohort, significant differences were also identified in the median scores (H statistic = 20.2, p = 0.001), attributed to the lower median scores in 2023 compared to 2021 and 2018 (Fig 5B). Finally, in the 2019 cohort, the difference in medians (H statistic = 44.5, p < 0.001) was explained by the lower median scores in the last two years compared to 2020, 2021, and 2022 (Fig 5C).

thumbnail
Fig 5. Scatter plot of scores in the first six progress tests by entry cohort.

(A) 2017 entry cohort, (B) 2018 entry cohort, (C) 2019 entry cohort. All cohorts had the opportunity to complete at least six progress tests before the medical internship. The dashed black horizontal line indicates half of the maximum score on a 20-point scale. The red central lines correspond to the median, and the bars represent the 1st and 3rd quartiles. The upper bars indicate p-values obtained from Dunn’s multiple comparisons test.

https://doi.org/10.1371/journal.pone.0330029.g005

In the generalized estimating equations (GEE) model, no consistent upward trend in scores over time was observed. When we formulated a model with time as a numerical variable, we found that, on average, scores decreased by 0.088 points per additional year (SE = 0.030, 95% CI: −0.147 to −0.029, p = 0.003), after controlling for the number of PT taken.

Another model, presented in Table 3, included time (year of PT administration), the number of PTs, and sex as explanatory variables. In the full model, the years 2021 and 2018 showed significant increases in scores compared to 2017 (+1.38 and +1.21, respectively), while scores in 2019, 2023, and 2024 did not differ significantly from those in 2017. These results remained consistent when the model only considered the number of PTs taken. In the reduced model, regardless of time, taking five PTs resulted in a 1.40-point increase in scores compared to taking four PTs, while taking six tests resulted in a 0.87-point increase compared to four. No clear trend of increasing scores with a higher number of PTs taken was observed. When the predicted scores, based on the GEE model estimates, were graphed, it was evident that, in all years of PT administration, students who took five PTs had higher scores (Fig 6A). Similarly, when scores were analyzed by entry cohort, all cohorts showed a sustained decline in scores from 2022 to 2024 (Fig 6B).

thumbnail
Table 3. Generalized Estimating Equations (GEE) models to assess the evolution of scores based on time, number of progress tests, and sex.

https://doi.org/10.1371/journal.pone.0330029.t003

thumbnail
Fig 6. Prediction of progress test scores based on the number of tests taken (A) and entry cohort (B).

The predictions were based on estimates from generalized estimating equations (GEE) models with the following independent variables: year of progress test administration, number of progress tests taken, and sex.

https://doi.org/10.1371/journal.pone.0330029.g006

When replicating the analysis for each entry cohort, different trends in scores over time were observed. In the 2017 cohort, a significant increasing trend in scores was found between 2018 and 2021, while no significant differences were observed in 2023 and 2024 compared to 2017. In the 2018 cohort, scores in 2019 and 2020 decreased significantly compared to the first PT. Subsequently, between 2022 and 2024, a declining trend in scores was observed. In the 2019 cohort, a significant increase in scores was seen in 2020 and 2021 compared to the baseline (2019), followed by a decrease in scores in 2022 and 2023, with the reduction in 2023 being significant. In all cohorts, taking five PTs was associated with higher scores (Table 4).

thumbnail
Table 4. Evolution of progress test scores in medical students by entry cohort.

https://doi.org/10.1371/journal.pone.0330029.t004

Discussion

We did not observe a clear and sustained upward trend in the scores across eight PTs administered at a medical school in Peru. Although the score trajectories varied by entry cohort, the scores, on average, showed a stable pattern. Students in the 2017 cohort had some years in which scores were higher than those obtained in the entry year. In the 2018 cohort, scores decreased in 2019, 2020, and 2024 compared to the first PT. In the 2019 cohort, score increases were observed only in 2020 and 2021, compared to the first year. In all PTs, the median scores tended to remain around half of the maximum score on the 20-point scale.

Our data show a moderate to strong linear correlation among the scores of the first five PTs. However, this correlation disappears for the sixth and seventh tests. Several factor may explain this finding. Across all entry cohorts and years of test administration, students who completed five PTs had higher scores compared to those who completed four or seven PTs. This pattern can be explained by the characteristics of our PT and the curriculum progression. The maximum number of tests a student can complete, assuming they pass all courses and fulfill the required academic credits each semester, is six. Students who completed seven PTs are those who were not promoted to the next year at some point in the curriculum, while those who completed four or fewer tests likely had insufficient and inconsistent exposure to the PT, mainly due to its non-mandatory nature at this medical school. Another possible explanation for declining participation over time is a decrease in student´s favorable perceptions of the PTs after repeated administrations. For example, students who completed five PTs reported lower satisfaction levels compared to those who had taken only one [23]. Additionally, given the non-mandatory nature of this PT experience, students in their final years may have opted out to focus on targeted preparation for the national medical licensing examination.

In our study, using a 20-point scale, the average scores did not exceed half of the maximum score. Additionally, in the longitudinal analysis, we consistently observed that scores did not increase as students progressed through the curriculum. This contrasts with previous studies where PTs were designed to measure cumulative knowledge expected at the end of medical training—often showing a clear increase in scores over time. In our context, however, the PT specifically assessed students’ knowledge based on the curricular content covered up to the point of each test administration. This design feature appears to be a key factor underlying the observed score stability across academic years. Therefore, rather than indicating a lack of academic progress, the absence of score growth in our study likely reflects the formative nature of the test and its alignment with students’ stage-specific learning milestones.

As previously mentioned, longitudinal studies using traditional PTs have reported score increases of varying magnitudes. A study conducted at the medical schools of the Universities of Groningen, Maastricht, and Nijmegen observed an increase in the average proportion of correct answers from 5.7% to 67.5% between the first and last PT, by the end of the sixth year of study [24]. At McMaster University, a similar trend was observed in four entry cohorts, with the proportion of correct answers rising from approximately 12–18% on the first test to nearly 50% on the final test [4]. In Saudi Arabia, an analysis of progress test results collected over a 10-year period from multiple medical schools found that the proportion of correct answers increased from 6% in the first year to 38.2% in the fifth year [17]. This pattern was also observed at the University of São Paulo in Brazil, where the proportion of correct responses rose from 32% in the first year to 56.5% in the sixth year [25]. This upward trajectory was further corroborated by a cross-sectional analysis of a 2015 national PT administered in several Brazilian medical schools, which showed that first-year students achieved an average of 32.38% correct responses, compared to 61.28% among sixth-year students [7].

Although our findings on PTs score trends are not directly comparable to those reported in most previous studies, we observed that medical students, on average, achieved scores equivalent to approximately 50% of the maximum possible on each administration of the test. This level of correct responses is comparable to that reported in other settings for students taking their final PT, typically at the end of the curriculum, after completion of the full curriculum [4,15,24,25]. Despite these results, 73.4% of students exposed to this PT reported that it helped them apply their knowledge to clinical contexts, 60.8% felt it allowed them to demonstrate their knowledge, and 52.1% considered it a fair assessment [23]. Additionally, among 26 graduates from the 2017 cohort and 25 from the 2018 cohort who took the Peruvian national medical licensing examination in 2023 and 2024, respectively, none failed the exam. These facts suggest that the observed PT performance was expected and did not negatively affect students’ academic progress or preparedness for licensure.

The analysis period included the COVID-19 pandemic. Despite this, the administration of PTs continued, although the testing modality changed in 2020 and 2021, shifting from in-person to online formats. During these two years, all three student cohorts obtained higher scores compared to the 2019 PT. This transition from face-to-face to virtual testing was adopted in several countries [18,19]. However, the impact of pandemic-related adaptations in medical education on performance in PTs remains a subject of ongoing debate. For instance, at the Charité – Universitätsmedizin Berlin, an increase in student scores was reported in the PTs administered in April and November 2022, compared to previous exams [19]. In contrast, a study conducted at two universities in Brazil found that students enrolled in clinical subjects, who experienced suspended hospital rotations during the pandemic, did not show a significant increase in scores in the 2020 PT compared to 2019 [18].

In 2022, PTs returned to an in-person format using printed materials. That year, a significant change was implemented: the number of distractors per item reduced from four to three. Among students from the 2017–2019 entry cohorts, we observed a sustained decline in PTs scores from 2022 onward. Reducing the number of distractors can help eliminate non-functional distractors, thereby increasing the distractor efficiency of the remaining items. This adjustment is directly related to the item difficulty index [26]. Studies have shown that in multiple-choice questions with three distractors, when all distractors are functional, the average item difficulty is around 56%. However, when only one distractor is functional, the item difficulty increases to approximately 74% [27]. An item difficulty index of 80% or higher typically indicates that a question is easier to answer [28].

Another key finding was the increase in PTs scores when students began their clinical subjects in the fourth year of study. No significant differences in scores were observed during the first three years, which focus on basic sciences. This result aligns with a previous study conducted in the same institution, using the same PT. In that study, students who had already started their clinical training scored higher on questions related to content from the first two years, even outperforming students currently taking those early-year subjects [29]. This improvement may be attributed to the enhanced vertical integration activities during the clinical training phase, which likely reinforce and consolidate the knowledge gained in earlier stages of the curriculum [30].

The study has limitations that should be considered when interpreting the results. The PT analyzed assesses knowledge related to the subjects completed by students at the time of the test. Most international experiences with PTs evaluate the expected knowledge at graduation, which complicates the comparison of our findings. However, it provides new insights into this particular type of PT. With the data analyzed, we were unable to conduct a psychometric evaluation of the eight PTs. This information would have provided greater context to explain the longitudinal trends in scores. This analysis pertains to a single medical school, so the findings and longitudinal score trajectories cannot be generalized to other public or private universities in Peru. However, it demonstrates the feasibility of implementation in the Peruvian context. Two characteristics of the PT should be considered. First, the non-mandatory nature of the test meant that not all students had the opportunity to take all six tests, which affected the estimation of longitudinal trajectories. Additionally, some irregular students, who failed courses and were not promoted to the next year, took more PTs than expected. On average, these students performed worse in these evaluations.

Based on these findings, we propose several recommendations. The developers of the analyzed PT could include a subset of questions that assess the knowledge expected of a medical student upon graduation. This would provide additional pedagogical insights from this PT experience. The analysis of longitudinal performance suggests that students should take a minimum of five PTs. The transition from basic sciences to clinical subjects in the curriculum leads to improvements in PT performance. However, the observed increase in the median scores by one point could be further enhanced by strengthening vertical integration strategies in the curriculum. Finally, it is necessary to evaluate the reliability and validity of the applied PTs.

Conclusion

This PT experience, designed to assess students’ knowledge based on their progression through the curriculum, provides longitudinal data that do not show a clear and sustained upward trend in PTs scores. This stable pattern of performance was consistently observed across all analyzed entry cohorts. Although the median scores fluctuated within each cohort, they tended to stabilize around the midpoint of the 20-point scale. Students who completed five or six PTs achieved higher predicted scores across all years of test administration. However, this association may be influenced by overall academic performance, as students who took five PTs were typically regular students who did not fail subjects. Finally, the transition from basic science to clinical subjects in the curriculum coincided with an increase in median scores.

Supporting information

References

  1. 1. Vleuten CPMVD, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem-based learning curriculum. Medical Teacher. 1996;18(2):103–9.
  2. 2. Görlich D, Friederichs H. Using longitudinal progress test data to determine the effect size of learning in undergraduate medical education - a retrospective, single-center, mixed model analysis of progress testing results. Med Educ Online. 2021;26(1):1972505. pmid:34459724
  3. 3. Freeman A, Van Der Vleuten C, Nouns Z, Ricketts C. Progress testing internationally. Med Teach. 2010;32(6):451–5. pmid:20515370
  4. 4. Blake JM, Norman GR, Keane DR, Mueller CB, Cunnington J, Didyk N. Introducing progress testing in McMaster University’s problem-based medical curriculum: psychometric properties and effect on learning. Acad Med. 1996;71(9):1002–7. pmid:9125989
  5. 5. Cecilio-Fernandes D, Bicudo AM, Hamamoto Filho PT. Progress testing as a pattern of excellence for the assessment of medical students’ knowledge: concepts, history, and perspective. Medicina (Ribeirão Preto). 2021;54(1):e173770.
  6. 6. Moursy NA, Hamsho K, Gaber AM, Ikram MF, Sajid MR. A systematic review of progress test as longitudinal assessment in Saudi Arabia. BMC Med Educ. 2025;25(1):100. pmid:39838466
  7. 7. Bicudo AM, Hamamoto Filho PT, Abbade JF, Hafner M de LMB, Maffei CML. Teste de Progresso em Consórcios para Todas as Escolas Médicas do Brasil. Rev bras educ med. 2019;43(4):151–6.
  8. 8. Johnson TR, Khalil MK, Peppler RD, Davey DD, Kibble JD. Use of the NBME Comprehensive Basic Science Examination as a progress test in the preclerkship curriculum of a new medical school. Adv Physiol Educ. 2014;38(4):315–20. pmid:25434014
  9. 9. Majeed GM, Islam J, Nandakumar G, Phoong K. Progress Testing in UK Medical Education: Evaluating Its Impact and Potential. Cureus. 2024;16(1):e52607. pmid:38249657
  10. 10. Alavarce DC, de Medeiros ML, de Araújo Viana D, Abade F, Vieira JE, Machado JLM, et al. The progress test as a structuring initiative for programmatic assessment. BMC Med Educ. 2024;24(1):555. pmid:38773470
  11. 11. Schuwirth LWT, van der Vleuten CPM. The use of progress testing. Perspect Med Educ. 2012;1(1):24–30. pmid:23316456
  12. 12. Albanese M, Case SM. Progress testing: critical analysis and suggested practices. Adv Health Sci Educ Theory Pract. 2016;21(1):221–34. pmid:25662873
  13. 13. Wrigley W, van der Vleuten CPM, Freeman A, Muijtjens A. A systemic framework for the progress test: strengths, constraints and issues: AMEE Guide No. 71. Med Teach. 2012;34(9):683–97. pmid:22905655
  14. 14. McNeish D, Dumas D, Torre D, Rice N. Modelling Time to Maximum Competency in Medical Student Progress Tests. Journal of the Royal Statistical Society Series A: Statistics in Society. 2022;185(4):2007–34.
  15. 15. Heeneman S, Schut S, Donkers J, van der Vleuten C, Muijtjens A. Embedding of the progress test in an assessment program designed according to the principles of programmatic assessment. Med Teach. 2017;39(1):44–52. pmid:27646870
  16. 16. Cecilio-Fernandes D, Nagtegaal M, Noordzij G, Tio RA. Cumulative assessment: Does it improve students’ knowledge acquisition and retention? Sci Med. 2018;28(4):31880.
  17. 17. Alamro AS, Alghasham AA, Al-Shobaili HA, Alhomaidan HT, Salem TA, Wadi MM, et al. 10 years of experience in adopting, implementing and evaluating progress testing for Saudi medical students. J Taibah Univ Med Sci. 2022;18(1):175–85. pmid:36398029
  18. 18. Hamamoto Filho PT, Cecilio-Fernandes D, Norcia LF, Sandars J, Anderson MB, Bicudo AM. Reduction in final year medical students’ knowledge during the COVID-19 pandemic: Insights from an interinstitutional progress test. Front Educ. 2022;7.
  19. 19. Sehy V, Roselló Atanet I, Sieg M, Struzena J, März M. Effects of COVID-19 Pandemic on Progress Test Performance in German-Speaking Countries. Education Research International. 2022;2022:1–9.
  20. 20. Flores Cohaila J. Asociación entre el promedio ponderado universitario y exámenes de progreso de ciencias básicas y ciencias clínicas frente al puntaje obtenido del ENAM 2020 en internos de medicina de la Universidad Privada de Tacna. Universidad Privada de Tacna. 2021. Available: https://repositorio.upt.edu.pe/handle/20.500.12969/1940
  21. 21. Romaní Romaní FR, Gutiérrez C. Correlación entre una evaluación sumativa escrita y el promedio ponderado en estudiantes de medicina humana. Inv Ed Med. 2022;11(43):37–50.
  22. 22. Schober P, Boer C, Schwarte LA. Correlation Coefficients: Appropriate Use and Interpretation. Anesth Analg. 2018;126(5):1763–8. pmid:29481436
  23. 23. Pachas-Mu A, Bouroncle-Derteano B, Romaní-Romaní F. Percepciones y satisfacción sobre una prueba de progreso en estudiantes de medicina. Educación Médica. 2025;26(3):101020.
  24. 24. Muijtjens AMM, Schuwirth LWT, Cohen-Schotanus J, Thoben AJNM, van der Vleuten CPM. Benchmarking by cross-institutional comparison of student achievement in a progress test. Med Educ. 2008;42(1):82–8. pmid:18181848
  25. 25. Tomic ER, Martins MA, Lotufo PA, Benseñor IM. Progress testing: evaluation of four years of application in the school of medicine, University of São Paulo. Clinics (Sao Paulo). 2005;60(5):389–96. pmid:16254675
  26. 26. Kheyami D, Jaradat A, Al-Shibani T, Ali FA. Item Analysis of Multiple Choice Questions at the Department of Paediatrics, Arabian Gulf University, Manama, Bahrain. Sultan Qaboos Univ Med J. 2018;18(1):e68–74. pmid:29666684
  27. 27. Chauhan GR, Chauhan BR, Vaza JV, Chauhan PR. Relations of the Number of Functioning Distractors With the Item Difficulty Index and the Item Discrimination Power in the Multiple Choice Questions. Cureus. 2023;15(7):e42492. pmid:37644928
  28. 28. Johari J, Sahari J, Wahab DA, Abdullah S, Abdullah S, Omar MZ, et al. Difficulty Index of Examinations and Their Relation to the Achievement of Programme Outcomes. Procedia - Social and Behavioral Sciences. 2011;18:71–80.
  29. 29. Romaní-Romaní F, Gutiérrez C, Azurin-Salazar J. Tendencia en la retención de conocimientos de ciencias básicas en una prueba de progreso entre estudiantes de Medicina. Educación Médica. 2023;24(4):100830.
  30. 30. Wijnen-Meijer M, van den Broek S, Koens F, Ten Cate O. Vertical integration in medical education: the broader perspective. BMC Med Educ. 2020;20(1):509. pmid:33317495