Acessibilidade / Reportar erro

Assessment of medical students’ Surgery knowledge based on Progress Test

ABSTRACT

Progress Testing (PT) is an assessment tool whose use has grown throughout Brazil in the last decade. PT makes it possible to assess the students’ knowledge gain throughout the undergraduate course and, for their interpretations to be valid, their items (questions) must have adequate quality from the point of view of content validity and reliability of results. In this study, we analyzed the psychometric characteristics of the items and the performance of students in the content area of surgery from 2017 to 2023. For the analyses, we used the assumptions of Classical Test Theory, Bloom’s taxonomy and Cronbach’s alpha reliability coefficient. The items were easy (average difficulty index between 0.3-0.4), with fair to good discrimination (discrimination index between 0.3-0.4) and with a predominance of medium to high taxonomy. Reliability remained substantial over the years (>0.6). Students’ knowledge gain in surgery was found to be progressive and more important from the 3rd year of the undergraduate course, reaching approximately 70-75% in the 6th year. This measurements framework can be replicated in other contexts for a better understanding of student learning and for qualification of evaluation processes.

Keywords:
Surgery; Educational Measurement; Medical Education; Psychometrics

RESUMO

O Teste de Progresso (TP) é uma ferramenta de avaliação cujo uso tem crescido em todo o Brasil na última década. O TP permite avaliar o ganho de conhecimento dos estudantes ao longo do curso de graduação e, para que suas interpretações sejam válidas, é preciso que seus itens (questões) tenham qualidade adequada do ponto de vista de validade de conteúdo e confiabilidade de resultados. Neste estudo, analisamos as características psicométricas dos itens e o desempenho dos estudantes na área de cirurgia do TP de 2017 a 2023. Para as análises, usamos os pressupostos da Teoria Clássica dos Testes, a taxonomia de Bloom e o coeficiente de fidedignidade alfa de Cronbach. Os itens se mostraram fáceis (índice de dificuldade média entre 0,3-0,4), com discriminação de regular a boa (índice de discriminação entre 0,3-0,4) e com predomínio de questões de média a alta taxonomia. A confiabilidade se manteve substancial ao longo dos anos (>0,6). O ganho de conhecimento dos estudantes em cirurgia é progressivo e mais importante a partir do 3º ano do curso de graduação, chegando a aproximadamente 70-75% no 6º ano. Este arcabouço de aferições pode ser replicado em outros contextos para melhor compreensão do aprendizado dos estudantes e para qualificação dos processos avaliativos.

Palavras-chave:
Cirurgia; Avaliação Educacional; Educação Médica; Psicometria

INTRODUCTION

The Progress Test (PT) is an assessment applied to medical students aiming to analyze the consecutive gain of knowledge throughout the course. The same test is applied to all students, from the first to the sixth year, with different tests for each application, which have a fixed periodicity and whose content is aimed at the level of the newly graduated doctor11 Schuwirth LW, van der Vleuten CP. The use of progress testing. Perspect Med Educ. 2012;1(1):24-30. doi: 10.1007/s40037-012-0007-2.
https://doi.org/10.1007/s40037-012-0007-...
.

PT was created in the Netherlands and the United States in the 1970s and aimed to change the culture of evaluating the teaching-learning process, with the principle of longitudinal evaluation and monitoring of the process effectiveness22 Van der Vleuten CPM, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem based learning curriculum. Med Teach. 1996;18(2):103-9. doi: 10.3109/01421599609034142.
https://doi.org/10.3109/0142159960903414...
,33 Arnold L, Willoughby TL. The quarterly profile examination. Acad Med. 1990;65(8):515-6. doi: 10.1097/00001888-199008000-00005.
https://doi.org/10.1097/00001888-1990080...
. Today, PT is recognized for its potential to provide detailed feedback for students, teachers, and the medical school itself, providing information on personal, group, curriculum, and institution performance44 Coombes L, Ricketts C, Freeman A, Stratford J. Beyond assessment: feedback for individuals and institutions based on the progress test. Med Teach. 2010;32(6):486-90. doi: 10.3109/0142159X.2010.485652.
https://doi.org/10.3109/0142159X.2010.48...
. Furthermore, PT reduces the endogeneity effect of assessments conducted within the same school, as, working with consortia of schools, there are multiple sources of origin for items (questions)55 Muijtjens AM, Schuwirth LW, Cohen-Schotanus J, van der Vleuten CP. Origin bias of test items compromises the validity and fairness of curriculum comparisons. Med Educ. 2007;41(12):1217-23. doi: 10.1111/j.1365-2923.2007.02934.x.
https://doi.org/10.1111/j.1365-2923.2007...
. For this reason, PT has proven to be a useful predictor of performance in professional certification exams or for medical residency66 Karay Y, Schauber SK. A validity argument for progress testing: Examining the relation between growth trajectories obtained by progress tests and national licensing examinations using a latent growth curve approach. Med Teach. 2018 Nov;40(11):1123-9. doi: 10.1080/0142159X.2018.1472370.
https://doi.org/10.1080/0142159X.2018.14...
,77 Hamamoto Filho PT, de Arruda Lourenção PLT, do Valle AP, Abbade JF, Bicudo AM. The Correlation Between Students' Progress Testing Scores and Their Performance in a Residency Selection Process. Med Sci Educ. 2019;29(4):1071-5. doi: 10.1007/s40670-019-00811-4.
https://doi.org/10.1007/s40670-019-00811...
.

In Brazil, PT has been used since the late 1990s and early 2000s88 Tomic ER, Martins MA, Lotufo PA, Benseñor IM. Progress testing: evaluation of four years of application in the school of medicine, University of São Paulo. Clinics (Sao Paulo). 2005;60(5):389-96. doi: 10.1590/s1807-59322005000500007.
https://doi.org/10.1590/s1807-5932200500...
. With almost 20 years of experience, the Interinstitutional Center for Studies and Assessment Practices in Medical Education (NIEPAEM) is the consortium that brings together public medical schools in the State of São Paulo, and its practices have been the basis for replicating the model throughout Brazil99 Bicudo AM, Hamamoto Filho PT, Abbade JF, Hafner MLMB, Maffei CML. Consortia of Cross-Institutional Progress Testing for All Medical Schools in Brazil. Rev Bras Educ Med. 2019;43(4):151-6. doi: 10.1590/1981-52712015v43n4RB20190018.
https://doi.org/10.1590/1981-52712015v43...
.

Traditionally, PT’s questions are divided into the areas stipulated for medical residency exams: clinics, pediatrics, surgery, gynecology, and public health1010 Ministério da Educação. Resolução no 01, de 14 de agosto de 2000. [Acesso em 18 Jul 2023]. Disponível em: <https://www.gov.br/mec/pt-br/residencia-medica/ementario-da-legislacao-da-residencia-medica>.
https://www.gov.br/mec/pt-br/residencia-...
. This division has been questioned for giving equal weight to areas with heterogeneous content extension, which may compromise the reliability of the evidence in subareas and, therefore, of the evidence itself (unpublished data). Still, given the historical series of PT application, it would be possible to obtain information about the teaching of surgery in Brazil today, particularly in schools in São Paulo. Therefore, the objective of this study was to analyze the characteristics of the items and the performance of students in surgery PT from 2017 to 2023.

METHODS

Study design

This is a cross-sectional, observational, analytical study conducted using information from the Interinstitutional Center for Studies and Assessment Practices in Medical Education (NIPAEM) database. It is a consortium of the following medical schools: Paulista State University (UNESP), University of São Paulo (USP, courses in Ribeirão Preto and Bauru), State University of Campinas (UNICAMP), Federal University of São Paulo (UNIFESP), Federal University of São Carlos (UFSCar), State University of Londrina (UEL), Faculty of Medicine of São José do Rio Preto (FAMERP) and Faculty of Medicine of Marília (FAMEMA). Until 2022, the group had the participation of the Regional University of Blumenau (FURB). This is a study based on an aggregated database with individualized information at the item level (questions). Therefore, there is no individualized information on students (sex/age). As per the NIEPAEM code of conduct, there is also no identification of performance by institution, avoiding comparisons that lead to the classification of schools.

We included data from tests administered annually from 2017 onwards, including the first application of the test in 2023, when the test began to be administered bi-annually. For the psychometric data of the questions, we considered only the grades of students in the sixth year, as the test is formulated for the level of recent graduates. To analyze progress, we considered the performance of all students. Until 2022, the PT surgery section had 20 items per test. As of 2023, the section now has 23 items. Table 1 contains the matrix of question themes. This matrix is followed to prepare the test, to guarantee similarity of content in different applications.

Table 1
Knowledge matrix used to prepare the questions.

For the analyses, we investigated the difficulty and discrimination index of the items (according to the Classical Test Theory), the taxonomic classification of the questions, and the test’s reliability coefficient (measured by Cronbach’s alpha coefficient).

Ethical considerations

Because it deals with a database study made available in an aggregated form without the possibility of individual identification of students, this study does not need to be assessed by an Ethics in Research Committee, in accordance with the legislation of the National Commission for Ethics in Research with Human Beings (CONEP)1111 Conselho Nacional de Saúde. Resolução no 674 de 06 de maio de 2022. [acesso em 05 Ago 2023]. Disponível em: <https://conselho.saude.gov.br/resolucoes-cns/2469-resolucao-n-674-de-06-de-maio-de-2022>.
https://conselho.saude.gov.br/resolucoes...
.

Data analysis

The difficulty level of each item was calculated as the percentage of errors in each item (i.e., the closer to 1, the more difficult the question). To classify the degree of difficulty of each item, we adopted the following values: above 0.8 - difficult; between 0.4 and 0.8 - average; below 0.4 - easy.

We calculated the discrimination index by the difference in correct answers for each item between the 27% of students with higher performance on the test and the 27% with lower performance. Thus, the index can vary from -1 to 1, and the closer to 1, the better the discrimination. We adopted the following values to classify the questions: ≥0.4 - good; ≥0.3 and below 0.4 - regular; ≥0.2 and below 0.3 - weak; <0.2 - deficient.

We computed the alpha reliability coefficient for each test according to the formula proposed by Cronbach1212 Cronbach LJ. Coefficient alpha and the internal structure of tests. Psychometrika. 1951;16(3):297-334. doi: 10.1007/BF02310555.
https://doi.org/10.1007/BF02310555...
. It refers to the internal consistency of the measure, that is, the extent to which the items measure the same construct. We adopted the following classification: > 0.8 -near perfect; from 0.8 to 0.61 - substantial; from 0.6 to 0.41 - moderate; from 0.4 to 0.21 - reasonable; <0.21 - small.

According to Bloom’s Taxonomy, later modified by Anderson and Krathwol, cognitive educational domains can be classified according to the complexity of cognitive processes into: knowledge, understanding, application, analysis, synthesis, and evaluation1313 Bloom BS. Taxonomy of educational objectives: the classification of education goals. Cognitive domain. Handbook 1. New York: Longman; 1956. ISBN-10: 0679302093; ISBN-13: 978-0679302094,1414 Anderson LW, Krathwohl DR, Airasian PW, et al. A taxonomy for learning, teaching, and assessing: A revision of Bloom's Taxonomy of Educational Objectives. New York: Addison Wesley Longman; 2001.. For taxonomic classification of items, they were classified according to the cognitive repertoire involved in their resolution as low (memorization), medium (understanding), or high taxonomy (application/analysis).

We calculated student performance as a function of the average percentage of correct answers for each year of graduation.

To analyze the temporal trend of psychometric indicators, we conducted a simple linear regression. In analyzing the difference in student performance, we performed a one-way ANOVA test followed by the Tukey test for paired comparisons between subsequent years of the undergraduate course. We considered p<0.05 as statistically significant.

The analyzes were conducted using GraphPad v. 9.5.0 (GraphPad Software Inc. San Diego, CA, USA) and SPSS (Statistical Package for Social Sciences, IBM Corp., Armonk, NY, USA).

RESULTS

Regarding the difficulty of the items, we observed that on the annual average, the items were easy, with an average difficulty index varying between 0.3 and 0.4 (Figure 1). Only the 2019 test had a higher average difficulty, close to 0.6. As for item discrimination, the average pointed to regular discrimination (between 0.3 and 0.4), with the 2019, 2021, and 2023 tests displaying an index close to or greater than 0.4 (good discrimination, Figure 1). The reliability analysis of the test, as measured by the reliability coefficient (Cronbach’s alpha), showed that, except for the 2017 test, all had a value greater than or equal to 0.6 (substantial internal consistency, Figure 1). In the temporal trend analysis of the indicators, all demonstrated stability: low angular coefficients and statistically non-significant (Table 2)..

Table 2
Linear regression results for behavioral trends of psychometric indicators in surgery in the progress test from 2017 to 2023.

Figure 1
Mean values with respective 95% confidence intervals for the difficulty (A) and discrimination (B) indices of the surgery items for each application of the progress test in the years 2017 to 2023. C: coefficient reliability values of the surgery area for the same years.

In the analysis of the items’ taxonomic classification, we observed a predominance of medium to high taxonomy questions (Figure 2), that is, there are few items that emphasize memorization of content and concepts, and more items that require greater cognitive complexity, with understanding, analysis, and application of contents.

Figure 2
Violin plot for the number of items by taxonomic classification. The larger diameter of the violin indicates a greater concentration of items in that classification.

As for student performance, we observed that there were progressive gains with each year of graduation, starting from an average of 25 to 35% correct answers in the first year, reaching 70-75% in the sixth. When comparing subsequent graduation, performance was different for almost all years, except for the comparison between the first and second years, which showed no difference for the test applying in 2017, 2018, 2021, and 2022 (Figure 3).

Figure 3
Performance of students in surgery in the progress test exemplified in the 2017, 2019, and 2022. The comparison of performance between subsequent series is always significant, except for the comparison between the first and second years. *p<0.05; ***p<0.0001; ns: not significant.

DISCUSSION

PT has been increasingly used in Brazilian medical schools. Faced with numerous discussions about the importance of external and serial assessments of medical students, PT appears to be a useful tool for allowing diagnoses on student performance and curriculum behavior and, ultimately, the effectiveness of the teaching-learning process1515 Cecilio-Fernandes D, Bicudo AM, Hamamoto Filho PT. Progress testing as a pattern of excellence for the assessment of medical students' knowledge - concepts, history, and perspective. Medicina (Ribeirão Preto). 2021;54(1):e-173770. doi: 10.11606/issn.2176-7262.rmrp.2021.173770.
https://doi.org/10.11606/issn.2176-7262....
,1616 Troncon LEA, Elias LLK, Osako MK, Romão EA, Bollela VR, Moriguti JC. Reflections on the use of the Progress Test in the programmatic student assessment. Rev Bras Educ Med. 2023;47(2):e076. doi: 10.1590/1981-5271v47.2-2022-0334.ing.
https://doi.org/10.1590/1981-5271v47.2-2...
.

There are valid criticisms of regarding PT’s limitation for inferences about specific disciplines. For example, a student’s error on a single anatomy question does not mean that the student does not have the necessary cognitive skills about anatomy. Therefore, for more accurate analyzes of a particular area, more in-depth approaches to the test results are necessary, or an increase in sampling of the area in question by increasing the number of its items in the test1717 Swanson DB, Case SM. Assessment in basic science instruction: directions for practice and research. Adv Health Sci Educ Theory Pract. 1997;2(1):71-84. doi: 10.1023/A:1009702226303.
https://doi.org/10.1023/A:1009702226303...
.

Therefore, we believe that this work, analyzing equivalent evidence in surgery (which makes up 1/6 to 1/5 of PT) across seven applications, keeping the matrix of covered knowledge fixed, allows for some conclusions.

Firstly, we noted that the surgery test is not a difficult one. The questions have low average difficulty ratings (below 0.4) and have remained stable over the years. Knowing that the contents covered in the test are at the level of a newly qualified doctor, this is a good indicator.

This observation is reinforced by the average discrimination index, which has also remained stable and above 0.3, therefore, with items regulating good discrimination. The interpretation of this indicator is that the items have been adequate in detecting students with good or deficient performance. Therefore, even on easy questions, schools and mentors can identify students with unsatisfactory performance, who deserve attention. We should emphasize that the 2019, 2021, and 2022 tests used pre-tested items, that is, they were items already used in previous PT versions and were chosen based on their good psychometric behavior. It is not by chance, therefore, that these were the tests in which we observed the best item discrimination averages.

This observation has an important practical implication for medical schools. Teachers usually repeat test questions - either due to lack of organization of a proper database, or due to limited creativity in writing new items1818 Boulet JR, McKinley DW, Whelan GP, Hambleton RK. The effect of task exposure on repeat candidate scores in a high-stakes standardized patient assessment. Teach Learn Med. 2003;15(4):227-32. doi: 10.1207/S15328015TLM1504_02.
https://doi.org/10.1207/S15328015TLM1504...

19 Wood TJ. The effect of reused questions on repeat examinees. Adv Health Sci Educ Theory Pract. 2009;14(4):465-73. doi: 10.1007/s10459-008-9129-z.
https://doi.org/10.1007/s10459-008-9129-...
-2020 O'Neill TR, Sun L, Peabody MR, Royal KD. The Impact of Repeated Exposure to Items. Teach Learn Med. 2015;27(4):404-9. doi: 10.1080/10401334.2015.1077131.
https://doi.org/10.1080/10401334.2015.10...
. It is recommended that items be reused with sufficient time intervals to minimize biases in remembering questions among students and their peers. With our data, we draw attention to the fact that teachers, in addition to ensuring intervals in the application of repeated items, should study the psychometric behavior of the questions after their use, identifying questions that are very easy, very difficult, or that do not show good discrimination, that is, do not contribute to an adequate assessment2121 Albanese M, Case SM. Progress testing: critical analysis and suggested practices. Adv Health Sci Educ Theory Pract. 2016;21(1):221-34. doi: 10.1007/s10459-015-9587-z.
https://doi.org/10.1007/s10459-015-9587-...
. The psychometric analysis of items can be done automatically on online assessment platforms, the use of which gained notoriety with the COVID-19 pandemic2222 Patael S, Shamir J, Soffer T, Livne E, Fogel-Grinvald H, Kishon-Rabin. Remote proctoring: Lessons learned from the COVID-19 pandemic effect on the large scale on-line assessment at Tel Aviv University. J Comput Assist Learn. 2022;38(6):1554-73. doi: 10.1111/jcal.12746.
https://doi.org/10.1111/jcal.12746...
.

The reliability coefficient is also an indicator that the PT surgery test is of superior quality, with indices that have remained above 0.6 and, therefore, with substantial internal consistency and comparable to values obtained internationally2323 Blake JM, Norman GR, Keane DR, Mueller CB, Cunnington J, Didyk N. Introducing progress testing in McMaster University's problem-based medical curriculum: psychometric properties and effect on learning. Acad Med. 1996;71(9):1002-7. doi: 10.1097/00001888-199609000-00016.
https://doi.org/10.1097/00001888-1996090...
. Obviously, the set of all items that comprise the test raises the coefficient to values above 0.8-0.9 (unpublished data). It must be recognized, however, that this coefficient is influenced by the score variance and, therefore, by the number of respondents2121 Albanese M, Case SM. Progress testing: critical analysis and suggested practices. Adv Health Sci Educ Theory Pract. 2016;21(1):221-34. doi: 10.1007/s10459-015-9587-z.
https://doi.org/10.1007/s10459-015-9587-...
. As we have students from nine schools, the sample size is large and naturally increases the variance of responses and the value of Cronbach’s alpha. However, the set of different psychometric indicators taken together suggests that the assessment has had a satisfactory quality.

To this set of indicators, we added the items’ taxonomic category, with a predominance of medium to high taxonomy questions. It has been demonstrated that items with a higher taxonomy have a better discrimination index than items with a lower one2424 Rush BR, Rankin DC, White BJ. The impact of item-writing flaws and item complexity on examination item difficulty and discrimination value. BMC Med Educ. 2016;16(1):250. doi: 10.1186/s12909-016-0773-3.
https://doi.org/10.1186/s12909-016-0773-...
,2525 Hamamoto Filho PT, Silva E, Ribeiro ZMT, Hafner MLMB, Cecilio-Fernandes D, Bicudo AM. Relationships between Bloom's taxonomy, judges' estimation of item difficulty and psychometric properties of items from a progress test: a prospective observational study. Sao Paulo Med J. 2020;138(1):33-9. doi: 10.1590/1516-3180.2019.0459.R1.19112019.
https://doi.org/10.1590/1516-3180.2019.0...
. Thus, the PT surgery items have required more clinical reasoning from students than memorization of facts or concepts, possibly approaching cognitive domains closer to the clinical practice of the newly graduated doctor.

Finally, regarding the performance of medical students, we observed gains in knowledge throughout the course, as expected. The interesting fact is that the gain is significant from the third year onwards, which reflects the reality of the curriculum of most Brazilian medical schools, in which the teaching of surgical technique and clinical practice occurs more frequently in the third year. Recently, it was demonstrated that curricular exposure of students to surgery content improves their performance on PT surgery items, although performance at the end of the course is similar among students, regardless of the curricular design2626 Wearn A, Bindra V, Patten B, Loveday BPT. Relationship between medical programme progress test performance and surgical clinical attachment timing and performance. Med Teach. 2023;45(8):877-84. doi: 10.1080/0142159X.2023.2186205.
https://doi.org/10.1080/0142159X.2023.21...
. The fact that sixth-year students have an average success rate close to 70-75% is comparable to that of other areas of knowledge and to what is reported in the international literature on PT22 Van der Vleuten CPM, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem based learning curriculum. Med Teach. 1996;18(2):103-9. doi: 10.3109/01421599609034142.
https://doi.org/10.3109/0142159960903414...
,2727 Nouns ZM, Georg W. Progress testing in German speaking countries. Med Teach. 2010;32(6):467-70. doi: 10.3109/0142159X.2010.485656.
https://doi.org/10.3109/0142159X.2010.48...
,2828 Cecilio-Fernandes D, Aalders WS, Bremers AJA, Tio RA, de Vries J. The Impact of Curriculum Design in the Acquisition of Knowledge of Oncology: Comparison Among Four Medical Schools. J Cancer Educ. 2018;33(5):1110-4. doi: 10.1007/s13187-017-1219-2.
https://doi.org/10.1007/s13187-017-1219-...
.

This study is not without limitations. Due to the very nature of PT, we cannot infer conclusions about specific learning in each surgical specialty. Obviously, as this is a knowledge assessment, it is also not possible to make any inferences about the teaching and learning of basic surgical skills that every doctor should have, nor about professional attitudes. It should also be noted that the data from the 2023 test correspond to the test administered at the end of the first semester, and not at the end of the year, since this year the frequency of the test was increased to twice a year. Furthermore, we do not have information at the individual student level to make other inferences based on covariates such as sex, age, and institution. As per the NIEPAEM code of conduct, each student’s performance information is only available to the student’s own school, and not to the group of schools.

Despite these limitations, this study provides useful information about the quality of PT for assessing surgical knowledge and provides more evidence about the knowledge gain curve of medical students. Together, we present a framework of assessment quality measurements that can be repeated in other contexts to qualify medical student assessment.

CONCLUSION

The surgery items that comprise the NIEPAEM Progress Test are not difficult, have good discrimination, favor clinical reasoning, and produce good reliability indicators. Students’ knowledge gain is significant from the third year of the undergraduate course and reaches 70 75% by the sixth year.

REFERENCES

  • 1
    Schuwirth LW, van der Vleuten CP. The use of progress testing. Perspect Med Educ. 2012;1(1):24-30. doi: 10.1007/s40037-012-0007-2.
    » https://doi.org/10.1007/s40037-012-0007-2
  • 2
    Van der Vleuten CPM, Verwijnen GM, Wijnen WHFW. Fifteen years of experience with progress testing in a problem based learning curriculum. Med Teach. 1996;18(2):103-9. doi: 10.3109/01421599609034142.
    » https://doi.org/10.3109/01421599609034142
  • 3
    Arnold L, Willoughby TL. The quarterly profile examination. Acad Med. 1990;65(8):515-6. doi: 10.1097/00001888-199008000-00005.
    » https://doi.org/10.1097/00001888-199008000-00005
  • 4
    Coombes L, Ricketts C, Freeman A, Stratford J. Beyond assessment: feedback for individuals and institutions based on the progress test. Med Teach. 2010;32(6):486-90. doi: 10.3109/0142159X.2010.485652.
    » https://doi.org/10.3109/0142159X.2010.485652
  • 5
    Muijtjens AM, Schuwirth LW, Cohen-Schotanus J, van der Vleuten CP. Origin bias of test items compromises the validity and fairness of curriculum comparisons. Med Educ. 2007;41(12):1217-23. doi: 10.1111/j.1365-2923.2007.02934.x.
    » https://doi.org/10.1111/j.1365-2923.2007.02934.x
  • 6
    Karay Y, Schauber SK. A validity argument for progress testing: Examining the relation between growth trajectories obtained by progress tests and national licensing examinations using a latent growth curve approach. Med Teach. 2018 Nov;40(11):1123-9. doi: 10.1080/0142159X.2018.1472370.
    » https://doi.org/10.1080/0142159X.2018.1472370
  • 7
    Hamamoto Filho PT, de Arruda Lourenção PLT, do Valle AP, Abbade JF, Bicudo AM. The Correlation Between Students' Progress Testing Scores and Their Performance in a Residency Selection Process. Med Sci Educ. 2019;29(4):1071-5. doi: 10.1007/s40670-019-00811-4.
    » https://doi.org/10.1007/s40670-019-00811-4
  • 8
    Tomic ER, Martins MA, Lotufo PA, Benseñor IM. Progress testing: evaluation of four years of application in the school of medicine, University of São Paulo. Clinics (Sao Paulo). 2005;60(5):389-96. doi: 10.1590/s1807-59322005000500007.
    » https://doi.org/10.1590/s1807-59322005000500007
  • 9
    Bicudo AM, Hamamoto Filho PT, Abbade JF, Hafner MLMB, Maffei CML. Consortia of Cross-Institutional Progress Testing for All Medical Schools in Brazil. Rev Bras Educ Med. 2019;43(4):151-6. doi: 10.1590/1981-52712015v43n4RB20190018.
    » https://doi.org/10.1590/1981-52712015v43n4RB20190018
  • 10
    Ministério da Educação. Resolução no 01, de 14 de agosto de 2000. [Acesso em 18 Jul 2023]. Disponível em: <https://www.gov.br/mec/pt-br/residencia-medica/ementario-da-legislacao-da-residencia-medica>.
    » https://www.gov.br/mec/pt-br/residencia-medica/ementario-da-legislacao-da-residencia-medica
  • 11
    Conselho Nacional de Saúde. Resolução no 674 de 06 de maio de 2022. [acesso em 05 Ago 2023]. Disponível em: <https://conselho.saude.gov.br/resolucoes-cns/2469-resolucao-n-674-de-06-de-maio-de-2022>.
    » https://conselho.saude.gov.br/resolucoes-cns/2469-resolucao-n-674-de-06-de-maio-de-2022
  • 12
    Cronbach LJ. Coefficient alpha and the internal structure of tests. Psychometrika. 1951;16(3):297-334. doi: 10.1007/BF02310555.
    » https://doi.org/10.1007/BF02310555
  • 13
    Bloom BS. Taxonomy of educational objectives: the classification of education goals. Cognitive domain. Handbook 1. New York: Longman; 1956. ISBN-10: 0679302093; ISBN-13: 978-0679302094
  • 14
    Anderson LW, Krathwohl DR, Airasian PW, et al. A taxonomy for learning, teaching, and assessing: A revision of Bloom's Taxonomy of Educational Objectives. New York: Addison Wesley Longman; 2001.
  • 15
    Cecilio-Fernandes D, Bicudo AM, Hamamoto Filho PT. Progress testing as a pattern of excellence for the assessment of medical students' knowledge - concepts, history, and perspective. Medicina (Ribeirão Preto). 2021;54(1):e-173770. doi: 10.11606/issn.2176-7262.rmrp.2021.173770.
    » https://doi.org/10.11606/issn.2176-7262.rmrp.2021.173770
  • 16
    Troncon LEA, Elias LLK, Osako MK, Romão EA, Bollela VR, Moriguti JC. Reflections on the use of the Progress Test in the programmatic student assessment. Rev Bras Educ Med. 2023;47(2):e076. doi: 10.1590/1981-5271v47.2-2022-0334.ing.
    » https://doi.org/10.1590/1981-5271v47.2-2022-0334.ing
  • 17
    Swanson DB, Case SM. Assessment in basic science instruction: directions for practice and research. Adv Health Sci Educ Theory Pract. 1997;2(1):71-84. doi: 10.1023/A:1009702226303.
    » https://doi.org/10.1023/A:1009702226303
  • 18
    Boulet JR, McKinley DW, Whelan GP, Hambleton RK. The effect of task exposure on repeat candidate scores in a high-stakes standardized patient assessment. Teach Learn Med. 2003;15(4):227-32. doi: 10.1207/S15328015TLM1504_02.
    » https://doi.org/10.1207/S15328015TLM1504_02
  • 19
    Wood TJ. The effect of reused questions on repeat examinees. Adv Health Sci Educ Theory Pract. 2009;14(4):465-73. doi: 10.1007/s10459-008-9129-z.
    » https://doi.org/10.1007/s10459-008-9129-z
  • 20
    O'Neill TR, Sun L, Peabody MR, Royal KD. The Impact of Repeated Exposure to Items. Teach Learn Med. 2015;27(4):404-9. doi: 10.1080/10401334.2015.1077131.
    » https://doi.org/10.1080/10401334.2015.1077131
  • 21
    Albanese M, Case SM. Progress testing: critical analysis and suggested practices. Adv Health Sci Educ Theory Pract. 2016;21(1):221-34. doi: 10.1007/s10459-015-9587-z.
    » https://doi.org/10.1007/s10459-015-9587-z
  • 22
    Patael S, Shamir J, Soffer T, Livne E, Fogel-Grinvald H, Kishon-Rabin. Remote proctoring: Lessons learned from the COVID-19 pandemic effect on the large scale on-line assessment at Tel Aviv University. J Comput Assist Learn. 2022;38(6):1554-73. doi: 10.1111/jcal.12746.
    » https://doi.org/10.1111/jcal.12746
  • 23
    Blake JM, Norman GR, Keane DR, Mueller CB, Cunnington J, Didyk N. Introducing progress testing in McMaster University's problem-based medical curriculum: psychometric properties and effect on learning. Acad Med. 1996;71(9):1002-7. doi: 10.1097/00001888-199609000-00016.
    » https://doi.org/10.1097/00001888-199609000-00016
  • 24
    Rush BR, Rankin DC, White BJ. The impact of item-writing flaws and item complexity on examination item difficulty and discrimination value. BMC Med Educ. 2016;16(1):250. doi: 10.1186/s12909-016-0773-3.
    » https://doi.org/10.1186/s12909-016-0773-3
  • 25
    Hamamoto Filho PT, Silva E, Ribeiro ZMT, Hafner MLMB, Cecilio-Fernandes D, Bicudo AM. Relationships between Bloom's taxonomy, judges' estimation of item difficulty and psychometric properties of items from a progress test: a prospective observational study. Sao Paulo Med J. 2020;138(1):33-9. doi: 10.1590/1516-3180.2019.0459.R1.19112019.
    » https://doi.org/10.1590/1516-3180.2019.0459.R1.19112019
  • 26
    Wearn A, Bindra V, Patten B, Loveday BPT. Relationship between medical programme progress test performance and surgical clinical attachment timing and performance. Med Teach. 2023;45(8):877-84. doi: 10.1080/0142159X.2023.2186205.
    » https://doi.org/10.1080/0142159X.2023.2186205
  • 27
    Nouns ZM, Georg W. Progress testing in German speaking countries. Med Teach. 2010;32(6):467-70. doi: 10.3109/0142159X.2010.485656.
    » https://doi.org/10.3109/0142159X.2010.485656
  • 28
    Cecilio-Fernandes D, Aalders WS, Bremers AJA, Tio RA, de Vries J. The Impact of Curriculum Design in the Acquisition of Knowledge of Oncology: Comparison Among Four Medical Schools. J Cancer Educ. 2018;33(5):1110-4. doi: 10.1007/s13187-017-1219-2.
    » https://doi.org/10.1007/s13187-017-1219-2
  • Funding source:

    none.

Publication Dates

  • Publication in this collection
    01 Dec 2023
  • Date of issue
    2023

History

  • Received
    06 Aug 2023
  • Accepted
    14 Oct 2023
Colégio Brasileiro de Cirurgiões Rua Visconde de Silva, 52 - 3º andar, 22271- 090 Rio de Janeiro - RJ, Tel.: +55 21 2138-0659, Fax: (55 21) 2286-2595 - Rio de Janeiro - RJ - Brazil
E-mail: revista@cbc.org.br