Analysis of Competitive Learning at University Level in Mexico via Item Response Theory

Open access


This paper presents a study of the multiple choice test from the eleventh knowledge tournament for Statistics I, in order to determine whether it instills competitive learning in university students. This research uses Item Response Theory (IRT). The results obtained show that only 27 students (13.43% of the total number of participants) have an acceptable level of ability (1.03 to 2.58), while the level of ability of the rest of the students is not satisfactory (-1.68 to 0.76). The participants are not a group of students seeking to test their knowledge of the subject or looking for an academic challenge. Better strategies for motivating students in terms of competitive learning must be found.

Aiken, L. R. (1979). Relationship between the item difficulty and discrimination indexes. Educational and Psychological Measurement, 39, 821–824.

Aiken, L. R. (2003). Test psicológicos y evaluación (11th ed.). Naucalpan: Pearson Education.

Awopeju, O. A., & Afolabi, E. R. I. (2016). Comparative analysis of classical test theory and item response theory based item parameter estimates of senior school certificate mathematics examination. European Scientific Journal, 12(28), 263–284.

Baker, F. B., & Kim, S.-H. (2017). The basics of item response theory using R. Springer International Publishing.

Balmori, S. Y., Delgadillo, G. H., & Méndez, R. I. (2011). Evaluación de un examen parcial de bioquímica. REB. Revista de Educación Bioquímica, 33(4), 3–7.

Best, J. W., & Kahn, J. V. (2006). Research in education. Pearson (10th ed.). Boston, MA: Pearson Education.

Birnbaum, A. (1968). Some latent trait models and their use in inferring an examinee’s ability. In F. Lord & M. Novick (Eds.), Statitsical Theories for Mental Test Scores (p. 592). Reading,MA: Addison-Wesley.

Cantador, I., & Conde, J. (2010). Effects of competition in education: a case study in an elearning environment. In IADIS international conference e-Learnig (pp. 11–18). Retrieved from

Carpio Cañada, J., Mateo Sanguino, T. J., Merelo Guervós, J. J., & Rivas Santos, V. M. (2015). Open classroom: enhancing student achievement on artificial intelligence through an international online competition. Journal of Computer Assisted Learning, 31(1), 14–31.

DiBattista, D., & Kurzawa, L. (2011). Examination of the quality of multiple-choice items on classroom tests. Canadian Journal for the Scholarship of Teaching and Learning, 2(2), 1–23.

DMC. (2017). Torneo de estadística. Retrieved from

Escudero, E. B., Reyna, N. L., & Morales, M. R. (2000). The level of difficulty and discrimination power of the basic knowledge and skills examination (EXHCOBA). Revista Electrónica de Investigación Educativa, 2(1), 1–16.

Fasli, M., & Michalakopoulos, M. (2015). Learning through game-like simulations. Innovation in Teaching and Learning in Information and Computer Sciences, 5(2), 1–11.

Finch, H., & French, B. (2015). Latent variable modeling with R. New York, NY: Routledge.

Furr, M. R., & Bacharach, V. R. (2013). Psychometrics: an introduction. Sage Publications (Second). Sage Publications, Inc.

Gajjar, S., Sharma, R., Kumar, P., & Rana, M. (2014). Item and test analysis to identify quality multiple choice questions (MCQS) from an assessment of medical students of Ahmedabad, Gujarat. Indian Journal of Community Medicine, 39(1), 17.

Hambleton, R. K., Swaminathan, H., & Rogers, H. J. (1991). Fundamentals of item response theory.

Hierro, L. Á., Atienza, P., & Pérez, J. L. (2014). Una experiencia de aprendizaje universitario mediante juegos de torneo en clase. REDU. Revista de Docencia Universitaria, 12(4), 415–436.

Ingale, A. S., Giri, P. A., & Doibale, M. K. (2017). Study on item and test analysis of multiple choice questions amongst undergraduate medical students. International Journal of Community Medicine and Public Health, 4(5), 1562–1565.

Johnson, D. W., & Johnson, R. T. (2002). Learning together and alone: overview and meta-analysis. Asia Pacific Journal of Education, 22(1), 95–105.

K., R., Hambleton, & Jones, R. W. (1993). Comparison of classical test theory and item response theory and their applications to test development. Educational Measurement: Issues and Practice, 12(3), 38–47.

Kim, S. L., & Sonnenwald, D. H. (2002). Investigating the relationship between learning style preferences and teaching collaboration skills and technology. Proceedings of the Asis Annual Meeting, 39, 64–73.

Lawrence, R. (2004). Teaching data structures using competitive games. IEEE Transactions on Education, 47(4), 459–466.

Lord, F. (1980). Applications of item response theory to practical testing problems. Hillsdale, NJ: Routledge.

Marie, S. M. J. A., & Edannur, S. (2015). Relevance of item analysis in standardizing an achievement test in teaching of physical science. Journal of Educational Technology, 12(3), 30–36.

McDonald, M. (2017). The nurse educators guide to assessing learning outcomes. Burlington,MA: Jones & Bartlett Learning.

Miller, M. D., Linn, R., & Gronlund, N. (2009). Measurement and assessment in teaching (10th ed.). upper Saddle River, NJ: Pearson Education.

Mitra, N., Nagaraja, H., Ponnudurai, G., & Judson, J. (2009). The levels of difficulty and discrimination indices in type a multiple choice questions of pre-clinical semester 1 multidisciplinary summative tests. International E-Journal of Science, Medicine & Education, 3(1), 2–7.

Muñiz, J. (2010). Las teorías de los test: teoría clásica y teoría de respuesta a los items. Papeles Del Psicólogo, 31(1), 57–66.

Owens, L., & Straton, R. G. (1980). The development of a cooperative, competitive, and individualised learning preference scale for students. British Journal of Educational Psychology, 50, 147–161.

Rao, C., Kishan Prasad, H., Sajitha, K., Permi, H., & Shetty, J. (2016). Item analysis of multiple choice questions: assessing an assessment tool in medical students. International Journal of Educational and Psychological Researches, 2(4), 201.

Rasch, G. (1980). Probabilistic models for some intelligence and attainment tests. Chicago, US: The University of Chicago Press.

Regueras, L. M., Verdú, E., Munoz, M. F., Perez, M. A., de Castro, J. P., & Verdú, M. J. (2009). Effects of competitive e-learning tools on higher education students: a case study. IEEE Transactions on Education, 52(2), 279–285.

Reise, S. P. (1990). A comparison of item- and person-fit methods of assessing model-data fit in IRT. Applied Psychological Measurement, 14(2), 127–137.

Rizopoulos, D. (2006). ltm: an R package for latent variable modeling and item response theory analyses. Journal of Statistical Software, 17(5), 1–25.

Rizopoulos, D. (2017). Package “ ltm.” Retrieved from

Romero, G. M. O., Rojas, P. A. D., Domínguez, O. R. L., Pérez, S. M. P., & Sapsin, K. G. (2015). Difficulty and discrimination of the items of the exams of reasearch methodology and statistics. Edumecentro, 7(2), 19–35.

Sinharay, S. (2003). Bayesian item fit analysis for dichotomous Item response theory models.

Thissen, D., & Wainer, H. (2001). Test scoring. Mahwah, NJ: Lawrence Erlbaum Associates, Inc.

Verhoeff, T. (1997). The Role of competitions in education. Future world: Educating for the 21st century. Retrieved from

Zamri Khairani, A., & Shamsuddin, H. (2016). Assessing item difficulty and discrimination indices of teacherdeveloped multiple-choice tests. In S. Fun Tang & L. Logonnathan (Eds.), Assessment for Learning Within and Beyond the Classroom (pp. 417–426). Springer Science & Business Media.