Analysis Of Instrument Test Of Historical Thinking Skills In Senior High School History Learning With Quest Programs

  • Ofianto Ofianto Universitas Negeri Padang
Keywords: Historical Thinking, Validity, Item Difficulty, Item Discrimination, Distractor Of Efficiency, Reliability

Abstract

This descriptive quantitative research aims to describe the quality of questions about history subjects in the West Sumatra District, including validity, item difficulty, item discrimination, distractor efficiency, and reliability. Objects in this study are question tests and 443 sheets of answer sheets, including MAN 3 Padang, MAN Pariaman, and MAN 3 Padang Panjang, a critical answer test, a test blueprint, and then its analysis with the quest program. This research shows that the content validity category was 28 (56%), and un-valid was 22 questions (44%). The difficulty questions were in the easy category, which was 1 question (2%); the immediate category was nine questions (18%), and the problematic category was 40 questions (80%). Item discrimination is weak category was 26 questions (52%), the enough category was 18 questions (36%), the good category was four questions (8%), the excellent questions were 1 question (2%), and the lousy category was 1 question (2%). Distractor efficiency was 47 questions (94%), and not function was three questions (6%). The average Kuder-Richarson formula 20 reliability coefficient was 0,53, which means it has enough reliability.

References

Alpusari, M. (2014). Analyze the basic concepts of Science 1 using the animated computer program version 4.0 for Windows. Primary, 3(2), 106–115.
Alwi, I. (2015). Empirical criteria in determining sample size in statistical hypothesis testing and grain analysis. Formative: Scientific Journal of Mathematics and Natural Sciences Education, 2(2).
Ambiyar, A. (2012). Professional Development Education Vocational High School Teacher Knowledge Century.
Arifin, Z. (2017). Criteria for instruments in a study. Journal of Theorems (the original mathematics research), 2(1).
Aslan, B., & Seker, H. (2017). Interactive Response Systems (IRS) Socrative Application Sample. Journal of Education and Learning, 6(1), 167-174.
Cattell, R. (Ed.). (2012). The scientific use of factor analysis in behavioral and life sciences. Springer Science & Business Media.
Chen, C. M., & Chung, C. J. (2008). A personalized mobile English vocabulary learning system is based on item response theory and the learning memory cycle. Computers & Education, 51(2), 624-645.
Choi, J., Kim, H., & Pak, S. (2018). Evaluation of Automatic Item Generation Utilities in Formative Assessment Application for Korean High School Students. Journal of Educational Issues, 4(1), 68-89.
Dudin, A. (2018). Case Study on Curriculum Implementation 2013 in Some Madrasah Aliyah in West Kalimantan Province: Case Study on Curriculum Implementation 2013 in Several Madrasah Aliyah in West Kalimantan Province. Dialogue, 41(1), 19–34.
Hamzah B. Uno & Satria Koni. (2012). Learning Assessment. Jakarta: Bumi Aksara.
Han, J. (2013). Scientific reasoning: Research, development, and assessment (Doctoral dissertation, The Ohio State University).
Hidayatulloh, M. A. (2013). Analysis of the quality of Arabic even semester exam questions for grade VIII for the 2012/2013 academic year with a quest program at SMP Muhammadiyah 3 Depok Sleman, Special Region of Yogyakarta (Doctoral Dissertation, Library Uin Sunan Kalijaga).
Kizlik, B. (2012). Measurement, assessment, and evaluation in education. Retrieved October, p. 10, 2015.
Kusaeri, K. (2018). The portrait of Madrasah Aliyah in Indonesia: A critical evaluation of the mathematics score in the national examination. In 2nd International Conference on Statistics, Mathematics, Teaching, and Research (pp. 1–7). IOP Publishing Ltd.
London, J., Matthews, K., & Grover, V. (2017). On meaning and measurement: A review of content validity in IS.
Nana Sudjana. (2005). Assessment of Teaching and Learning Process Results. Bandung: Remaja Rosdakarya.
Patrick, D. L., Burke, L. B., Gwaltney, C. J., Leidy, N. K., Martin, M. L., Molsen, E., & Ring, L. (2011). Content validity—establishing and reporting the evidence in newly developed patient-reported outcomes (PRO) instruments for medical product evaluation: ISPOR PRO Good Research Practices Task Force report: part 2—assessing respondent understanding. Value in Health, 14(8), 978-988.
Şahin, S. (2019). An analysis of English language testing and evaluation course in English language teacher education programs in Turkey: Developing language assessment literacy of pre-service EFL teachers.
Setiadi, H. (2016). Implementation of assessment in the 2013 Curriculum. Journal of educational research and evaluation, 20(2), 166–178.
Solekhah, F. M. A. (2018). Development of a high-level thinking ability test instrument on Newton's law material on motion.
Solichin, M. (2017). Analysis of question variance, difficulty level, validity of test items, interpretation of test results, and validity of predictions in educational evaluation. Dirasat: Journal of Islamic Management and Education, 2(2), 192–213.
Sudijono, A. (2001). Introduction to educational evaluation. PT. King Grafindo Persada.
Sunarti & Selly Rahmawati. (2014). Assessment in the 2013 Curriculum Helps Teachers and Prospective Teachers Know the Steps of Learning Assessment. Yogyakarta: Andi OFFSET.
Wasserman, E. A., & Young, M. E. (2010). Same–different discrimination: The keel and backbone of thought and reasoning. Journal of Experimental Psychology: Animal Behavior Processes, 36(1), 3.
Zainal Arifin. (2012). Learning Evaluation. Bandung: PT Remaja Rosdakarya.
Published
2021-09-15
How to Cite
Ofianto, O. (2021). Analysis Of Instrument Test Of Historical Thinking Skills In Senior High School History Learning With Quest Programs. Indonesian Journal of History Education, 6(2), 227-240. https://doi.org/10.15294/ijhe.v6i2.27648