ANALYSIS OF DIFFERENTIAL ITEM FUNCTIONING (DIF) IN THE FINAL SCHOOL ASSESSMENT INSTRUMENT FOR THE CHEMISTRY SUBJECT
Abstract
Purpose – Fair assessment is a fundamental pillar for ensuring that evaluation results accurately reflect students' abilities without bias. This study aims to identify Differential Item Functioning (DIF) in the Final School Assessment (PAS) instrument for the Chemistry subject, based on four demographic variables: gender, family economic status, residential location, and school of origin.
Methodology – This study uses a quantitative design with a descriptive-exploratory approach. The research subjects were responses from 1,840 twelfth-grade students at high schools in Maros Regency. Data analysis was conducted using an Item Response Theory (IRT) approach in R (version 2024.4.2, Build 764). After assumption tests (unidimensionality and local independence) and a model fit test, the 1-Parameter Logistic (1PL) Model was selected as the most suitable. DIF detection was performed using Raju’s Area Measures.
Findings – The analysis results showed that the assumptions of unidimensionality and local independence were met. Out of 30 items, five showed statistically significant DIF (p < 0.05): one item based on gender (Item 11), one based on residential location (Item 26), and three based on economic status (Items 19, 23, and 24). No items showed DIF by school of origin. Although statistically detected, the effect size analysis showed that all DIF items fell into Category 'A' (negligible) according to ETS criteria.
Contribution – This study provides empirical evidence regarding the fairness of an assessment instrument developed by a teacher association (MGMP). It highlights how non-academic factors can manifest as measurable differences in performance. This study affirms the importance of DIF analysis as a standard procedure in the quality assurance of assessment instruments to maintain fairness for all students.Keywords
Full Text:
PDFReferences
Anizar, & Sardin. (2023). Evaluasi pada kurikulum Merdeka dan pemanfaatan hasil penilaiannya. Edupedia Publisher.
Amelia, R., et al. (2022). Deteksi bias gender pada instrumen evaluasi belajar kimia dengan metode Mantel-Haenszel. Jurnal Tarbiyah, 29(2).
Chalmers, P. (2022). Multidimensional item response theory. CRAN.
Guo, H., Lu, R., Johnson, M. S., & McCaffrey, D. F. (2022). Alternative Methods for Item Parameter Estimation: From CTT to IRT (Research Report No. RR-22-12). In ETS Research Report Series (Vol. 12, Issue 1). https://doi.org/10.1002/ets2.12355
Gyamfi, A., & Acquaye, R. (2023). Parameters and Models of Item Response Theory ( IRT ): A Review of Literature. Acta Educationis Generalis, 13(3), 68–78. https://doi.org/10.2478/atd-2023-0022
Istiqlal, M., Istiyono, E., Widihastuti, Sari, D. K., Danni, R., & Safitri, I. (2024). Construction of Mathematics Cognitive Test Instrument of Computational Thinking Model for Madrasah Aliyah Students. Nazhruna: Jurnal Pendidikan Islam, 7(2), 475–492. https://doi.org/10.31538/nzh.v7i2.4425
Istiyono, E., Dwandaru, W. S. B., Fenditasari, K., Ayub, M. R. S., & Saepuzaman, D. (2022). The Development of a Four-Tier Diagnostic Test Based on Modern Test Theory in Physics Education. European Journal of Educational Research, 12(1), 371–385. https://doi.org/10.12973/eu-jer.12.1.371
Hardiyanti, N. (2018). Komparasi metode dalam mendeteksi sensitivitas bias item soal USBN Kimia SMA Negeri di Kabupaten Bone dengan pendekatan teori respon item [Tesis, Universitas Negeri Makassar].
Lerner, J. Y., McCubbins, M. D., & Renberg, K. M. (2021). The efficacy of measuring judicial ideal points: The mis-analogy of IRTs. International Review of Law and Economics, 68, 1–11. https://doi.org/10.1016/j.irle.2021.106020
Magis, D., & Raiche, G. (2022). Collection of methods to detect dichotomous differential item functioning (DIF). CRAN.
Nor, R., Pendidikan, E., Pascasarjana, P., & Yogyakarta, U. N. (2017). IMPLEMENTASI ITEM RESPONSE THEORY SEBAGAI BASIS ANALISIS KUALITAS BUTIR SOAL DAN KEMAMPUAN KIMIA SISWA KOTA YOGYAKARTA Implementation of Item Response Theory for Analysis of Test Items Quality and Students ’ Ability in Chemistry. 2(1), 1–12.
Pratama, D., dkk. (2021). Analisis differential item functioning (DIF) pada skala sikap moderasi beragama siswa. Proyeksi, 18(1), 116–131.
Pristiwaluyo, T., & Syamsuddin, S. (2021). Development of an instrument for teachers’ attitudes towards academic supervision performed by supervisors in schools of special education. Journal of Educational Science and Technology (EST), 7(1), 40–49.
Priyadi, S. (2024). Differential item functioning (DIF): Sebuah analisis bibliometrik. Edukatif: Jurnal Ilmu Pendidikan, 6(4), 3975–3989.
Rust, J., Kosinski, M., & Stillwell, D. (2020). Modern Psychometrics. In Modern Psychometrics. https://doi.org/10.4324/9781315637686
Safitri, I., & Ansyari, R. (2024). The Calibration of Science Achievement Test Based on Integrated Islamic Curriculum. 8th International Conference on Education and Multimedia Technology (ICEMT), 329–336. https://doi.org/10.1145/3678726.3678727
Safitri, I., Lestarani, D., Imtikhanah, R. D. N. W., Akbarini, N. R., Sari, M. W., Fitrah, M., & Hapsan, A. (2024). TEORI PENGUKURAN DAN EVALUASI. CV. Ruang Tentor.
Safitri, I., Rosnawati, R., & Ansyari, R. (2024). Estimasi Kesalahan Pengukuran dalam Penilaian Sidang Skripsi: Generalizability Theory Analysis. Afeksi: Jurnal Penelitian Dan Evaluasi Pendidikan, 5(1), 162–168. https://doi.org/https://doi.org/10.35672/afeksi.v5i1.220
Samritin, S. (2022). Identifikasi muatan differential item functioning pada data Ujian Nasional Matematika. Journal on Education, 4(4), 1675–1684.
Setiawan, A. (2020). Pendeteksian DIF pada perangkat tes objektif penilaian akhir semester IPA dengan menggunakan permodelan Rasch. PSEJ (Pancasakti Science Education Journal), 5(2), 23–29.
Thiyagarajan, A., James, T. G., & Marzo, R. R. (2022). Psychometric properties of the 21-item Depression, Anxiety, and Stress Scale (DASS-21) among Malaysians during COVID-19: a methodological study. Humanities and Social Sciences Communications, 9(1), 1–8. https://doi.org/10.1057/s41599-022-01229-x
Yin, R. K. (2018). Case study research: Design and methods (6th ed.). SAGE Publications.
Wahyuni, A. (2022). Detection of gender bias using DIF (differential item functioning) analysis on the item test of the school examination in Yogyakarta. Jurnal Evaluasi Pendidikan, 13(1).
DOI: https://doi.org/10.36987/jes.v13i2.7781
Refbacks
- There are currently no refbacks.
Copyright (c) 2026 Sangkala Maros

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.







1.jpg)






