EVALUATING THE CONSISTENCY OF SENIOR HIGH SCHOOL CHEMISTRY LABORATORY PRACTICE ASSESSMENT USING G-THEORY AND D-STUDY
DOI:
https://doi.org/10.26740/ujced.v14n3.p227-237Keywords:
laboratory practice assessment, reliability, generalizability theory, decision studyAbstract
This study aims to evaluate the reliability of chemistry laboratory practice assessments for Grade XII senior high school students by applying Generalizability Theory (G-theory) and Decision Study (D-study). A total of 79 students were involved in the research, each randomly assigned to perform one of six available chemistry laboratory practice tasks: electrolytes and nonelectrolytes, exothermic and endothermic reactions, enthalpy of neutralization between HCl and NaOH, acid-base titration, identification of acidic-basic properties, and electrolysis of CuSO₄. Each student was assessed independently by two chemistry teachers based on seven performance criteria: equipment selection, procedure, data reading, analysis, conclusion, cleanliness, and time efficiency. The G-study was conducted using a nested-crossed model in which students were nested within laboratory practice tasks and crossed with raters. The results revealed that variance due to raters (43.2%) and residual error (42.2%) dominated the total score variance, while the variance attributed to students nested within laboratory practice was relatively low (14.6%). The D-study produced a generalizability coefficient (Eρ²) of 0.41 and a dependability index (Φ) of 0.26, indicating low reliability for both relative and absolute decisions. A D-study simulation demonstrated that increasing the number of raters and laboratory practice tasks improved reliability. An optimal configuration of six tasks assessed by nine raters is required to achieve an Eρ² ≥ 0.80. These findings underscore the importance of well-designed assessment systems, consistent rater training, and diverse task coverage to ensure fair and dependable laboratory practice scoring. G-theory and D-study prove to be valuable tools for enhancing the quality of performance-based assessments in science education.
Downloads
References
Hofstein, A., and Lunetta, V. N. 2004. The Laboratory in Science Education: Foundations for The Twenty-First Century. Science Education, Vol. 88, No. 1, pp. 28–54.
Kang, N. H., and Wallace, C. S. 2005. Secondary Science Teachers’ Use of Laboratory Activities: Linking Epistemological Beliefs, Goals, and Practices. Science Education, Vol. 89, No. 1, pp. 140–165.
Harlen, W., and Qualter, A. 2018. The Teaching Of Science in Primary Schools (7th ed.). Routledge.
Gipps, C. V. 1994. Beyond Testing: Towards a Theory of Educational Assessment. Routledge.
Abrahams, I., and Millar, R. 2008. Does Practical Work Really Work? A Study of The Effectiveness of Practical Work as A Teaching and Learning Method in School Science. International Journal of Science Education, Vol. 30, No. 14, pp. 1945–1969.
Brookhart, S. M. 2013. How to Create and Use Rubrics for Formative Assessment and Grading. ASCD.
Fan, X., and Sun, S. 2014. Generalizability Theory as a Unifying Framework of Measurement Reliability in Adolescent Research. The Journal of Early Adolescence, Vol. 34, No. 1, pp. 38–65.
Darling-Hammond, L., and Adamson, F. 2013. Developing Assessments of Deeper Learning: The Costs and Benefits of Using Tests That Help Students Learn. Stanford Center for Opportunity Policy in Education.
Palm, T. 2008. Performance Assessment and Authentic Assessment: A Conceptual Analysis of The Literature. Practical Assessment, Research & Evaluation, Vol. 13, No. 4, pp. 1–11.
Cronbach, L. J., Gleser, G. C., Nanda, H., and Rajaratnam, N. 1972. The Dependability of Behavioral Measurements: Theory of Generalizability for Scores and Profiles. John Wiley & Sons.
Monteiro, S., Sullivan, G. M., and Chan, T. M. 2020. Generalizability Theory Made Simple(R): An Introductory Primer to G-Studies. Journal of Graduate Medical Education, Vol. 12, No. 4, pp. 365–370.
Bloch, R., and Norman, G. 2012. Generalizability Theory for The Perplexed: A Practical Introduction and Guide: AMEE Guide No. 68. Medical Teacher, Vol. 34, No. 11, pp. 960–992.
Trejo-Mejía, J. A., Sánchez-Mendiola, M., Méndez-Ramírez, I., and Martínez-González, A. 2016. Reliability Analysis of The Objective Structured Clinical Examination Using Generalizability Theory. Medical Teacher, Vol. 38, No. 1, pp. 44–49.
Briesch, A. M., Swaminathan, H., Welsh, M., and Chafouleas, S. M. 2014. Generalizability Theory: A Practical Guide to Study Design, Implementation, and Interpretation. Journal of School Psychology, Vol. 52, No. 1, pp. 13–35.
National Research Council. 2006. America's Lab Report: Investigations in High School Science. The National Academies Press.
Agustian, H. Y., and Seery, M. K. 2017. Reasserting The Role of Pre-Laboratory Activities in Chemistry Education: A Proposed Framework for Their Design. Chemistry Education Research and Practice, Vol. 18, No. 4, pp. 518–532.
Kemendikbudristek. 2022. Peraturan Menteri Pendidikan, Kebudayaan, Riset, dan Teknologi Republik Indonesia Nomor 21 Tahun 2022 tentang Standar Penilaian Pendidikan. Jakarta: Kementerian Pendidikan, Kebudayaan, Riset, dan Teknologi.
Brennan, R. L. 2001. Generalizability Theory. Springer-Verlag.
Shavelson, R. J., and Webb, N. M. 1991. Generalizability Theory: A Primer. Sage Publications.
American Educational Research Association. 2014. American Educational Research Association, American Psychological Association, and National Council on Measurement in Education, Standards for Educational and Psychological Testing. Washington, DC: American Educational Research Association.
Kane, M. T. 2006. Validation. In R. L. Brennan (Ed.), Educational Measurement (4th ed., pp. 17–64). American Council on Education/Praeger.
Lane, S., and Stone, C. A. 2006. Performance Assessment. In R. L. Brennan (Ed.), Educational Measurement (4th ed., pp. 387-431). American Council on Education/Praeger.
Cardinet, J., Johnson, S., and Pini, G. 2010. Applying Generalizability Theory Using EduG. Routledge.
Nunnally, J. C., and Bernstein, I. H. 1994. Psychometric Theory (3rd ed.). McGraw-Hill.
Ruiz-Primo, M. A., and Shavelson, R. J. 1996. Problems and Issues in The Use of Concept Maps in Science Assessment. Journal of Research in Science Teaching, Vol. 33, No. 6, pp. 569–600.
Gott, R., and Duggan, S. 1995. Investigative Work in The Science Curriculum: Developing Science and Technology Education. Open University Press.
Vispoel, W. P., Morris, C. A., and Kilinc, M. 2018. Applications of Generalizability Theory and Their Relations to Classical Test Theory and Structural Equation Modeling. Psychological Methods, Vol. 23, No. 1, pp. 1–26.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 UNESA Journal of Chemical Education

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
The license terms of the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) requires attribution to the original creator, permits non-commercial use, and does not allow for the application of legal or technological restrictions on others' use.
Abstract views: 32
,
PDF Downloads: 30









