Is it harder to know or to reason? Analyzing two-tier science assessment items using the Rasch measurement model

Asia-Pacific Science Education - Tập 1 Số 1 - 2015
Gavin W. Fulmer1, Hye‐Eun Chu2, David F. Treagust3, Knut Neumann4
1National Institute of Education, Nanyang Technological University, ᅟ, Singapore
2School of Education, Macquarie University, Sydney, Australia
3Science & Mathematics Education Centre, Curtin University, Perth, Australia
4Leibniz Institute for Science and Mathematics Education (IPN), University of Kiel, Kiel, Germany

Tóm tắt

Từ khóa


Tài liệu tham khảo

Adams, R. J., Wu, M. L., & Wilson, M. R. (2012). Reference manual for ACER ConQuest 3. Camberwell, Australia: ACER.

Andersson, B., & Kärrqvist, C. (1983). How Swedish pupils, aged 12–15 years, understand light and its properties. European Journal of Science Education, 5(4), 387–402. doi: 10.1080/0140528830050403

Bao, L., Cai, T., Koenig, K., Fang, K., Han, J., Wang, J., et al. (2009). Learning and scientific reasoning. Science, 323, 586–7.

Bell, B., & Cowie, B. (2001). The characteristics of formative assessment in science education. Science Education, 85(5), 536–553.

Bennett, J., & Hogarth, S. (2009). Would you want to talk to a scientist at a party? High school students’ attitudes to school science and to science. International Journal of Science Education, 31, 1975–98.

Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education, 5, 7–74.

Bond, T. G., & Fox, C. M. (2007). Applying the Rasch Model: Fundamental Measurement in the Human Sciences (2nd ed.). Mahwah: Lawrence Erlbaum Associates.

Boone, W. J., Staver, J. R., & Yale, M. S. (2014). Rasch analysis in the human sciences. Dordrecht, the Netherlands: Springer.

Chandrasegaran, A. L., Treagust, D. F., & Mocerino, M. (2007). The development of a two-tier multiple-choice diagnostic instrument for evaluating secondary school students’ ability to describe and explain chemical using multiple levels of representation. Chemical Education Research and Practice, 8, 293–307.

Chiu, M.-H., Guo, C.-J., & Treagust, D. F. (2007). Assessing students’ conceptual understanding in science: An introduction about a national project in Taiwan. International Journal of Science Education, 29, 379–90.

Chu, H.-E., & Treagust, D. F. (2009). A stratified study of students’ understanding of basic optics concepts in different contexts using two‐tier multiple‐choice items. Research in Science & Technological Education, 27, 253–65.

Chu, H.-E., & Treagust, D. F. (2014). Secondary students’ stable and unstable optics conceptions using contextualized questions. Journal of Science Education and Technology, 23(2), 238–51. doi: 10.1007/s10956-013-9472-6 .

Chu, H.-E., Treagust, D. F, Lim, G. S. E., & Chandrasegaran, A. L. (2015). Efficacy of multiple choice items: Do two-tier multiple-choice diagnostic items have the power to measure students’ conceptions similar to open-ended items? Paper presented at the 11th European Science Education Research Association (ESERA) conference, Helsinki, Finland.

DeVellis, R. F. (2012). Scale development: Theory and Applications (3rd ed.). Thousand Oaks: Sage.

Ding, L. (2014). Verification of causal influences of reasoning skills and epistemology on physics conceptual learning. Physical Review Special Topics - Physics Education Research, 10(2), 023101.

Duit, R. (2009). Bibliography: Students’ alternative frameworks and science education (IPN).

Duit, R., & Treagust, D. F. (2003). Conceptual change: a powerful framework for improving science teaching and learning. International Journal of Science Education, 25, 671–88.

Fetherstonaugh, T., & Treagust, D. F. (1992). Students’ understanding of light and its properties: teaching to engender conceptual change. Science Education, 76, 653–72.

Fulmer, G. W. (2014). Undergraduates’ attitudes toward science and their epistemological beliefs: Positive effects of certainty and authority beliefs. Journal of Science Education and Technology, 23, 198–206. doi: 10.1007/s10956-013-9463-7 .

Fulmer, G. W., Liang, L. L., & Liu, X. (2014). Applying a force and motion learning progression over an extended time span using the Force Concept Inventory. International Journal of Science Education, 36, 2918–36. doi: 10.1080/09500693.2014.939120 .

Galili, I., & Hazan, A. (2000). The influence of an historically oriented course on students’ content knowledge in optics evaluated by means of facets-schemes analysis. American Journal of Physics, 68, S3–15.

Johnson, P., & Tymms, P. (2011). The emergence of a learning progression in middle school chemistry. Journal of Research in Science Teaching, 48, 849–77. doi: 10.1002/tea.20433 .

La Rosa, C., Mayer, M., Patrizi, P., & Vicentini-Missoni, M. (1984). A stratified study of students’ understanding of basic optics concepts in different contexts using two‐tier multiple‐choice items. Research in Science & Technological Education, 27, 253–65.

Langley, D., Ronen, M., & Eylon, B. S. (1997). Light propagation and visual patterns: Preinstruction learners’ conceptions. Journal of Research in Science Teaching, 34, 399–424.

Lawson, A. E. (1978). Development and validation of the classroom test of formal reasoning. Journal of Research in Science Teaching, 15, 11–24.

Lawson, A. E. (2000). Classroom Test of Scientific Reasoning: Multiple choice version. Arizona State University: Author.

Liu, X. (2010). Using and developing measurement instruments in science education. Charlotte: Information Age Publishing.

Liu, O. L., Lee, H.-S., & Linn, M. C. (2011). Measuring knowledge integration: Validation of four-year assessments. Journal of Research in Science Teaching, 48, 1079–107. doi: 10.1002/tea.20441 .

Nehm, R. H., & Ha, M. (2011). Item feature effects in evolution assessment. Journal of Research in Science Teaching, 48, 237–56.

Neumann, I., Neumann, K., & Nehm, R. (2011). Evaluating instrument quality in science education: Rasch-based analyses of a Nature of Science test. International Journal of Science Education, 33, 1373–405. doi: 10.1080/09500693.2010.511297 .

Oliver, M., & Venville, G. (2011). An exploratory case study of Olympiad students’ attitudes towards and passion for science. International Journal of Science Education, 33, 2295–322. doi: 10.1080/09500693.2010.550654 .

Taber, K. S., & Tan, K. C. D. (2011). The insidious nature of ‘hard-core’ alternative conceptions: Implications for the constructivist research programme of patterns in high school students’ and pre-service teachers’ thinking about ionisation energy. International Journal of Science Education, 33, 259–97. doi: 10.1080/09500691003709880 .

Tamir, P. (1971). An alternative approach to the Construction of multiple choice test items. Journal of Biological Education, 5, 305–307.

Tamir, P. (1989). Some issues related to the use of justifications to multiple-choice answers. Journal of Biological Education, 23, 285–92.

Tan, K. C. D., Goh, N. K., Chia, L. S., & Treagust, D. F. (2002). Development and application of a two-tier multiple choice diagnostic instrument to assess high school students’ understanding of inorganic chemistry qualitative analysis. Journal of Research in Science Teaching, 39, 283–301. doi: 10.1002/tea.10023 .

Treagust, D. F. (1988). Development and use of diagnostic tests to evaluate students’ misconceptions in science. International Journal of Science Education, 10, 159–69. doi: 10.1080/0950069880100204 .

Treagust, D. F. (1995). Diagnostic assessment of students’ science concepts. In S. Glynn & R. Duit (Eds.), Learning science in the schools: Research reforming. Mahwah: Lawrence Erlbaum Associates.

Treagust, D. F., Jacobowitz, R., Gallagher, J. L., & Parker, J. (2001). Using assessment as a guide in teaching for understanding: A case study of a middle school science class learning about sound. Science Education, 85, 137–57.

Tsui, C.-Y., & Treagust, D. (2009). Evaluating secondary students’ scientific reasoning in genetics using a two-tier diagnostic instrument. International Journal of Science Education, 32, 1073–98. doi: 10.1080/09500690902951429 .

Wiggins, G., & McTighe, J. (1998). Understanding by design. Alexandria: Association for Supervision and Curriculum Development [ASCD].