Abstract
Researchers are trying to develop assessments for inquiry practices to elicit students’ deep science learning, but few studies have examined the relationship between students’ doing, i.e. performance assessment, and writing, i.e. open responses, during inquiry. Inquiry practices include generating hypotheses, collecting data, interpreting data, warranting claims, and communicating findings [1]. The first four practices involve “doing” science, whereas the last involves writing scientific explanations, i.e. arguing using evidence. In this study, we explored whether what students wrote in their constructed responses reflected what they did during science inquiry in the Inq-ITS system. Results showed that more than half of the students’ writing did not match what they did in the environment. Findings revealed multiple types of students in the messy middle, which has implications for both teacher instruction and intelligent tutoring systems, such as Inq-ITS, in terms of providing real-time feedback for students to address the full complement of inquiry practices [1].
The original version of this chapter was revised. The spelling of the third author’s name was corrected. The erratum to this chapter is available at https://doi.org/10.1007/978-3-319-61425-0_83
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Generation Science Standards Lead States: Next generation science standards: for states, by states. National Academies Press, Washington (2013)
National Research Council: A framework for K-12 science education: practices, crosscutting concepts, and core ideas. National Academies Press, Washington (2012)
Gobert, J.D., Sao Pedro, M., Raziuddin, J., Baker, R.S.: From log files to assessment metrics: measuring students’ science inquiry skills using educational data mining. J. Learn. Sci. 22, 521–563 (2013). doi:10.1080/10508406.2013.837391
Shute, V.J.: Stealth assessment in computer-based games to support learning. In: Tobias, S., Fletcher, J.D. (eds.) Computer Games and Instruction, pp. 503–524. IAP (2011)
Sao Pedro, M.A., Gobert, J., Baker, R.S.: The development and transfer of data collection inquiry skills across physical science microworlds. In: Paper presented at the American Educational Research Association Conference (2012)
Sao Pedro, M., Baker, R., Gobert, J., Montalvo, O., Nakama, A.: Leveraging machine-learned detectors of systematic inquiry behavior to estimate and predict transfer of inquiry skill. User Model. User Adap. 23, 1–39 (2013)
Chen, Z., Klahr, D.: All other things being equal: acquisition and transfer of the control of variables strategy. Child Dev. 70, 1098–1120 (1999)
Tschirgi, J.E.: Sensible reasoning: a hypothesis about hypotheses. Child Dev. 51, 1–10 (1990)
Baker, R.S., Clarke-Midura, J., Ocumpaugh, J.: Towards general models of effective science inquiry in virtual performance assessments. J. Comp. Assist. Learn. 32, 267–280 (2016)
Gotwals, A.W., Songer, N.B.: Reasoning up and down a food chain: using an assessment framework to investigate students’ middle knowledge. Sci. Educ. 94, 259–281 (2010)
Roberts, R., Gott, R., Glaesser, J.: Students’ approaches to open-ended science investigation: the importance of substantive and procedural understanding. Res. Pap. Ed. 25, 377–407 (2010). doi:10.1080/02671520902980680
Roberts, R., Johnson, P.: Understanding the quality of data: a concept map for ‘the thinking behind the doing’ in scientific practice. Curriculum J. 26, 345–369 (2015)
Rupp, A.A., Gushta, M., Mislevy, R.J., Shaffer, D.W.: Evidence-centered design of epistemic games: measurement principles for complex learning environments. J. Technol. Learn. Assess. 8, 1–48 (2010)
National Research Council: Knowing what students know: the science and design of educational assessment. In: Pellegrino, J.W., Chudowsky, N., Glaser, R. (eds.) National Academies Press, Washington (2001)
Quellmalz, E., Timms, M., Schneider, S.: Assessment of student learning in science simulations and games. National Research Council, Washington (2009)
McNeill, K., Lizotte, D.J., Krajcik, J., Marx, R.W.: Supporting students’ construction of scientific explanations by fading scaffolds in instructional materials. J. Learn. Sci. 15, 153–191 (2006). doi:10.1207/s15327809jls1502_1
Liu, O.L., Lee, H.S., Linn, M.C.: Multifaceted assessment of inquiry-based science learning. Ed. Assess. 15, 69–86 (2010). doi:10.1080/10627197.2010.491067
Martınez-Cortizas, A., Pontevedra-Pombal, X., Garcıa-Rodeja, E., Novoa-Munoz, J.C., Shotyk, W.: Mercury in a Spanish peat bog: archive of climate change and atmospheric metal deposition. Science 284, 939–942 (1999)
Toulmin, S.: The Uses of Argument. Cambridge University Press, Cambridge (1958)
Sadler, T.D.: Informal reasoning regarding socioscientific issues: a critical review of research. J. Res. Sci. Teach. 41, 513–536 (2004)
Sandoval, W.A., Millwood, K.A.: The quality of students’ use of evidence in written scientific explanations. Cogn. Instruct. 23, 23–55 (2005)
Hogan, K., Maglienti, M.: Comparing the epistemological underpinnings of students’ and scientists’ reasoning about conclusions. J. Res. Sci. Teach. 38, 663–687 (2001). doi:10.1002/tea.1025
Gobert, J.D., Pallant, A.R., Daniels, J.T.: Unpacking inquiry skills from content knowledge in geoscience: a research and development study with implications for assessment design. Int. J. Learn. Technol. 5, 310–334 (2010)
Glaesser, J., Gott, R., Roberts, R., Cooper, B.: The roles of substantive and procedural understanding in open-ended science investigations: using fuzzy set qualitative comparative analysis to compare two different tasks. Res. Sci. Ed. 39, 595–624 (2009). doi:10.1007/s11165-008-9108-7
Shrout, P.E., Fleiss, J.L.: Intraclass correlations: uses in assessing rater reliability. Psychol. Bull. 86, 420–428 (1979). doi:10.1037/0033-2909.86.2.420
Acknowledgement
The research reported here was supported by Institute of Education Sciences (R305A120778) to Janice Gobert.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Li, H., Gobert, J., Dickler, R. (2017). Dusting Off the Messy Middle: Assessing Students’ Inquiry Skills Through Doing and Writing. In: André, E., Baker, R., Hu, X., Rodrigo, M., du Boulay, B. (eds) Artificial Intelligence in Education. AIED 2017. Lecture Notes in Computer Science(), vol 10331. Springer, Cham. https://doi.org/10.1007/978-3-319-61425-0_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-61425-0_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-61424-3
Online ISBN: 978-3-319-61425-0
eBook Packages: Computer ScienceComputer Science (R0)