I’ve been thinking a bit lately about how I/we/the system records learning. My thinking on this has been challenged a few times, and was kicked into gear way back in 2015 when I saw an amazing learning conversation around a thermos flask take place, that these students couldn’t write down but could explain very clearly, and link other ideas they had learned. (a reflection from the time is here ). More recently, this has been playing on my mind as my yr 10’s have been completing a Science capability task.
Now Science capabilities can be a tricky thing when you get to the nitty gritty of them, but are also amazing simple too. I don’t feel our school has really got to grips with them yet (although we are certainly improving all the time), and we are certainly too focussed on the ‘paper trail’ evidence for assessing them. Which has got me back to thinking more explicitly about how I gather evidence for learning.
With this Science Capabilities task, the context was ocean acidification and how increasing the CO2 in the atmosphere increases the dissolved CO2 in the water, which decreases the pH/increases the acidity of the ocean. This then has an impact on creatures in the ocean with carbonate shells or exoskeletons. We investigated the impact of concentration on acids on the reaction rate for carbonates. Reaction rates really sits at NCEA Level one rather than yr 10, so we have tried to scaffold the prior learning and expectations around reaction rates for students final answers.
As part of the task, students where asked to design an experiment that gave them ‘quality data’ (reproducable, no overlaps etc). Students were given a range of material to try and design their method, as we had previously used gas displacement, a lime water test, amount of bubble produced by adding dish washing liquid etc. There was much fun as the bubbles produced from the highest concentration of acid shot out of the test tubes and all over the bench tops and floors. But this is where the really rick conversations started happening, as students realised that this method was the ‘most fun’, it did not produce the more reliable results. Then others discovered that if they used small volumes of acid, the lower concentrations stopped reacting before the amount of gas produced was detectable. Others found it difficult to distinguish between the two most concentrated acids because the times were very similar.
The learning conversation between the students and the students and myself over these 2-3 hours were amazing. It felt really good, I was excited to be in class, the students seemed to be enjoying it, and the frustrations and successes were palpable. Students were really gaining knowledge from the various practical tasks as they were trying to ascertain the best method for them to use. Students saw that they got ‘the same’ end result while using a different method to another group, and had conversation around which way might be better or worse. The conversations around the ‘why’ the reactions where proceeding that way were amazing.
And yet, while it was a valuable learning experience for me and the students, I feel a lot of that valuable learning was lost. How do I record those conversations that were had? How do I translate that into the students written work, or scaffold the questions better so they are encouraged to incorporate more of that learning into their answers?
Or, do I need to? Is it enough that those conversations took place, and they do not need to be included in the gathered data for reports and feedback?
And, how to I replicate those conversation in other settings?? How do I get students to see more explicitly that there is often more than one way to come to a similar answer, and that discovering the path is often much more exciting than getting to the end of it. How to I make sure more of the learning experiences I offer are open ended??
And how do I ensure I capture the ideas that students don’t write down??