Tests show U.S. students struggle to explain answers


A second type of test, Interactive Computer Tasks, went beyond what had previously been measured, testing how students ran their own experiments in simulated environments.

American children do much better identifying the correct answers to simple scientific tasks than using evidence from their experiments to explain those answers.

The National Assessment of Educational Progress, often called the Nation’s Report Card, asked students in grades four, eight and 12 to perform actual experiments to apply principles they learn in the classroom on a practical level. The results of the 2009 tests were released June 19.

“That tells us that our science teaching isn’t getting us as far as we need to go,” said Chris Dede, professor from Harvard Graduate School of Education.

Katherine Carroll, an 11th- and 12th-grade chemistry teacher in Waterboro, Maine, said even her best students struggle to explain their conclusions in the lab reports they turn in for her class. She found them more accustomed to questions with one right answer.

“Teachers have moved towards teaching more knowledge, as opposed to the understanding behind that knowledge,” Carroll said.

Like Carroll, Dede said kids’ difficulty explaining is old news to most teachers and parents, but this is the first time they have concrete evidence demonstrating the problem.

“Having something that is more than just anecdotes, that is rigorous research across a wide range of students, is very helpful, because it’s a better form of evidence on which to make decisions,” Dede said.

The first test, called Hands On Tasks (HOTs), allotted students 40 minutes to conduct experiments with physical objects. This allowed for a richer analysis of their understanding of the subject than pencil and paper tests can provide, according to Alan Friedman, chairman of the National Assessment Governing Board’s Assessment Development Committee.

HOTs, however, are nothing new. NAEP tests used them as far back as 1996.

Friedman said the second type of test, Interactive Computer Tasks (ICTs), went beyond what had previously been measured, testing how students ran their own experiments in simulated natural or laboratory environments with the ability to go back, adjust variables, and correct their mistakes on a computer.

“This is a set of skills which in the real world is invaluable,” Friedman said, “and which before this we’d never been able to know if students could do this or not.”

Though Friedman said the computer tests are “dramatically more expensive” to design, traditional assessments cannot measure these same skills.

During ICTs, just over a quarter of high school seniors could both select and explain their correct answers about heating and cooling. Double that amount—54 percent—in the eighth grade group could support correct conclusions with evidence, but only 15 percent of fourth grade students could do the same in their experiment.

Want to share a great resource? Let us know at submissions@eschoolmedia.com.

 

We’re Celebrating 25 Years with 25 Giveaways!

Enter Each Day to Win the Daily Gift Card Giveaway

and the Grand Prize drawing for an

Apple iPad!


Visit eSchool News each day through April 1, 2023 to enter the daily $25 Gift Card drawing.
Each daily entry counts as one entry for the grand prize drawing. See details and rules.
Giveaway is open only to legal residents of the fifty (50) United States and Canada who are employed full- or part-time in K-12 education.