Science Outreach Assessment?

This summer I am working with a student on trying to start to get some reasonable assessment of one of our ongoing oddball-ish outreach projects.  Working with a local videographer, we've been making 2 minute mini-profiles of astrobiologists who work here at LSU (Louisiana State University) - two minute summaries of who they are, what they work on, and why they work on it.  Rather than do the seemingly most typical "talking head" style of video, we've tried to do something that more resembles a mash-up between Bill Nye and Independent Film (yes, it is difficult to do that in your head, but there it is).  Three are in editing mode right now and three more are scheduled to be filmed soon - but the focus of this blog entry is: How do we assess these?

Here is my understanding of how we are "supposed" to do it:  We walk into the astronomy summer camp group that we will be showing them to, and immediately give them a quiz on the content in the videos, THEN we show them the videos, then we give them the same exact quiz again.  For control groups, we are apparently supposed to swap out the videos for a standard talk that has the same content.  You then look at the gain from Test 1 to Test 2 for the 2 different groups.  I first heard about this "Test Them, Teach Them, then Test Them Again" strategy for education assessment only about a year and a half ago, or so - but I've been told that IT IS THE ABSOLUTE NORM.

When I first heard about this assessment strategy I went: huh?  When I describe the strategy to students who are working with me, they go: huh?  When I describe it to other biochemists, they go: huh?  But any education person I talk to says that this is the way it is done.  Clearly, however, the pre-test primes the audience for the information.  It just seems so odd that no-one has come up with a widely acceptable alternative to this pre-test/post-test approach.

I've been wondering if maybe this strategy is a way to avoid having to have large test pools.  It just seems that if you only gave tests after the videos, and your audience numbers were small, of course audience member variations in "prior knowledge" could totally skew the results.  But what if you test, say 150 kids after the videos and 150 kids after the "standard lecture" (no pre-test for either of them) - if you had randomly sorted the 300 into their 2 pools it just seems like any "prior-knowledge bias" of some audience members would be overwhelmed by the numbers.  If your difference between groups were small, you might have to worry - but if the difference is large and statistical, it seems it would be difficult to argue that too many people in one group had too much prior knowledge.

It just seems like either strategy has drawbacks, but for some reason educational assessment researchers have decided that priming people with which answers to look for (by giving them the pre-test) is okay, while risking the probability that some people will have prior knowledge is not okay. Every education researcher I have talked to this about has acknowledge that this is, in fact, one of the choices being made in designing assessment studies this way, but that everybody does it and everybody expects it.

Anyway, next week we are going to start doing some assessment, and I feel confident that we will be learning a lot more about assessment itself as we proceed.

More like this

One of the more contentious recurring topics around here over the years has been education policy, mostly centering around the question of teacher evaluation and teacher's unions. It's probably the subject for which there's the biggest gap between my opinions and those of some of my regular readers…
Kevin Drum has done a couple of education-related posts recently, first noting a story claiming that college kids study less than they used to, and following that up with an anecdotal report on kids these days, from an email correspondent who teaches physics. Kevin's emailer writes of his recent…
I've been pretty quiet about educational matters of late, for the simple reason that I was too busy teaching to say much. The dust having settled a bit, though, I thought I would put some notes here about what I did this past term, and what worked. I had two sections of the introductory Newtonian…
As noted in previous posts, I've been trying something radically different with this term's classes, working to minimize the time I spend lecturing, and replace it with in-class discussion and "clicker questions." I'm typing this while proctoring the final exam for the second of the two classes I'm…

It sounds like a lovely project. I was just curious how you decided to assess it in the end and if it worked?