By Matin Durrani
Here in the Physics World office our attention was caught last week by a story in the Times Higher Education. It reported on a lecture given by Simon Singh at the 2:AM conference in Amsterdam, in which the broadcaster, author and former particle physicist criticized some projects that are designed to boost the public’s interest in science, but which, he feels, are not value for money.
The story mentioned several projects facing Singh’s ire, one of which was the 2005 dance Constant Speed that was created to mark the centenary of Einstein’s annus mirabilis. It was commissioned by the Institute of Physics, which publishes Physics World, so naturally Singh’s comments piqued my interest.
Singh apparently told delegates in Amsterdam that “People hate physics, they hate ballet; all you’ve done is allowed people to hate things more efficiently. I just don’t understand how this gets vast amounts of money.” (Incidentally, Constant Speed is a dance, not a ballet; apparently there’s a difference.)
According to Times Higher Education, Singh feels the best engagement consists of “cheap, grass-roots” projects, praising YouTube video series such as Numberphile, Sixty Symbols and Computerphile, which have been watched millions of times over.
I agree with Singh that these kinds of short videos are a great example of conveying the excitement of science to lots of people for relatively little money. In fact, here at Physics World we have our own version called 100 Second Science.
But I’m not sure the two efforts can be compared. Science outreach (the YouTube videos) and art–science collaborations (Constant Speed) aren’t necessarily one and the same thing. Outreach sometimes involves art, but often it doesn’t. What’s more, good art–science collaborations aren’t just about teaching science concepts.
What Singh is right to raise, however, is the importance of ensuring that when money does get spent on public engagement of whatever type, those projects need to be properly organized, accounted for and planned. The sums involved are rarely enormous, but every penny counts and shouldn’t be wasted.
Evaluating the success of any project, however, is a hard thing to do. Is it better to get, say, 100 attendees who already have an interest in science or 50 attendees who previously had never been to a science-outreach event before? I guess it all depends on the purpose of the project.
There are many ways of measuring success but, on pure numbers alone, I’d say it’s more important that science outreach does more than just preach to the converted. That’s why I particularly enjoyed LightFest at the Library of Birmingham last month – an event to mark the International Year of Light that attracted ordinary library-goers who wouldn’t otherwise have gone to something billed as science outreach.
The bottom line, though, is how can you best quantify the success of outreach?
I was kind of hoping you’d answer your question.
Perhaps it would be a good topic for an article for Physics World?
We can, I think, all agree that outreach should involve giving accurate accounts of phenomena of public interest. Sadly, the the 100 Second Science series mentioned STILL includes the fundamentally flawed (indeed, ludicrous) account of how aircraft fly.
I raised this with the IoP Education Dept two years ago, and nothing has been done about replacing it.
Hi Robert…I was involved in those discussions with colleagues in the education department about that particular video. Yes it is still there, but as a result of those discussions, the Physics World team is now more aware of which topics in physics one has to be careful when trying to explain. I’ll get back to you by email about this as I know we started discussing this between the two of us.
Measuring the success of science outreach shouldn’t be difficult. You set certain criteria at the outset, someone funds you for achieving those, either you meet them or not.
The trickier part is agreeing a set of criteria that is measurable.
I suspect that is where Simon Singh might have an issue. Some outreach is done without methods to measure it in place. It doesn’t mean the outreach is wrong, but the evaluation of it certainly is wrong.
From a funder’s point of view there is the additional complication of knowing what criteria to value. Does it really matter as long as they communicate it clearly and consistently?
Often the outreach of any kind is not effective, because it suffers from being casual and loose-ended in its planning, funding and execution.
Please remember to keep all comments completely on-topic. Also, please keep your comments brief — at a maximum of 250 words.
Trackback: » Evaluating outreach