Got a lot of angry replies to this one so here's a more in-depth thread about why the Behavioural Insights Team irritates me so much... https://twitter.com/sophie_e_hill/status/1238400198155649029
1⃣ The Nudge Unit continually cites research that has been debunked or doesn't replicate.

e.g. their "menu for change" project, which they say is based on "the latest and most well-evidenced behavioural science" cites Brian Wansink 6 times!! https://www.bi.team/publications/a-menu-for-change/
2⃣ Sometimes when the Nudge Unit cites credible academic research they... don't seem to have actually read it?

Here's a blog post where they claim that mandatory sexual harassment training can backfire, with a link to this meta-analysis:

https://www.bi.team/blogs/more-than-a-few-bad-apples-what-behavioural-science-tells-us-about-reducing-sexual-harassment/
So, unlike the Nudge Unit, I actually read this paper.

It finds:
-mandatory trainings are more effective at changing behaviour
-mandatory trainings produce positive reactions, though voluntary trainings produce even MORE positive reactions

That's not the same as a backlash!!?
P.S. Of course, there is some research out there that argues for the "backlash" effect, though it was not cited by the Nudge Unit here and I think it's deeply flawed for other reasons! https://twitter.com/sophie_e_hill/status/1201610408362156032?s=20
3⃣ The Nudge Unit frequently draws strong conclusions from quasi-experimental studies, without acknowledging how these designs complicate the inferences we can draw.
For example, here's a recent paper published on their interventions on how UK jobseekers benefit is administered.

David Halpern has frequently cited this work as the work he is most proud of.

https://www.cambridge.org/core/journals/journal-of-public-policy/article/behavioural-insight-and-the-labour-market-evidence-from-a-pilot-study-and-a-large-steppedwedge-controlled-trial/F953A370AD7D8A0427FF0772098D7C74
The goal of this intervention is to increase "off-flow". Basically to try to get people to stop claiming jobseekers benefit as quickly as possible.

(Note: they are not measuring transitions into work, just when people stop claiming the benefit...)
The first "experiment" was a pilot study in Loughton Jobcentre.

Treatment assignment was not randomized. It was based on whether a claimant was allocated to an adviser on the first floor or ground floor of the Jobcentre.

But don't worry, the allocation was arbitrary.
Narrator: it was not arbitrary.
They also find a pretty big treatment effect on the, uh, control group? Once again raising the concern that the "invervention" here was basically a ton of political pressure to get people off benefits, rather than a new way of helping people look for work.
Never mind, that was just a pilot. Let's move on to the MAIN experiment. I'm sure this one will be a lot more rigorous.
ok good good ok hmm WHAT
So to summarise, we have 9 clusters of Jobcentres. They have been put into 3 strata based on a subjective assessment of "readiness". The most "ready" clusters will receive the intervention first. But they're going to randomise the order within strata.
So there are two basic issues:

1. In any stepped wedge design, it's hard to disentangle the effect of treatment from a secular time trend. But it's even worse here because the roll-out explicitly conditions on Jobcentre readiness, which is plausibly related to the outcome.
There's almost zero discussion of these very very obvious concerns in the paper. They don't even include strata fixed effects in the main results, you have to go to the appendix for that. No reference to, e.g. the wild cluster bootstrap.

Reviewer #2, where u at?
Here's their takeaway:

"it is not possible to be entirely unequivocal about these results"

but also

"we are therefore able to be unequivocal in arguing that its national rollout is justified and backed by evidence"

????
4⃣ It's not just problems with the experimental design. Studies published by the Nudge Unit show alarmingly frequent issues with basic implementation.

Here's a study where they lost 75% (!) of their data lmao
actually it's more like 82% ?? they recruited 4,000 participants but were only left with 750
Actual quote (this is Halpern talking about research on social capital):

"The academic literature ran out of road at that point. But we have been able to work on actual interventions."
What is this intervention that the academic community has been wilfully ignoring?

Different types of 10 minute icebreakers.
and, uh, here are the results 🧐
Don't worry though, I'm sure people like David Halpern and their spectacularly-sized egos aren't going to cause too much harm by, say, eroding public trust in science and ultimately causing the deaths of many people! Everything is fine! https://twitter.com/BBCMarkEaston/status/1237694665824047111?s=20
You can follow @sophie_e_hill.
Tip: mention @twtextapp on a Twitter thread with the keyword “unroll” to get a link to it.

Latest Threads Unrolled: