I live for a high disagree-to-upvote ratio
PurpleAir contribute all of their sensors to IQAir also! So you can get a very comprehensive sense of air quality very quickly and compare private and public sources.
I suspect the synthesis here is that unguided is very effective when adhered to, but the main challenge is adherence. The reason to believe this is that there is usually a strong dosage effect in psychotherapy studies, and that the Furukawa study I posted in the first comment found that the only value humans provided was for adherence, not effect size.
Unfortunately, this would then cause big problems, because there is likely a trial bias affecting adherence, potentially inflating estimates by 4× against real-world data. I’m surprised that this isn’t covered in the literature, and my surprise is probably good evidence that I have something wrong here. This is one of the reasons I’m keen to study our intervention’s real-world data in a comparative RCT.
You make a strong point about the for-profit space and relative incentives, which is partly why, when I had to make a decision between founding a for-profit unguided app and joining Kaya Guides, I chose the guided option. As you note, the way the incentives seem to work is that large for-profits can serve LMICs only when profit margins are competitive with expanding further in HICs. This is the case for unguided apps, because translation and adaptation is a cheap fixed cost. But as soon as you have marginal costs, like hiring humans (or buying books, or possibly, paying for AI compute), it stops making sense. This is why BetterHelp have only now begun to expand beyond the U.S. to other rich countries.
But I think you implicitly note—if one intervention has zero marginal cost, then surely it’s going to be more cost-effective and therefore more attractive to funders? One model I’ve wondered about for an unguided for-profit is essentially licensing its core technology and brand to a non-profit at cost, which would then receive donations, do translations, and distribute in other markets.
The devil’s in the details here. The meta-analysis you cite includes an overall estimate for unguided self-help which aggregates over different control condition types (waitlist, care-as-usual, others). When breaking down by control condition, and adding Karyotaki et al. (2021), which looks at guided vs unguided internet-based studies:
Now, Tong et al. (2024) does seem to find higher effects in general, but 32% of the interventions studied included regular human encouragement. These conditions found effect sizes of 0.62, compared to 0.47 for no support (I wish these would disaggregate against control conditions, but alas).
Tong has significantly more self-guided studies than Cuijpers. When both limit to just low risk-of-bias studies, Tong reports an effect sizes of 0.43 (not disaggregated across controls, unfortunately). Cuijpers reports 0.44 for waitlist controls, and 0.13 for care as usual, for the same restriction. So Tong has included more high risk-of-bias studies, which is quite dramatically lifting their effect size.
Now, as Cuijpers and Karyotaki are both co-authors on the Tong analysis, I’m sure that there’s value in including those extra studies, and Tong probably makes me slightly update on the effectiveness of unguided self-help. But I would be wary about concluding that Tong is ‘normal’ and Cuijpers is ‘unusually pessimistic’; probably the inverse is true.
(All up, though, I think that it’s quite likely that unguided interventions could end up being more optimally cost-effective than guided ones, and I’m excited to see how this space develops. I’d definitely encourage people to try and pursue this more! I don’t think the case for unguided interventions rests on their relative effectiveness, but much more on their relative costs.)
I don’t have a strong sense here. The Jesus-Romero study is good and persuasive, but to be convinced, I’d want to see a study of revealed preferences rather than stated ones. To illustrate, one reason we think apps might be better is because our digital ads reach users while they’re on media apps like Instagram, which is a behaviour people are probably likely to engage in when depressed. I think there’s probably a much lower activity threshold to click on an ad and book a call with us, than there is to remember you ordered/were sent a book and to do an exercises from it.
Regardless, it’s likely that digital interventions are much cheaper (again, probably about $1 per participant in engineering vs. ~$5 (??) for book printing and delivery, assuming both interventions spend $1 on social media), and can scale much faster (printing and delivery requires a lot of extra coordination and volume prediction). There’s a good reason many for-profit industries have digitised; it’s just much cheaper and more flexible.
G’day, welcome to the Forum!
I help lead a highly cost-effective EA bibliotherapy charity in India. I agree with most of your points, and in fact, Charity Entrepreneurship’s original report into cost-effective psychotherapy interventions recommended buying physical books and distributing in much the same way you suggest. My charity, Kaya Guides, was incubated from this idea by CE in 2022. We have learned a lot since then, that might add some colour to your post:
Apps are much cheaper than physical books: We deliver our intervention over WhatsApp, which, at scale, will cost us a maximum of like, 20c per participant. The only reason not to do this is in internet-poor countries, but India has a high and rapidly accelerating rate of internet adoption among their under-resourced depressed population.
You don’t need to distribute randomly, just use digital ads: We recruit new participants for ~$0.60 via Instagram ads that bounce users directly to WhatsApp. Meta are a trillion dollar company because they are extremely good at identifying people who might want to use your product; ~70% of users who make it to our depression questionnaire screen for depression.
Adding weekly calls likely doubles to quadruples your adherence: The bibliotherapy you’re referring to, as far as I can infer from your cost model, is also known as ‘unguided self-help’. I should note that the meta-analyses you link are mostly of ‘guided self-help’, which is when the books are paired with regular contact with a therapist. Guided self-help has indistinguishable effect sizes from individual and group therapy. You may be interested in this 2019 meta-analysis which looks at 155 RCTs of different delivery formats for CBT, which finds that, relative to care as usual, guided self-help has double the effect size of unguided self-help. The reasons why aren’t perfectly understood, but the general belief is that the therapists don’t provide much in the way of extra help, but just provide a sense of accountability that helps participants make their way through the entire curriculum. See this study, which found that ‘human encouragement’ had significant effects on retention, but no unique-to-human element had a significant effect on effect sizes directly.
FWIW, I wrote about the difference between unguided and guided self-help previously on the Forum; the main things that have changed in my thinking since is that (a) you can get the cost of counsellors down further than previously thought, including if you augment them with AI, and (b) the adherence of unguided interventions seems lower than I thought. Anyway, we’ll be doing an RCT on this soon hopefully to put the issue to bed a bit more thoroughly :)
I also have updated estimates of the cost-effectiveness relative to cash transfers. Using the Happier Lives Institute’s methodology, we probably create around 1.28 wellbeing-years of value per participant (including spillovers to other members of their household), which is a bit less than top group therapy charities such as StrongMinds and Friendship Bench. But we currently treat people for around $50 per person, should reach ~$20 per person next year, and ~$5–7 per person at scale (I intend to write a post on this soon). Cash transfers, meanwhile, create about 9.22 wellbeing-years when administered by GiveDirectly, and cost about $1,200.
Costly signals like hunger strikes are only likely to persuade public opinion if the public actually hears about them, which is only going to happen if the issue already has some level of public salience. (Whereas protests are better for building public salience, because they’re better suited for mass turnout).
I then also argued that a costly signal like this is unlikely to persuade people at Anthropic, who are already unusually familiar with the debate around AI safety or pausing AI and don’t hold a duty of care over the protestor. It’s at this point that the harms (including inspiring other hunger strikers) overwhelm the benefits.
I think we can cleave a reasonable boundary around hunger strikes specifically. They work well when the person you’re striking against has a duty of care over you so is forced to address your protest (ex. Guantanamo Bay), or if you’ll garner significant attention (ex. Gandhi). These, I think, reasonably outweigh the harms to the subjects and people they might inspire.
Risking other forms of harm in a protest has a different character. This is because those harms are often inflicted by the subject of the protest, which fairly reliably causes backlash or exposes some inconvenient aspect of the oppressor (ex. Civil rights). This risk is significant, but IMHO can be justified by its likelihood to further the aims of the protest.
(I guess it’s possible to believe that it’s morally good to face harms when protesting even if you’re absolutely certain you will achieve nothing, because you believe that self-sacrifice for your protest is good in and of itself. But I suspect in your example the only option for protest was to face harm, whereas this person has extensive non-harmful avenues available to them)
I don’t have a good heuristic for distinguishing between a hunger strike as legitimate protest and delusional self-harm, but this feels more toward the latter.
I think we can game this out, though. The purpose of a hunger strike as protest is to (a) generate public attention (it’s an extreme, costly signal, which can cause other people to update their beliefs) and (b) coerce an actor into doing something to prevent the nominal cause of your strike.
In this case, we can clearly rule out (b). But I also think that (a) is highly unlikely—AI safety isn’t an issue on the verge of public salience, and Anthropic aren’t even the most famous company neglecting it. In almost all cases, this will merely garner attention among a group of people who are already amenable to their message within EA and rationalism. Cynically, we might even recognise this has having a third, hidden purpose, which is to demonstrate fealty to others in the AI safety radicalism movement.
I would suggest that as the benefits of such a protest are minimal, we should turn to the harms. Not only is this person harming themself, if this gets attention within their ingroups, they’re encouraging others to do the same for the same attention. As such, I’ve strongly downvoted this post and reported it to the moderation team. I strongly believe we should not be making this behaviour visible to others.
Yep—Beast Philanthropy actually did an AMA here in the past! My takeaway was that the video comes first, so that your chances of a partnership would greatly increase if you can make it entertaining. This is somewhat in contrast with a lot of EA charities, which are quite boring, but I suspect on the margins you could find something good.
What IMHO worked for GiveDirectly in that video, and for Shrimp Welfare in their public outreach, has been the counterintuitiveness of some of these interventions. Wild animals, cultured meat, shrimp, are more likely to fit in this bucket than corporate campaigns for chickens I reckon.