Our 2025 evaluations of evaluators

This page presents the results from Giving What We Can’s 2025 iteration of our ongoing project to evaluate impact-focused evaluators and grantmakers (which we collectively refer to as ‘evaluators’). We use this project to decide which evaluators we rely on for our giving recommendations and to advise our cause area funds.

In the 2025 round, we completed two evaluations that informed our donation recommendations in global health and wellbeing for the 2025 giving season and beyond. Specifically, we evaluated:

As with our 2023 and 2024 rounds, there are substantial limitations to these evaluations, but we nevertheless think that they are a significant improvement to a landscape in which there were previously no independent evaluations of evaluators’ work.

On this page, we:

In future iterations of this project, we aim to conduct new evaluations of our existing vetted evaluators, re-evaluate evaluators we do not yet rely on and expand beyond the eight evaluator programmes we’ve covered so far. We also hope to further improve our methodology.

For more context on why and how we evaluate evaluators, see our main evaluating evaluators page.

Key takeaways from each of our 2025 evaluations

GiveWell

Based on our evaluation, we have decided to continue including GiveWell's Top Charities, Top Charities Fund and All Grants Fund in GWWC's list of recommended programmes and to continue allocating a portion of GWWC's Global Health and Wellbeing Fund to GiveWell's All Grants Fund. As GiveWell met our bar in our 2023 evaluation, our task was to determine whether their evaluation quality had been maintained and whether there were significant issues we had previously missed. Our decision is based on two main considerations:

  1. Firstly, we continue to think GiveWell's approach serves a variety of sufficiently plausible worldviews amongst donors who prioritise promoting near-term human health and wellbeing
  2. Secondly, our quality checks on one Top Charity evaluation (Helen Keller Intl's vitamin A supplementation programme) and two marginal grant evaluations (Taimaka's malnutrition treatment and Technical Support Unit grants) imply that GiveWell maintains high evaluation standards, with careful validation of inputs, comprehensive analyses, and effective use of external expertise

We also note GiveWell's progress across all three areas for improvement we identified in our 2023 evaluation — transparency and legibility, forecast reviews, and uncertainty handling — demonstrating their commitment to continuous improvement. Whilst we think there remains room for further improvement, particularly in the legibility of grant evaluations and justification of subjective inputs, we continue to think that GiveWell's reputation for providing high-quality recommendations and grants is justified, and we expect them to maintain their position as a leading source of impact-focused recommendations in global health and wellbeing.

For more information, please see our 2025 evaluation report for GiveWell.

The Happier Lives Institute (HLI)

Following our 2025 investigation of the Happier Lives Institute (HLI), we've decided not (yet) to include HLI's recommended charities in our list of recommended charities and funds and do not plan to allocate a portion of GWWC's Global Health and Wellbeing Fund budget to HLI's recommended charities at this point. However, this was an unusually difficult decision, and reasonable people could disagree with our conclusion.

HLI is filling an important gap — identifying opportunities for donors who strongly prioritise life improvements over life-saving benefits. We found much to appreciate:

  • Their reports are very transparent and comprehensive, with their psychotherapy evaluation receiving a positive independent review from the Unjournal
  • The researchers we interacted with were thoughtful, open-minded, and genuinely responsive to feedback
  • They've made impressive contributions: identifying issues in GiveWell's deworming analysis, conducting foundational wellbeing research, and sparking valuable discussions about how philosophical beliefs affect one’s view of charity effectiveness
  • They've made remarkable progress given their small team and organisational age and we think their recommendations are already worth considering as highly promising donation options

Despite these strengths, we couldn't confidently conclude that HLI's process reliably identifies opportunities at least as cost-effective as GiveWell's for donors with a life-improving focus:

  1. Competitiveness of GiveWell Top Charities: When we adapted HLI's AMF evaluation to use assumptions from their recent Taimaka evaluation, AMF's cost-effectiveness increased substantially — with a (highly uncertain) point estimate comparable to HLI's Top Charity recommendations. This used only HLI's own assumptions, suggesting GiveWell Top Charities may be reasonably competitive under HLI's worldview.
  2. Implementation evidence: We have concerns about how HLI weighs charity-specific evidence for psychotherapy programmes, particularly given Friendship Bench's low attendance rates and M&E results indicating substantially smaller effects than meta-analytic estimates.
  3. Process maturity: Whilst we think HLI has made impressive progress given their organisational age and capacity, there remain aspects of their evaluation processes that we would like to see more developed before relying on them. These include clearer recommendation criteria and more consistent assumptions across evaluations.

We emphasise: our conclusion is consistent with HLI's charities being highly cost-effective and potentially even more cost-effective than GiveWell Top Charities by HLI’s worldview. Our concern is about process reliability, not individual charity quality. We'll continue considering HLI's recommendations for our 'other supported programmes', and we believe their recommendations are worth consideration by donors — particularly those with strong life-improving preferences.

We're optimistic about HLI's trajectory and look forward to re-evaluating them in a future round of evaluating evaluators.

For more information, please see our 2025 evaluation report for HLI.

Because we decided to not (yet) rely on the Happier Lives Institute, GiveWell — whom we re-evaluated in 2025 — remains the only evaluator we rely on the global health and wellbeing cause area. As such, we plan to continue:

  • Relying on GiveWell’s funds and Top Charity recommendations for our own recommendations in global health and wellbeing
  • Allocating 100% of our Global Health and Wellbeing Fund to GiveWell’s All Grants Fund

The process for our 2025 evaluations

As discussed above, a key goal for our evaluations project was to decide which evaluators to rely on for our recommendations and grantmaking. We were additionally interested in providing guidance to other effective giving organisations, providing feedback to evaluators, and improving incentives in the effective giving ecosystem.

For each evaluator, our evaluation aimed to transparently and justifiably come to tailored decisions on whether and how to use its research to inform our recommendations and grantmaking. Though each evaluation is different — because we tried to focus on the most decision-relevant questions per evaluator — the general process was fairly consistent in structure:

  1. We began with a general list of questions we were interested in, and we used this list to generate some initial requests for information.
  2. After receiving this information, we tried to define decision-relevant cruxes of the evaluation: the questions that, when answered, would guide our decisions on whether and how to rely on the evaluator for our recommendations and grantmaking. These differed for each evaluator.
  3. We shared these cruxes along with some additional (more specific) questions with evaluators and domain experts, asking for feedback — in some cases, changing our cruxes as a result.
  4. We then investigated these cruxes — asking further questions and iterating on our overall view — until we felt confident that we could make a justifiable decision. We intermittently tried to share our thinking with evaluators so that we could receive feedback and new information that would help with our investigation.

Note: we were highly time-constrained in completing this second iteration of our project. On average, we spent about 20 researcher-days per organisation we evaluated (in other words, one researcher spending four full work weeks), of which only a limited part could go into direct evaluation work — a lot of time needed to go into planning and scoping, discussion on the project as a whole, communications with evaluators, and so on.