Our 2024 evaluations of evaluators

This page presents the results from Giving What We Can’s 2024 iteration of our ongoing project to evaluate impact-focused evaluators and grantmakers (which we collectively refer to as ‘evaluators’). We use this project to decide which evaluators we rely on for our giving recommendations and to advise our cause area funds.

In the 2024 round, we completed three evaluations that informed our donation recommendations for the 2024 giving season. Specifically, we evaluated:

As with our 2023 round, there are substantial limitations to these evaluations, but we nevertheless think that they are a significant improvement to a landscape in which there were no independent evaluations of evaluators’ work.

On this page, we:

  • Share the key takeaways from each of our 2024 evaluations and link to the full reports
  • Include an update explaining our decision to remove The Humane League’s corporate campaigns program from our list of recommended programs
  • Lay out the resulting changes to our recommended programs and cause area funds
  • Describe our process for choosing which evaluators we considered in 2024
  • Share our process for evaluating evaluators in 2024

In future iterations of this project, we aim to conduct new evaluations of our existing vetted evaluators, re-evaluate evaluators we do not yet rely on and expand beyond the seven evaluator programs we’ve covered in 2023 and 2024. We also hope to further improve our methodology.

For more context on why and how we evaluate evaluators, see our main evaluating evaluators page.

Key takeaways from each of our 2024 evaluations

Global health and wellbeing

Founders Pledge Global Health and Development Fund (FP GHDF)

Based on our evaluation, we have decided not to currently include FP GHDF in our list of recommended charities and funds and do not plan to allocate a portion of GWWC’s Global Health and Wellbeing Fund budget to the FP GHDF at this point. However, we still think FP GHDF is worth considering for impact-focused donors and we will continue to host the program on the GWWC donation platform.

Some takeaways that inform this decision include:

  • Our belief that FP GHDF evaluations do not robustly demonstrate that the opportunities they fund exceed their stated bar of 10x GiveDirectly in expectation, due to the presence of errors and insufficiently justified subjective inputs we found in some of their BOTECs that could cause the estimated cost-effectiveness to fall below this threshold.
  • Our uncertainty about the future quality of evaluations following the recent departure of the senior researcher responsible for managing the fund and our belief that there are currently insufficient peer review processes, such as red-teaming, in place to reassure us that the fund quality will remain at least consistent following this transition.

Taken together, these issues left us unable to justifiably conclude that the FP GHDF is currently competitive with GiveWell’s recommendations.

While our decision is not to recommend FP GHDF at this time, we would like to emphasise that we did not conclude that the marginal cost-effectiveness of the GHDF is unambiguously not competitive with GiveWell’s recommended charities and funds — in fact, we think the GHDF might be competitive with GiveWell now or in the near future. Instead, our finding is that we can’t currently justifiably conclude that the fund is competitive with GiveWell in expectation, based on the evidence we have seen in our evaluation and the uncertainty surrounding staffing changes.

We also highlight positive findings from our evaluation. For example, we think that FP GHDF’s strategy of funding promising early-stage organisations plausibly has significant positive externalities, and the fact that FP GHDF has made several grants to organisations before GiveWell indicates they have some track record for identifying these promising funding opportunities.

For more information, please see our 2024 evaluation report for FP GHDF.

Animal welfare

Animal Charity Evaluators’ Movement Grants (ACE MG)

Based on our evaluation, we have decided to include ACE MG in our list of recommended charities and funds and we expect to allocate half of GWWC’s Effective Animal Advocacy Fund to ACE MG.

The key finding informing this decision is that ACE MG’s recent marginal grants and overall grant decision-making look to be of sufficiently high quality to be competitive with our current top recommendation in animal welfare: EA Funds’ Animal Welfare Fund (EA AWF) — by our evaluation of EA AWF in 2023. This is in large part due to our view that ACE MG has improved meaningfully since our previous evaluation, when we concluded that ACE MG performed slightly less strongly on several proxies for the marginal cost-effectiveness of its grants. In particular, we noted improvements in the quality of ACE MG’s:

  • Marginal grantmaking
  • Weighting of geographical neglectedness
  • Approach to grant selection and sizing

We did find what we think to be room for improvement in ACE MG’s approach to more clearly integrating scope, particularly when comparing between applicants implementing different types of interventions. However, we don’t think this room for improvement affects the ACE MG’s position as being — to our knowledge — among the best places to recommend to donors.

We do not believe that our investigation identified ACE MG to be justifiably better or worse than EA AWF in terms of marginal cost-effectiveness — based on our 2023 evaluation of EA AWF — and so we continue to recommend EA AWF as a donation option alongside ACE MG, and to allocate half of our Effective Animal Advocacy Fund to EA AWF.

For more information, please see our 2024 evaluation report for ACE MG.

Animal Charity Evaluators’ Charity Evaluation Program

Based on our evaluation, we have decided not to currently include ACE’s Recommended Charities or the Recommended Charity Fund (RCF) in our list of recommended charities and funds nor to allocate a portion of GWWC’s Effective Animal Advocacy Fund to these programs at this point in time. However, we still think ACE’s Recommended Charities are worth considering for impact-focused donors and we may consider inviting some of these programs to apply to become supported programs on the GWWC donation platform in early 2025.

Some takeaways that informed our decision include:

  • While ACE has made improvements to its approach for capturing scope and cost-effectiveness since our 2023 evaluation, we are not yet convinced that these play a sufficient role in ACE’s ultimate charity recommendation decisions for us to be confident that ACE reliably makes scope sensitive recommendations.
  • We believe that ACE’s Charity Evaluation Program’s current approach does not sufficiently emphasise marginal cost-effectiveness as the most decision-relevant factor in evaluation decisions. For example, ACE primarily evaluates on the basis of organisations’ existing programs, and would benefit from putting additional emphasis on those that would be funded on the margin. This is in direct contrast to ACE’s Movement Grants, which explicitly evaluates programs that ACE would be influencing funding to on the margin.

Additionally, while our decision is not to rely on ACE’s Charity Evaluation program, we would like to emphasise that we are not making the case that the marginal cost-effectiveness of (any particular) ACE Recommended Charity is clearly not competitive with ACE MG or the EA AWF. In fact, we think it is plausible that some of these charities are competitive with our recommended funds, and may be great options to consider for donors with particular worldviews and who have time to delve deeper into ACE’s materials. Instead, our conclusion is that we can’t justifiably conclude that the program consistently identifies charities competitive with the funding opportunities supported by our current recommendations (ACE MG and EA AWF), based on the evidence we have seen in our evaluation.

For more information, please see our 2024 evaluation report for ACE’s Charity Evaluation Program.

Based on the results of this research, we’ve made some changes to our list of recommended programs and management of our cause area fund in the animal welfare cause area. For reasons described above and below, following our 2024 evaluations, we have made the following changes to our list of recommended programs:

  • Removed The Humane League’s corporate campaigns program (see below)
  • Added ACE’s Movement Grants program

Correspondingly, we now plan to allocate 50% of our Effective Animal Advocacy Fund to ACE’s Movement Grants program and 50% to EA Funds’ Animal Welfare Fund (EA AWF), where previously this fund was allocated to 50% EA AWF and 50% The Humane League.

Because we decided to not (yet) rely on the Founders Pledge Global Health and Development Fund, GiveWell — whom we evaluated in 2023 — remains our only vetted evaluator in the global health and wellbeing cause area. As such, we plan to continue:

  • Relying on GiveWell’s funds and Top Charity recommendations for our own recommendations in global health and wellbeing
  • Allocating 100% of our Global Health and Wellbeing Fund to GiveWell’s All Grants Fund

As we did not evaluate any evaluators in the reducing global catastrophic risks cause area, EA Funds’ Long-term Future Fund (EA LTFF) and Longview’s Emerging Challenges Fund (LV ECF) — both of which we evaluated in 2023 — remain our vetted evaluators in this cause area. As such, we plan to continue:

The Humane League’s corporate campaigns program

As part of our review of our recommended programs for the 2024 Giving Season, we have made the decision to remove The Humane League’s (THL) corporate campaigns program from our list of recommended programs. Our reason for this is that we believe we can no longer justifiably recommend THL’s program alongside our existing recommendations in animal welfare.

While our decision was to no longer recommend THL’s program, we want to be clear that our decision does not reflect any negative update on THL’s work. Instead, this is reflective of evidence being out of date, us maintaining the scope of the evaluating evaluators project, and keeping to our principles of usefulness, justifiability and transparency.

Specifically, this conclusion is predominantly based on the following:

  • We generally base our charity and fund recommendations on our evaluating evaluators project, for reasons explained here. This year, the evaluators and grantmakers we’ll rely on – based on our evaluations – will be EA Funds’ Animal Welfare Fund and ACE’s Movement Grants, and neither of these currently recommends THL as a charity (as neither makes charity recommendations).
  • In THL’s case last year, as explained in our report, we went beyond the scope of our usual project by using evidence provided by three further evaluators (Founders Pledge, Rethink Priorities and Open Philanthropy) to supplement the recommendation of an evaluator we had evaluated but decided not to rely on at that point (ACE) in order to ultimately recommend THL’s program. This was based on our overarching principles of usefulness, transparency and justifiability, via our judgement that providing a competitive alternative to the EA AWF – if we transparently and justifiably could – would be valuable.
  • However, because we expect the approach of grantmakers/evaluators to change more slowly than the work/programs of an individual charity, we are not comfortable relying on the information we used last year without undertaking a reinvestigation1, which we don’t expect to do (see below)
  • Given we now have two recommendations in animal welfare, we consider it somewhat less useful to look further into THL (or any other individual program in animal welfare) and we don’t think we can currently justify going beyond the scope of our evaluating evaluators project in the same way as we could last year.

THL’s corporate campaigns remain a supported program on our platform for donors who wish to continue supporting their work, and though we can no longer justifiably recommend them, we still think they are an option worth considering for impact-focused donors with particular worldviews and time to engage with the evidence supporting their corporate campaigns program.

How we chose which evaluators to look into in 2024

We explain the reasons for choosing to evaluate each evaluator within the relevant 2024 report. Among other reasons, in 2024, these choices were informed by:

  • The extent to which we expected evaluators to be impact-focused from our existing knowledge about them.
  • Where our donors give — we wanted to prioritise evaluators whose funds and charity recommendations our donors were currently supporting the most, so that the results would be most useful to them.
  • How likely we thought we would be to reach a justifiable conclusion for the evaluation in our limited timeframe, based on our expectations of the complexity of the evaluation

The choice of which evaluators to prioritise affects our overall recommendations. For example, because we have not yet evaluated Founders Pledge’s Global Catastrophic Risks Fund, we do not recommend the fund. This lack of comprehensiveness is one of our project’s initial key limitations. We try to partially account for this by highlighting promising alternatives to our recommendations on our donation platform, and providing resources for donors to investigate these further.

The process for our 2024 evaluations

As discussed above, a key goal for our evaluations project was to decide which evaluators to rely on for our recommendations and grantmaking. We were additionally interested in providing guidance to other effective giving organisations, providing feedback to evaluators, and improving incentives in the effective giving ecosystem.

For each evaluator, our evaluation aimed to transparently and justifiably come to tailored decisions on whether and how to use its research to inform our recommendations and grantmaking. Though each evaluation is different — because we tried to focus on the most decision-relevant questions per evaluator — the general process was fairly consistent in structure:

  1. We began with a general list of questions we were interested in, and we used this list to generate some initial requests for information.
  2. After receiving this information, we tried to define decision-relevant cruxes of the evaluation: the questions that, when answered, would guide our decisions on whether and how to rely on the evaluator for our recommendations and grantmaking. These differed for each evaluator.
  3. We shared these cruxes along with some additional (more specific) questions with evaluators and domain experts, asking for feedback — in some cases, changing our cruxes as a result.
  4. We then investigated these cruxes — asking further questions and iterating on our overall view — until we felt confident that we could make a justifiable decision. We intermittently tried to share our thinking with evaluators so that we could receive feedback and new information that would help with our investigation.

Note: we were highly time-constrained in completing this second iteration of our project. On average, we spent about 10 researcher-days per organisation we evaluated (in other words, one researcher spending two full work weeks), of which only a limited part could go into direct evaluation work — a lot of time needed to go into planning and scoping, discussion on the project as a whole, communications with evaluators, and so on.