Published using Google Docs
How Important is Exploration and Prioritization? (Apr 2025 Published Version)
Updated automatically every 5 minutes

How Important is Exploration and Prioritization?

Written: May 2022.  Published: Apr 2025.

In this document I demonstrate that under many circumstances, one should spend lots of effort (and likely more than commonly assumed[1]) on exploration (i.e. looking for new projects) and prioritization (i.e. comparing known projects), rather than exploitation (i.e. directly working on a project). Furthermore, I provide heuristics that help one decide how much effort to spend on exploration and prioritization.    

I assume that the goal is to maximize expected altruistic impact, where by “expected” I’m referring to the mathematical expectation. The actor here can be an individual or a community, and a project can be anything ranging from a 1-week personal project to an entire cause area.

Section 1 will deal with exploration, section 2 will deal with prioritization, and section 3 will test and apply the theory in the real world. Mathematical details can be found in the appendix.

Key Findings

This is an explorative study, so the results are generally of low confidence, and everything below should be seens as hypotheses. The main goal is to suggest paths for future research, rather than to provide definitive answers.

Theoretical Results

Practical Results

Instrumental Results

Note that the following claims are results (based on mostly qualitative arguments) rather than assumptions.

Important Limitations


1 Exploration: Discovering New Projects

1.1 Impact of All Potential Projects is Heavy-Tailed

First we need to figure out the impact distribution of all potential projects. As an example, we can take a look at the cost-effectiveness distribution of health interventions: (source)

This is a heavy-tailed distribution, which means roughly that a non-negligible portion of all interventions manages to reach a cost-effectiveness that is orders of magnitude higher than the median one. As another piece of evidence, the success of for-profit startups also follows a heavy-tailed distribution (the for-profit startup space is arguably analogous to the philanthropy space).

This heavy-tail phenomena occurs in the longtermist space as well. For example, the classical argument for longtermism holds that reducing existential risk has an EV orders of magnitude higher than most other altruistic interventions. Carl Shulman’s comments to this GiveWell blog post also fleshed out a case for heavy-tailed distribution of intervention effectiveness when taking into account longtermist interventions.

On the other hand, Brian Tomasik has argued that charities usually don't differ astronomically in cost-effectiveness, based on several considerations including, most notably, flow-through effects. However:

  1. When non-TOC impact doesn’t matter to our specific purpose of investigation, or
  2. When we’re (almost) completely clueless about the sign and size of non-TOC impact (even after investigating it), or
  3. When we only care about TOC impact and not the spooky non-TOC impact, or
  4. When non-TOC impact is strongly and positively correlated with TOC impact, or
  5. When non-TOC impact isn’t large (compared to TOC impact) for top projects. More concretely, we need the following statements to hold:

From now on I’ll use “total impact” to refer to the sum of TOC and non-TOC impact.

To sum up,

Claim 1: TOC impact of different projects follow a heavy-tailed distribution.

1.2 TOC Impact of All Potential Projects Follows a Power Law

When talking about heavy-tailed distributions, two distributions come to mind: the log-normal distribution, and the Pareto distribution (aka power law). These two distributions are perhaps the two most common heavy-tailed distributions in real world, with the latter being much more heavy-tailed than the former.

For the sake of convenience, we shall adopt one of these two distributions as an approximation of the TOC impact distribution. Here I argue that Pareto distribution (power law) is a better approximation than log-normal distribution. Note that these two distributions are notoriously hard to distinguish empirically, which is part of the reason why I have relatively low confidence in this conclusion.

A commonly seen argument for the log-normalness of TOC impact is that:

  1. When estimating TOC impact, we usually multiply many factors together to get a cost-effectiveness value (with the unit being DALY/$ or something like that).
  2. According to the central limit theorem, by multiplying many independent random factors together, the product that we get follows a log-normal distribution.
  3. Therefore, cost-effectiveness values of different projects follow a log-normal distribution.

This argument has a big flaw. The central limit theorem assumes that there are a fixed number of random factors, while it’s usually not the case for EA prioritization. For example, it’s hard to believe that evaluating global poverty interventions involves that same set of factors as evaluating X-risk reduction interventions. Indeed, it’s exactly the factors related to the long-term impact of human civilization (e.g. the size of the galaxy), which are absent in the evaluation of global poverty interventions, that make X-risk reduction look vastly more appealing to longtermists. What’s more, the independence assumption in step 2 seems dubious too.

After accounting for the difference in the set of factors involved, and the (likely positive) correlation between factors, we should expect to get a distribution skewer than log-normal distribution. Therefore, we can conclude that the TOC impact distribution is likely more heavy-tailed than log-normal distribution. (Note that this does not imply the TOC impact distribution is closer to Pareto than to log-normal)

Next, let’s examine how heavy-tailed the two distributions exactly are.

For Pareto distribution:

For log-normal distribution:

I find that the implication of Pareto distribution (constant ratio between project  and project ) fits my impression about EA’s search for global priority. On the other hand, the implication of log-normal distribution (rapidly increasing ratio with increased ) is rather unintuitive.

The “constant ratio” implication of Pareto distribution also fits well with the exponential progress of technology[10], and searching for high-efficiency designs in the design space is arguably somewhat analogous to searching for high-impact projects in the altruistic project space. One may worry that the altruistic project space is much more complex than the design space of, say, integrated circuits. However, more relevant than the progress of any single piece of technology, may be the continued spring of new ideas and inventions in all sectors of society, which sustains the exponential growth of the economy. If we see society as a big piece of “technology”, then its design space seems at least as complex as the altruistic landscape, and yet we managed to improve this “technology” at an exponential rate. A notable counterargument to this line of reasoning might be Holden Karnofsky’s post, which claims that “the effective altruism community’s top causes and ‘crucial considerations’ seem to have (exclusively?) been identified very early”, and “there are some reasons to think that future ‘revolutionary crucial considerations’ will be much harder to find, if they exist.”

Carl Shulman, in his comments to this GiveWell blog post, also argued that log-normal distribution is too thin-tailed, which echoes my point here. Note that his argument applies only to TOC impact and not to total impact.

Conclusion: I consider Pareto distribution to better reflect the degree of heavy-tailedness that the TOC impact distribution displays. Note that I don’t consider the arguments above to be very strong, so I have rather low confidence in the conclusion. I do think the distribution is more likely to be (close to) Pareto than to be (close to) log-normal, but the space of all possible heavy-tailed distributions is so vast, that it’s hard to have too much confidence on one single family of distributions.

To sum up,

Claim 2: The distribution of TOC impact across different projects resembles the Pareto distribution in terms of heavy-tailedness.

I expect the distribution of total impact to be lighter-tailed than that of TOC impact, because:

  1. It introduces additive terms as opposed to only multiplicative terms.
  2. “Astronomical stakes”-style arguments (e.g. the one supporting X-risk reduction) applies to most, if not all, interventions, as long as we take into account non-TOC impact.
  1. Other arguments by Brian Tomasik (which apply to total impact and not to TOC impact).

Overall I think the distribution of total impact is more likely to be close to / lower than log-normal, than to be close to Pareto, in terms of heavy-tailedness. Note that I haven’t spent very much time thinking about this. Also I’m unsure whether log-normal distributions or other distributions with lighter tails are more likely.

1.3 Lots of Effort Should Go Into Exploration

Assuming that TOC impact follows a Pareto distribution, let’s first determine the shape parameter . While finding the exact value of  is hard, we may instead try to find an upper bound.

According to this report by the Center for Global Development, the cost-effectiveness of global health interventions displays a phenomena where “if we funded all of these interventions equally, 80 percent of the benefits would be produced by the top 20 percent of the interventions”. This implies a shape parameter of  .

After taking into account longtermist interventions, I expect the TOC impact distribution to be significantly more skew than only considering global health, because longtermist TOCs explicitly take long-term effects into account, where there’s more uncertainty (thus higher variance) involved compared to the near term; and because longtermist interventions often focus more on research, politics, etc. which are apparently more heavy-tailed than distributing bednets. This leads to a significantly smaller . But nevertheless,  can still serve as an upper bound.

Next, let’s try to derive the optimal amount of effort that should go into the search for new projects.

I model the exploration-exploitation process as follows:

  1. We have 1 unit of time in total. In the first  unit of time (), we repeatedly draw random samples from the TOC impact distribution.
  1. In the remaining  unit of time, we work on the one project drawn in step 1 that has the highest impact. Our eventual impact is the product of time spent and project quality.

The expected final impact is approximately  units, where  denotes (approximately) the quality of the project we find, and  denotes the time we spend working on it. Solving for the maximum, the best  turns out to be , which equals  when  . In other words, we should spend 46.3% of all effort on the exploration phase (step 1).

Then let’s try to account for diminishing returns. I model the process as follows:

  1. We have 1 unit of time in total. In the first  unit of time (), we repeatedly draw random samples from the TOC impact distribution.
  2. In the remaining  unit of time, we work on the top  projects drawn in step 1, optimally choosing  and allocating the time in order to maximize total utility. Here I use an isoelastic utility function with , shifted to the left by 1 unit to keep it non-negative.
  1. The isoelastic utility function is adopted to model diminishing returns. It has the property that the “degree of bending” at every point on its curve stays constant, which fits intuitively with how diminishing returns should work. This “degree of bending” is specified by the parameter ; the higher  is, the stronger diminishing returns are. Below are plots of the isoelastic utility function with  and  respectively, both shifted to the left by 1 unit.

A numerical optimization (with some approximating) suggests that  (assuming ), meaning we should spend 65.6% of all effort on the exploration phase.

We can also choose a smaller  to have weaker diminishing returns ( means no diminishing returns), or choose a larger  to have stronger diminishing returns. Below is a plot of  over , assuming  .  

Recall that  is only an upper bound of , but decreasing  further doesn’t turn out to significantly change , so we can take the 65.6% result as an approximate best guess.

Finally: why are we using TOC impact when doing the modeling, rather than total impact?

To sum up,

Claim 3: We, as a community and as individuals, should spend more than half of our effort on searching for potential new projects, rather than working on known projects.

In contrast:

Overall, the EA community as a whole currently seems to allocate approximately 9~12% of its resources to cause prioritization.

Note that donation and grantmaking work (e.g. pouring money into GiveDirectly) is often more scalable than research, community building, management, and other types of work, and this may be part of the reason why the portion of resources spent on cause prioritization is so low. However, the 12% figure and the 10% figure only counted people and not money, so the low scalability of human resources (compared to money) shouldn’t be a big problem there.

Taking one step further, one may argue that cause prioritization research is especially unscalable, and much less scalable than cause-specific research, community building, etc. I don’t have much to say about whether this claim is true, and whether it is enough to justify the 9~12% figure. I’d love to receive input on this.


2 Prioritization: Comparing Known Projects

Unless otherwise specified, all “impact” in section 2 refer to TOC impact rather than total impact.

2.1 Potential TOC Impact of Any Particular Project Follows a Power Law

From now on let’s turn our attention to evaluating known projects. We no longer care about the impact distribution of all potential projects; Instead we focus on how the impact of any particular project is distributed across all potential scenarios.

More rigorously speaking, the distribution can be defined as follows:

Here I argue that such a distribution is roughly as heavy-tailed as the Pareto distribution (aka power law), if not more. I present two arguments here:

  1. A Pareto distribution predicts that, conditional on our project having an impact of at least  units, there’s a constant probability that the project reaches an impact of  (in contrast to the probability decreasing rapidly with increased ), which fits my intuition.
  2. For long-termist projects, the expected impact is usually dominated by a tiny slice of possibility. In fact, the expected impact seems to be infinite. This indicates that the tail is at least as heavy as Pareto.
  1. One may (very reasonably) object to Pascal’s wager and thus modify the expected value decision framework to disallow wagers. Often, (variants of) argument 2 still holds under the modified framework.

Argument 1:

Argument 1 is isomorphic to the argument in section 1.2 about tail distributions. One may wonder whether there’re some connections that we can draw between the two arguments, and between the two distributions (the impact distribution across projects, and the impact distribution across scenarios).

I think the two distributions (and therefore the two arguments) are orthogonal, in the sense that any conceivable impact-distribution-across-projects can, in principle, be combined with any conceivable impact-distribution-across-scenarios. As a result, the two arguments ought to be evaluated separately. More details on this in the appendix.

Argument 2:

Argument 2a: Do anti-wager modifications undermine argument 2?

As we have seen, among all the possible modifications above, some break Pareto-ness (call these “type I” modifications), some preserve Pareto-ness but disallow infinite mean (type II), and some preserve both Pareto-ness and infinite mean (type III). Alternatively, we can choose not to make any modification and use the plain expected value framework, which results in a Pareto distribution with infinite mean (type IV).

In the rest of section 2, we will accommodate type II,III and IV. More specifically:

To sum up,

Claim 4: For any particular project, the distribution of its TOC impact across all potential scenarios is (roughly) at least as heavy-tailed as the Pareto distribution.

I expect the distribution of total impact to be lighter-tailed than that of TOC impact, because:

  1. It introduces additive terms, as opposed to only multiplicative terms.
  2. It seems prima facie plausible that “how long a project’s impact lasts, absent lock-in events” follows an exponential distribution (which means the dilution rate is constant in each year), which is lighter-tailed than Pareto (and even log-normal) distribution.

However, I don’t have much to say about exactly how heavy-tailed it is.

Sidenote on the seriousness of the problem posed by wagers:

2.2 Prioritization Works Like Bilocation

An implication of the impact distribution (of any particular project) being heavy-tailed, is that a large part of its impact (say, 90%) comes from the tail (occupying, say, 10% probability).

Now suppose that there’re two projects, A and B, in front of us, and we’ll work on one of them. The impact distributions of the two projects are identical but independent. We can directly start working on A (or B if you want), but alternatively we can carry out the following procedure:

The good thing about this procedure is that:

We will work only on A or B, but capture most of the expected impact from both A and B. By prioritizing between A and B, we manage to achieve something similar to bilocation.

This is a feature of heavy-tailed distributions. Had the impact followed a normal distribution, there wouldn’t be much to gain from prioritizing between equally good projects; we can at best expect to get one or two standard deviations of additional impact.

Note that this feature relies on the independence assumption. Because A and B are independent, their tails are largely disjoint, and so we can expect to capture both of their tails by prioritizing.

Now let’s quantify this. Suppose that the impact distributions of both projects follow a Pareto distribution (with parameters  and , the latter of which doesn’t matter much to us). Then , the extra gain from prioritizing (compared to working only on an arbitrary project), equals  . Below is a plot of  over .

 is undefined for , because the expected impact is infinite in this case. Note that infinite expected value doesn’t necessarily lead to being Pascal-mugged, as shown by the last three approaches in argument 2a of section 2.1 .[13]

For long-termist projects, I would guess that  holds (that is, the impact distribution has infinite mean), the reason for which is presented in argument 2 of section 2.1 . To avoid dealing with infinities, I shall assume that for long-termist projects,  is larger than but sufficiently close to 1, as a nearest approximation. This gives , meaning we can double our impact by prioritizing.

The choice of  is heavily influenced by subjective factors like one’s policy towards wagers, so very reasonable people may still disagree with me on whether . However, even conservatively choosing  (meaning that 80% impact comes from 20% scenarios) gives  (meaning we can gain 63% more impact from prioritizing), so the claim  isn’t essential here.

In the above model, we only evaluated A instead of both A and B. If we evaluate both A and B, and work on the one with higher impact, then we should expect to have an approximately  boost in impact, which is 100% for  and 82% for  .

Below is a plot of  over .

If prioritization increases our cost-effectiveness by , then we should be willing to do such prioritization as long as it takes less than  of our time, assuming the gain from direct work is linear in the amount of time spent. For  this is 50%, for  this is 39%, and for  this is 45%.

To sum up,

Claim 5: When prioritizing between two equally good longtermist projects that aren’t strongly correlated, the prioritization is worth doing as long as it takes  of total effort to find out the actual impact of at least one project, or to compare which one of the two projects will eventually be more impactful.

One important thing to note is the unrealisticness of “finding out the actual impact of one project” or “comparing which project will eventually be more impactful”, as doing so implies reducing uncertainty to zero, which is clearly impossible. In reality what we can achieve through prioritization will be more modest, and therefore the portion of effort we should spend on prioritization will likewise be more modest than the  indicated here. This limitation also applies to the findings in section 2.3 .

For short-termist projects I’m unclear about what  is appropriate, and therefore prefer to leave them out. Unless otherwise specified, I’ll be focusing on longtermist projects in the remaining parts of section 2, though the conclusions will likely apply to some extent to shortermist projects.

Before we proceed, there’s still one question to ask: why are we using TOC impact (which justifies the use of a Pareto impact distribution), rather than total impact, when modeling prioritization?

  1. Because prioritization is based on evaluation of projects, and TOC impact is much easier to evaluate than non-TOC impact.
  1. Because, as a matter of fact, we often evaluate only TOC impact (or a few most salient pathways of impact) when we evaluate projects. This might be suboptimal (or not), but currently this is how most things are done.
  1. For reasons ii.~v. in section 1.1, which (as previously mentioned) I have mixed attitudes on.
  2. For simplicity.

2.3 A Multi-Project Model of Prioritization

The reasoning in the two-project case naturally generalizes to the multi-project case. Instead of two projects, we now have  projects, the impacts of which are independent and identically distributed.

Instead of simply picking one arbitrary project and working on it, we can identify the project whose actual impact is largest, and devote all the remaining effort to it. The expected impact of the best project is approximately  times that of an arbitrary project, so here the improvement ratio  equals .[14] 

For longtermist projects I will assume  (that is, the impact distribution has infinite mean). If you disagree with this and adopt an  that’s slightly higher than 1, the conclusions in this section should still (approximately) hold because they’re already approximate results, but the approximation error will likely be much larger.

Assuming , we get  (which means we can capture the expected impact of all  projects just by working on the best one!), and so the acceptable ratio of effort spent on prioritization is  if  isn’t very small. This means we should be willing to spend as little as  of our effort on direct work, as long as the remaining effort spent on prioritization successfully leads us to the ex-post best project. Also, we can assume that the  unit of effort spent on prioritization are evenly allocated to each of the  projects, which means  unit for each.

To state the finding in another way,

Claim 6a: When faced with  ( isn’t too small) equally-good longtermist projects that aren’t correlated, we should act as if the  tasks of evaluating every project are each as important as working on the best project that we finally identify, and allocate our effort evenly across the  tasks, as long as we are able to identify the ex-post best project by working on the  evaluation tasks.

 

Longtermist projects are strongly and positively correlated, as most of them rely on the assumption that the future will be large and/or positive (there may also be other, more subtle ways of correlation). This seriously violates the independence assumption that we previously made. Thus a natural next step is to relax the independence assumption.

It’s not an easy job to determine the dependence structure among different projects, and I don’t have a very satisfying way of doing that. Here I present two approaches: a white-box approach and a black-box approach; the results of which are only meant as ballpark estimates and are to be compared with each other.

The white-box approach: clarify where the correlations come from, and model the correlation structure accordingly.

The black-box approach: use a model that doesn’t necessarily correspond to the underlying mechanism, but that displays traits we expect to see in this problem, and then fill in best-guess estimates of parameters.

The two results that we get here,  and , imply a much smaller boost from prioritization than in the independent case.

Recall that “we should be willing to do such prioritization as long as it takes less than  of our time”, so the fraction of time to spend on direct work should be , which equals  or  in the two cases here.

To sum up,

Claim 6b: When the  projects are strongly correlated, prioritization becomes much less important than in the no-correlation case. However, one should still be willing to spend only a small portion (something like  or ; note that this is already much higher than the  in Claim 6a) of one’s effort on direct work, and to spend the remaining portion on prioritization.


3 Examples and Applications

In this final part of the document, I try to put theory into the real world.

Section 3.1 explains what kind of real-world cases can be meaningfully compared with our model (“desiderata”), and what factors influence the conclusion about the importance of E&P (“factors”).

Section 3.2 examines a few real-world cases - not necessarily related to EA - in order to verify the claims in section 3.1 .

Section 3.3 tries to apply the insights that we gained, in an EA context.

3.1 Desiderata to Check For, and Factors to Watch

In the real world, it’s often hard to tell apart exploration and prioritization, so in section 3 I don’t distinguish between exploration and prioritization, and will use E&P to refer to them collectively. That is, we will only distinguish between two types of work: E&P and exploitation.

In order to draw a line between those cases that can be meaningfully compared with our model and those that can’t, here I give two desiderata of the real-world cases that we will examine:

  1. The goal can be roughly described as maximizing some meaningful kind of (expected) utility.
  1. There is a fixed total budget (monetary or not) for E&P + exploitation.

Different examples comply (or don’t comply) with our previous claims (that lots of effort should be spent on E&P) to different degrees. I think the differences are mainly due to three factors:

  1. Difficulty of reducing uncertainty by E&P: In some cases you can reduce uncertainty to zero if you work hard enough on E&P, but in some other cases E&P can’t help you much.
  2. Comparative scalability of E&P vs exploitation: E&P can be scalable (i.e. weak or no diminishing returns) or not, and exploitation can be scalable or not. Holding other factors constant, the more scalable you are, the more resources you deserve.
  1. Heavy-tailedness of opportunities: There’s no point doing E&P, if all opportunities are equally valuable. The more heavy-tailed the distribution is, the more value E&P will bring.

To sum up,

Claim 7: In real world, the most important factors deciding the applicability of our model are difficulty of reducing uncertainty by E&P, comparative scalability of E&P vs exploitation, heavy-tailedness of opportunities, fixed total budget, and utility-maximization objective.

We will test this claim in section 3.2 .

3.2 Real-World Examples

In this section we look at real-world examples, in order to test our theory. All examples satisfy the two desiderata in section 3.1 . Each example will be presented in the following format:

Name of the example (What E&P and exploitation respectively means in this example)

(color of the name will represent the resource share of E&P in this example: highest, lowest)

For the three factors listed, red-ish colors roughly mean “positively contributing (to the importance of E&P)”, while blue-ish colors mean “negatively contributing”. Every colored segment will be followed by a number from 0-5 representing the color number: highest (5), lowest (0). These numbers can also be seen as subjective scores assigned to each entry.

Caveat: I have very limited domain knowledge for many examples below, so it’s very possible that I make false claims about specific examples (and please correct me when I make them!).

Example(s) where E&P receive >50% resources

Catching an escaped criminal (5) (E&P: searching for the hiding place; exploitation: arresting)

Building small deep learning models (5) (E&P: experimenting to find the best design;

                              exploitation: the final training run)

Example(s) where E&P receive 10%-50% resources

Building large language models (2) (E&P: experimenting to find the best design;

                                                           exploitation: the final training run)

Example(s) where E&P receive <10% resources

Hiring a lawyer (1) (E&P: finding the lawyer;

        exploitation: pay the attorney fees and go through the litigation process)

Grantmaking of Open Philanthropy (0) (E&P: evaluating applications;

    exploitation: sending the grant money)

Venture Capital (0) (E&P: evaluating startups; exploitation: investing in chosen startups)

For these examples, it’s fair to say that the three factors predict E&P’s resource share quite well. In particular, the total score for the three factors is perfectly monotone with respect to the score for E&P’s resource share. However, this result should be taken with a grain of salt, since:

3.3 Applications in the EA Community

In this section we apply the framework to examples in the EA community. We will continue to use the format used in section 3.2:

Name of the example (What E&P and exploitation respectively means in this example)

And here is what I get. The subjective scores are my own (very crude) impression; feel free to fill in your estimates.

(Entries are sorted in decreasing order of total score.)

Cultivating talent (E&P: identifying promising individuals[24];

     exploitation: supporting selected individuals[25] in their development)

Cause prioritization / Cause-specific work (E&P: cause prioritization;

         exploitation: cause-specific work)

Within-cause prioritization / Direct work (E&P: within-cause prioritization;

     exploitation: direct work)

Career choice (E&P: choosing a career path; exploitation: progressing on the chosen path)

Employee recruitment (E&P: trying to identify the best applicant;

  exploitation: hiring the chosen applicant and pay them salary)

Grantmaking (E&P: evaluating applications; exploitation: sending the grant money)

To sum up,

Claim 8: In EA, the three areas where E&P deserves the largest portions of resources (relative to the total resources allocated to that area) are


Appendix: Mathematical Details

Appendix section Ax.y contains the omitted mathematical details in section x.y .

A1.2

Heavy-tailedness of Pareto and its implication: (we’ll follow the notations here)

Heavy-tailedness of log-normal and its implication:

A1.3

Proposition 1: If we independently draw  samples from a Pareto distribution with parameters  and , the expected maximum can be approximated by .

Proposition 2: If we independently draw  samples from a Pareto distribution with parameters  and , the expected -th largest value  can be approximated by .

Proposition 3: If we independently draw  samples from a Pareto distribution with parameters  and , the expected maximum is precisely , where  is the generalized binomial coefficient.

Proposition 4: If we independently draw  samples from a Pareto distribution with parameters  and , the expected -th largest value  is precisely , which equals .

Proof of proposition 1-4:

The exploration-exploitation process without diminishing returns:

The exploration-exploitation process with diminishing returns:

A2.1

For the mathematical details behind argument 1, see A1.1.2 .

Are the impact-distribution-across-projects (abbr. IDAP) and the impact-distribution-across-scenarios (abbr. IDAS) orthogonal?

Do anti-wager modifications undermine argument 2?

A2.2

The expression for the improvement ratio , assuming we had only evaluated project A:

The expression for the improvement ratio , assuming we had evaluated both projects:

Why is it that “we should be willing to do such prioritization as long as it takes less than  of our time”?

A2.3

First, the independent case.

The expression for the improvement ratio , assuming we had evaluated all  projects:

Then, the inter-dependent case.

The white-box approach:

The black-box approach:

Lemma 1: , where  is any real number.

Proof:


[1] In the current situation, the EA community as a whole seems to allocate approximately 9~12% of its resources to cause prioritization. But there’s some nuance to this - see the last parts of section 1.3 .

[2] Meaning I assign 0.45 probability to [the statement being true (in the real world)], and 0.2 probability to [the statement being true (in the real world) and my model being mostly right about the reason].

[3] Which, by the way, is quite an unrealistic assumption. This assumption is also shared by claim 6b.

[4] By taking this factor into account, we’ve dealt with the unrealistic assumption (“we are able to identify the ex-post best project”) in claim 6a and 6b.

[5] Including, for example, providing opportunities for individuals to test fit.

[6] Conditional on the heavy-tailedness comparison here being meaningful, which isn’t obvious. Same for similar comparisons elsewhere in this document.

[7] Subject to caveats about Pascal’s wager. See section 2.1 .

[8] A more realistic model might be a hierarchical one, where projects have sub-projects and sub-sub-projects, etc., and you need to do some amount of prioritization at every level of the hierarchy.

[9] Meaning I assign 0.8 probability to [the statement being true], and 0.8 probability to [the statement being true and my model being mostly right about the reason].

[10] For any technology X, assume that a constant amount of resources are spent each year on developing X. If we observe an exponential increase in X’s efficiency, we can infer that it always takes a constant amount of resources to double X’s efficiency, regardless of its current efficiency - which points to a Pareto distribution. This “amount of work needed to double the efficiency” may have been slowly increasing in the case of integrated circuits, but far slower than what a log-normal distribution would predict.

[11] Here the diminishing returns mean “saving 108 lives is less than 105 times as good as saving 103 lives, not because we’re scope-insensitive but because we’re risk averse and 108 is usually much more speculative and thus riskier than 103.”

[12] Isoelastic utility functions are good representatives of the broader class of HARA utility functions, which, according to Wikipedia, is “the most general class of utility functions that are usually used in practice”.

[13] Under the “fundamental assumptions” or “sense check with intuition” approach, the true distribution has finite mean, but the Pareto distribution used for approximating the true distribution has infinite mean. Under the “heuristics” approach, the true distribution itself has infinite mean.

[14] u is defined in section 2.2; it stands for the extra gain in “how good is the project that we work on” resulting from prioritization, compared to working only on an arbitrary project. For example, if prioritization increases the project quality from 1 DALY/$ to 2 DALY/$, then u=100%=1.0 .

[15] See the “⍺” column of the “Revenue” rows in table 1 of the paper.

[16] Copulas are used for modeling the dependence structure between multiple random variables. A reversed Clayton copula is a copula that shows stronger correlation when the variables take larger values. Mathematical knowledge about the (reversed) Clayton copula (and about copulas in general) isn’t needed for reading this section.

[17] This is a very crude guess, and my 90% confidence interval will likely be very, very wide.

[18] Note that by choosing r=⅓ I’m underestimating (to a rather small extent) the strength of correlation. I’ll briefly revisit this in a later footnote.

[19] Recall that we underestimated the strength of correlation by choosing r=⅓, so here u=(log k)-1 is an overestimation of the boost from prioritization, though I think the extent of overestimation is rather small.

[20] reversed because it’s negatively correlated with importance of E&P

[21] For GPT-3, 12% of compute is spent on training smaller models than the final 175B-parameter one, according to table D.1 of the GPT-3 paper, though it’s unclear whether that 12% is used for exploration/comparison, or simply checking for potential problems. Google’s T5 adopted a similar approach of experimenting on smaller models, and they made it clear that those experiments were used to explore and compare model designs, including network architectures. Based on the data in the paper I estimate that 10%-30% of total compute is spent on those experiments, with high uncertainty.

[22] I’m not counting referral fees into E&P, since they’re usually charged on the lawyer’s side while I’m mainly examining the client’s willingness to pay. Plus, it’s unclear what portion of clients use referral services, and how much referral services help improve the competence of the lawyer that you find.

[23] This is based on a simple ballpark estimate, and so I don’t provide details here.

[24] Including, for example, providing opportunities for individuals to test fit.

[25] The extent to which to prioritize promising individuals, is often discussed under the title of “elitism”. Also here’s some related research.

[26] Including, for example, providing opportunities for individuals to test fit.

[27] I know little about macroeconomics, and this claim is of rather low confidence.