effective altruism

All posts tagged effective altruism

Book review: Doing Good Better, by William MacAskill.

This book is a simple introduction to the Effective Altruism movement.

It documents big differences between superficially plausible charities, and points out how this implies big benefits to the recipients of charity from donors paying more attention to the results that a charity produces.

How effective is the book?

Is it persuasive?

Probably yes, for a small but somewhat important fraction of the population who seriously intend to help distant strangers, but have procrastinated about informing themselves about how to do so.

Does it focus on a neglected task?

Not very neglected. It’s mildly different from similar efforts such as GiveWell’s website and Reinventing Philanthropy, in ways that will slightly reduce the effort needed to understand the basics of Effective Altruism.

Will it make people more altruistic?

Not very much. It mostly seems to assume that people have some fixed level of altruism, and focuses on improving the benefits that result from that altruism. Maybe it will modestly redirect peer pressure toward making people more altruistic.

Will it make readers more effective?

Probably. For people who haven’t given much thought to these topics, the book’s advice is a clear improvement over standard habits. It will be modestly effective at promoting a culture where charitable donations that save lives are valued more highly than donations which accomplish less.

But I see some risk that it will make people overconfident about the benefits of the book’s specific strategies. An ideal version of the book would instead inspire people to improve on the book’s analysis.

The book provides evidence that donors rarely pay attention to how much good a charity does. Yet it avoids asking why. If you pay attention, you’ll see hints that donors are motivated mainly by the desire to signal something virtuous about themselves (for example, see the book’s section on moral licensing). In spite of that, the book consistently talks as if donors have good intentions, and only need more knowledge to be better altruists.

The book is less rigorous than I had hoped. I’m unsure how much of that is due to reasonable attempts to simplify the message so that more people can understand it with minimal effort.

In a section on robustness of evidence, the book describes this “sanity check”:

“if it cost ten dollars to save a life, then we’d have to suppose that they or their family members couldn’t save up for a few weeks, or take out a loan, in order to pay for the lifesaving product.”

I find it confusing to use this as a sanity check, because it’s all too easy to imagine that many people are in desperate enough conditions that they’re spending their last dollar to avoid starvation.

The book alternates between advocating doing more good (satisficing), and advocating the most possible good (optimizing). In practice, it mostly focuses on safe ways to produce fairly good results.

The book barely mentions existential risks. If it were literally trying to advocate doing the most good possible, it would devote a lot more attention to affecting the distant future. But that’s much harder to do well than what the book does focus on (saving a few more lives in Africa over the next few years), and would involve acts of charity that have small probabilities of really large effects on people who are not yet born.

If you’re willing to spend 50-100 hours (but not more) learning how to be more effective with your altruism, then reading this book is a good start.

But people who are more ambitious ought to be able to make a bigger difference to the world. I encourage those people to skip this book, and focus more on analyzing existential risks.

This post is partly a response to arguments for only donating to one charity and to an 80,000 Hours post arguing against diminishing returns. But I’ll focus mostly on AGI-risk charities.

Diversifying Donations?

The rule that I should only donate to one charity is a good presumption to start with. Most objections to it are due to motivations that diverge from pure utilitarian altruism. I don’t pretend that altruism is my only motive for donating, so I’m not too concerned that I only do a rough approximation of following that rule.

Still, I want to follow the rule more closely than most people do. So when I direct less than 90% of my donations to tax-deductible nonprofits, I feel a need to point to diminishing returns [1] to donations to justify that.

With AGI risk organizations, I expect the value of diversity to sometimes override the normal presumption even for purely altruistic utilitarians (with caveats about having the time needed to evaluate multiple organizations, and having more than a few thousand dollars to donate; those caveats will exclude many people from this advice, so this post is mainly oriented toward EAs who are earning to give or wealthier people).

Diminishing Returns?

Before explaining that, I’ll reply to the 80,000 Hours post about diminishing returns.

The 80,000 Hours post focuses on charities that mostly market causes to a wide audience. The economies of scale associated with brand recognition and social proof seem more plausible than any economies of scale available to research organizations.

The shortage of existential risk research seems more dangerous than any shortage of charities which are devoted to marketing causes, so I’m focusing on the most important existential risk.

I expect diminishing returns to be common after an organization grows beyond two or three people. One reason is that the founders of most organizations exert more influence than subsequent employees over important policy decisions [2], so at productive organizations founders are more valuable.

For research organizations that need the smartest people, the limited number of such people implies that only small organizations can have a large fraction of employees be highly qualified.

I expect donations to very young organizations to be more valuable than other donations (which implies diminishing returns to size on average):

  • It takes time to produce evidence that the organization is accomplishing something valuable, and donors quite sensibly prefer organizations that have provided such evidence.
  • Even when donors try to compensate for that by evaluating the charity’s mission statement or leader’s competence, it takes some time to adequately communicate those features (e.g. it’s rare for a charity to set up an impressive web site on day one).
  • It’s common for a charity to have suboptimal competence at fundraising until it grows large enough to hire someone with fundraising expertise.
  • Some charities are mainly funded by a few grants in the millions of dollars, and I’ve heard reports that those often take many months between being awarded and reaching the charities’ bank (not to mention delays in awarding the grants). This sometimes means months when a charity has trouble hiring anyone who demands an immediate salary.
  • Donors could in principle overcome these causes of bias, but as far as I can tell, few care about doing so. EA’s come a little closer to doing this than others, but my observations suggest that EA’s are almost as lazy about analyzing new charities as non EA’s.
  • Therefore, I expect young charities to be underfunded.

Why AGI risk research needs diversity

I see more danger of researchers pursuing useless approaches for existential risks in general, and AGI risks in particular (due partly to the inherent lack of feedback), than with other causes.

The most obvious way to reduce that danger is to encourage a wide variety of people and organizations to independently research risk mitigation strategies.

I worry about AGI-risk researchers focusing all their effort on a class of scenarios which rely on a false assumption.

The AI foom debate seems superficially like the main area where a false assumption might cause AGI research to end up mostly wasted. But there are enough influential people on both sides of this issue that I expect research to not ignore one side of that debate for long.

I worry more about assumptions that no prominent people question.

I’ll describe how such an assumption might look in hindsight via an analogy to some leading developers of software intended to accomplish what the web ended up accomplishing [3].

Xanadu stood out as the leading developer of global hypertext software in the 1980s to about the same extent that MIRI stands out as the leading AGI-risk research organization. One reason [4] that Xanadu accomplished little was the assumption that they needed to make money. Part of why that seemed obvious in the 1980s was that there were no ISPs delivering an internet-like platform to ordinary people, and hardware costs were a big obstacle to anyone who wanted to provide that functionality. The hardware costs declined at a predictable enough rate that Drexler was able to predict in Engines of Creation (published in 1986) that ordinary people would get web-like functionality within a decade.

A more disturbing reason for assuming that web functionality needed to make a profit was the ideology surrounding private property. People who opposed private ownership of home, farms, factories, etc. were causing major problems. Most of us automatically treated ownership of software as working the same way as physical property.

People who are too young to remember attitudes toward free / open source software before about 1997 will have some trouble believing how reluctant people were to imagine valuable software being free. [5] Attitudes changed unusually fast due to the demise of communism and the availability of affordable internet access.

A few people (such as RMS) overcame the focus on cold war issues, but were too eccentric to convert many followers. We should pay attention to people with similarly eccentric AGI-risk views.

If I had to guess what faulty assumption AGI-risk researchers are making, I’d say something like faulty guesses about the nature of intelligence or the architecture of feasible AGIs. But the assumptions that look suspicious to me are ones that some moderately prominent people have questioned.

Vague intuitions along these lines have led me to delay some of my potential existential-risk donations in hopes that I’ll discover (or help create?) some newly created existential-risk projects which produce more value per dollar.

Conclusions

How does this affect my current giving pattern?

My favorite charity is CFAR (around 75 or 80% of my donations), which improves the effectiveness of people who might start new AGI-risk organizations or AGI-development organizations. I’ve had varied impressions about whether additional donations to CFAR have had diminishing returns. They seem to have been getting just barely enough money to hire employees they consider important.

FLI is a decent example of a possibly valuable organization that CFAR played some hard-to-quantify role in starting. It bears a superficial resemblance to an optimal incubator for additional AGI-risk research groups. But FLI seems too focused on mainstream researchers to have much hope of finding the eccentric ideas that I’m most concerned about AGI-researchers overlooking.

Ideally I’d be donating to one or two new AGI-risk startups per year. Conditions seem almost right for this. New AGI-risk organizations are being created at a good rate, mostly getting a few large grants that are probably encouraging them to focus on relatively mainstream views [6].

CSER and FLI sort of fit this category briefly last year before getting large grants, and I donated moderate amounts to them. I presume I didn’t give enough to them for diminishing returns to be important, but their windows of unusual need were short enough that I might well have come close to that.

I’m a little surprised that the increasing interest in this area doesn’t seem to be catalyzing the formation of more low-budget groups pursuing more unusual strategies. Please let me know of any that I’m overlooking.

See my favorite charities web page (recently updated) for more thoughts about specific charities.

[1] – Diminishing returns are the main way that donating to multiple charities at one time can be reconciled with utilitarian altruism.

[2] – I don’t know whether it ought to work this way, but I expect this pattern to continue.

[3] – they intended to accomplish a much more ambitious set of goals.

[4] – probably not the main reason.

[5] – presumably the people who were sympathetic to communism weren’t attracted to small software projects (too busy with politics?) or rejected working on software due to the expectation that it required working for evil capitalists.

[6] – The short-term effects are probably good, increasing the diversity of approaches compared to what would be the case if MIRI were the only AGI-risk organization, and reducing the risk that AGI researchers would become polarized into tribes that disagree about whether AGI is dangerous. But a field dominated by a few funders tends to focus on fewer ideas than one with many funders.

Book review: Poor Economics: A Radical Rethinking of the Way to Fight Global Poverty by Abhijit V. Banerjee and Esther Duflo.

This book gives an interesting perspective on the obstacles to fixing poverty in the developing world. They criticize both Jeffrey Sach and William Easterly for overstating how easy/hard it is provide useful aid to the poor by attempting simple and sweeping generalizations, where Banerjee and Duflo want us to look carefully at evidence from mostly small-scale interventions which sometimes produce decent results.

They describe a few randomized controlled trials, but apparently there aren’t enough of those to occupy a full book, so they spend more time on less rigorous evidence of counter-intuitive ways that aid programs can fail.

They portray the poor as mostly rational and rarely making choices that are clearly stupid given the information that is readily available to them. But their cognitive abilities are sometimes suboptimal due to mediocre nutrition, disease, and/or stress from financial risks. Relieving any of those problems can sometimes enable them to become more productive workers.

The book advocates mild paternalism in the form of nudging weakly held beliefs about health-related questions where people can’t easily observe the results (e.g. vaccination, iodine supplementation), but probably not birth control (the poor generally choose how many children to have, although there are complex issues influencing those choices). They point out that the main reason people in developed countries make better health choices is due to better defaults, not more intelligence. I wish they’d gone a bit farther and speculated about how many of our current health practices will look pointlessly harmful to more advanced societies.

They give a lukewarm endorsement of microcredit, showing that it needs to be inflexible to avoid high default rates, and only provides small benefits overall. Most of the poor would be better off with a salaried job than borrowing money to run a shaky business.

The book fits in well with Givewell’s approach.

Book review: Reinventing Philanthropy: A Framework for More Effective Giving, by Eric Friedman.

This book will spread the ideas behind effective altruism to a modestly wider set of donors than other efforts I’m aware of. It understates how much the effective altruism movement differs from traditional charity and how hard it is to implement, but given the shortage of books on this subject any addition is valuable. It focuses on how to ask good questions about philanthropy rather than attempting to find good answers.

The author provides a list of objections he’s heard to maximizing the effectiveness of charity, a majority of which seem to boil down to the “diversification of nonprofit goals would be drastically reduced”, leading to many existing benefits being canceled. He tries to argue that people have extremely diverse goals which would result in an extremely diverse set of charities. He later argues that the subjectivity of determining the effectiveness of charities will maintain that diversity. Neither of these arguments seem remotely plausible. When individuals explicitly compare how they value their own pleasure, life expectancy, dignity, freedom, etc., I don’t see more than a handful of different goals. How could it be much different for recipients of charity? There exist charities whose value can’t easily be compared to GiveWell’s recommended ones (stopping nuclear war?), but they seem to get a small fraction of the money that goes to charities that GiveWell has decent reasons for rejecting.

So I conclude that widespread adoption of effective giving would drastically reduce the diversity of charitable goals (limited mostly by the fact that spending large amount on a single goal is subject to diminishing returns). The only plausible explanation I see for peoples’ discomfort with that is that people are attached to beliefs which are inconsistent with treating all potential recipients as equally deserving.

He’s reluctant to criticize “well-intentioned” donors who use traditional emotional reasoning. I prefer to think of them as normally-intentioned (i.e. acting on a mix of selfish and altruistic motives).

I still have some concerns that asking average donors to objectively maximize the impact of their donations would backfire by reducing the emotional benefit they get from giving more than it increases the effectiveness of their giving. But since I don’t expect more than a few percent of the population to be analytical enough to accept the arguments in this book, this doesn’t seem like an important concern.

He tries to argue that effective giving can increase the emotional benefit we get from giving. This mostly seems to depend on getting more warm fuzzy feelings from helping more people. But as far as I can tell, those feelings are very insensitive to the number of people helped. I haven’t noticed any improved feelings as I alter my giving to increase its impact, and the literature on scope insensitivity suggests that’s typical.

He wants donors to treat potentially deserving recipients as equally deserving regardless of how far away they are, but he fails to include people who are distant in time. He might have good reasons for not wanting to donate to people of the distant future, but not analyzing those reasons risks making the same kind of mistake he criticizes donors for making about distant continents.

Book Review: Let Their People Come: Breaking the Gridlock on Global Labor Mobility by Lant Pritchett.
This book is primarily written for economists and academics in related fields, but most of it can be understood by an average person.
I was a little hesitant to read this book because I suspected it would do little more than reinforce my existing beliefs. There were certainly parts of the book that I would have been better off skipping for that reason.
But one important effect of the book was to convince me that the effects on the poor of migration to wealthier countries is so large compared to things like “foreign aid” and free trade that anyone trying to help the poor by influencing government policies shouldn’t spend any time thinking about how to improve “foreign aid” or trade barriers.
I’ve long been wondering how to respond to remarks such as Jimmy Carter’s ‘We are the stingiest nation of all’ based the U.S.’s low “foreign aid” to GDP ratio. Pointing out that “foreign aid” is mostly wasted or even harmful requires too much analysis of lots of not-too-strong evidence. Pritchett shows that the wealth affects of allowing the poor to work in rich countries should dominate any measure of how those rich countries treat the poor. By that measure, adjusting for country size, the U.S. ranks better than countries in the EU, but is embarrassingly callous compared to the United Arab Emirates, Kuwait, and Jordan.
The book addresses both moral and selfish arguments for restricting immigration. It treats the selfish arguments (even those based on myths) as problems that can’t be overcome, but which can be reduced via compromises. These pragmatic parts of the book are too ordinary to be worth much.
The sections about moral arguments are more powerful. He clearly demonstrates a large blind spot in the moral vision of those who think they’re opposed to all discrimination but who aren’t offended by discrimination on the basis of the nationality a person was assigned at birth. But he exaggerates when he claims that nationality is the only exception to a widely agreed on outrage at discrimination based on “condition of birth”. Discrimination based on date of birth still gets wide support (e.g. the drinking age). And if you’re born as a conjoined twin, don’t expect much protection from surgery that looks about as moral as brain surgery designed to cure a child’s homosexuality should.
Perhaps this book is one small step toward creating a movement with a slogan such as “Tear down that kinder, gentler Berlin wall!”.

This post is a response to a challenge on Overcoming Bias to spend $10 trillion sensibly.
Here’s my proposed allocation (spending to be spread out over 10-20 years):

  • $5 trillion on drug patent buyouts and prizes for new drugs put in the public domain, with the prizes mostly allocated in proportion to the quality adjusted life years attributable to the drug.
  • $1 trillion on establishing a few dozen separate clusters of seasteads and on facilitating migration of people from poor/oppressive countries by rewarding jurisdictions in proportion to the number of immigrants they accept from poorer / less free regions. (I’m guessing that most of those rewards will go to seasteads, many of which will be created by other people partly in hopes of getting some of these rewards).

    This would also have a side affect of significantly reducing the harm that humans might experience due to global warming or an ice age, since ocean climates have less extreme temperatures, seasteads will probably not depend on rainfall to grow food, and can move somewhat to locations with better temperatures.
  • $1 trillion on improving political systems, mostly through prizes that bear some resemblance to The Mo Ibrahim Prize for Achievement in African Leadership (but not limited to democratically elected leaders and not limited to Africa). If the top 100 or so politicians in about 100 countries are eligible, I could set the average reward at about $100 million per person. Of course, nowhere near all of them will qualify, so a fair amount will be left over for those not yet in office.
  • $0.5 trillion on subsidizing trading on prediction markets that are designed to enable futarchy. This level of subsidy is far enough from anything that has been tried that there’s no way to guess whether this is a wasteful level.
  • $1 trillion existential risks
    Some unknown fraction of this would go to persuading people not to work on AGI without providing arguments that they will produce a safe goal system for any AI they create. Once I’m satisfied that the risks associated with AI are under control, much of the remaining money will go toward establishing societies in the asteroid belt and then outside the solar system.
  • $0.5 trillion on communications / computing hardware for everyone who can’t currently afford that.
  • $1 trillion I’d save for ideas I think of later.

I’m not counting a bunch of other projects that would use up less than $100 billion since they’re small enough to fit in the rounding errors of the ones I’ve counted (the Methuselah Mouse prize, desalinization and other water purification technologies, developing nanotech, preparing for the risks of nanotech, uploading, cryonics, nature preserves, etc).

Book review: Poverty and Discrimination by Kevin Lang.
This book is designed to make you feel less sure of your knowledge, and it succeeds in that goal. That’s a worthy accomplishment, although it provides much less satisfaction than a book that provides a grand vision for solving problems would. At some abstract intellectual level I liked the book, but my gut feelings often told me that reading the book was unrewarding work that I shouldn’t do unless it was assigned reading for a course I needed.
The book will dissatisfy anyone who wants to view politics as a fight between good and evil. For many issues such as the minimum wage, he provides strong arguments that the effects are small enough that we should doubt whether the issue is worth fighting about.
He gives good explanations of why it’s hard to even have clear concepts of poverty and discrimination by providing examples of how seemingly trivial or unobservable differences can create results that our intuitions say are important to our moral rules.
He provides clear evidence that some discrimination still exists, and then thoroughly explains why there’s large uncertainty about how harmful it is. He presents one moderately unrealistic model in which discrimination is common but doesn’t affect wages. Then he presents a somewhat more realistic model in which a tiny bit of discrimination produces large wage differences. But those wage differences may overstate the harm done, because they’re partly due to minorities spending less on education and to women pursuing careers in lower risk occupations or careers which allow more flexibility to take time off.
There are only a handful of places where I doubted his objectivity.
He reports one study showing evidence of racial discrimination in home loans, but fails to mention any of the contrary evidence such as the Anderson and Vanderhoff paper showing higher marginal default rates for blacks.
The final few pages on policy implications seem poorly thought out compared to the rest of the book (he says that’s the least important chapter of the book). He claims that income taxes on the bottom quintile can be reduced to zero by a 10% increase on the top quintile, but that claim depends on assumptions about how reported income changes in response to tax increases. He doesn’t indicate what assumptions his claim depends on.
He claims “The high rate of incarceration in the United States and the high level of inequality are related.” He gives a plausible theory about why inequality causes the wealthy in some countries to spend a lot protecting their wealth from the poor, but provides no evidence connecting that theory to U.S. incarceration rates.

Bernie Sanders has introduced a bill to replace patent monopoly protection for drugs with awards based in part on Quality Adjusted Life Years added by the drugs.
This would eliminate the harm due to monopoly pricing. It might also cause some research to be redirected from “me-too” drugs to more innovative drugs. But I suspect that it’s common enough for what initially looks like a “me-too” drug to end up having valuable advantages that such an effect will be minor.
It would probably be a bigger help to people in developing nations than all the government spending misleadingly labeled as foreign aid.
Because politics will ensure that the idea is implemented suboptimally, I would prefer that something similar (e.g. patent buyouts) be implemented by a more responsible institution such as the Gates Foundation. But the patent system has enough problems that even this imperfectly written bill might improve on the status quo.
One strange effect of political reality is that the rewards are apportioned according to either benefits to U.S. patients or world patients, and the bill provides an awfully vague description of which rule will apply to which drug.
The bill allocates 10% of the rewards to orphan drugs, presumably because the lives of people with those diseases are worth more than those with common diseases.
The bill claims generics cost 85% less than patented drugs, but gets that figure from comparing overall generic prices with overall patented prices. If the cost of manufacturing drugs differs for old and new drugs, that will be misleading. The estimates I’ve found for same-drug price declines after generic competition starts suggest the price decline is more like 30% to 50%. So the bill’s claim that it can be financed by the reduced Federal government drug spending appear to be fiction.
Besides, if it were self-financing that way, wouldn’t it indicate a big reduction in the rewards to drug development? I want to see a good analysis of why $80 billion a year is adequate to substitute for patent exclusivity. My crude attempts at analyzing it suggests it’s too low, but not by a large amount.
(HT Alex Tabarrok)

Book review: The Bottom Billion: Why the Poorest Countries are Failing and What Can Be Done About It by Paul Collier
This very eloquent and mostly thoughtful book about the world’s poorest countries will offend ideologues of all stripes. Collier’s four different explanations for poverty traps (war, presence of natural resources, bad neighbors blocking trade routes, and corruption) clearly place him as a fox rather than a hedgehog without being complex enough that they can rationalize any result (although they can probably rationalize more results than an ideal set of explanations would). He blames both villains in poor countries and thoughtless voters in wealthy countries.
Collier sees that globalization has benefited most nations, but provides plausible mechanisms by which globalization can harm some (e.g. through enabling capital flight).
Collier mostly thinks like a good economist, but his prior work for the World Bank biases him to be overly optimistic about improving such institutions. He recognizes the incentives that cause bureaucrats to be too risk averse, but then makes a cryptic claim that the British government understands the problem and is spending money to fix it. He vaguely implies that this is a venture capital-like fund, but fails to say whether they replicated the key venture capital feature of providing unusually large rewards to employees who produce unusually good results. His silence on this subject leads me to suspect that he’s asking us to blindly trust institutions that have a long track record of avoiding results-oriented incentives.
He also shows misplaced faith in authority when he tries to calculate the value to the world of rescuing a failed state by using George Bush’s calculation that the benefits of installing a good government in Iraq exceeded the expected $100 billion cost. That might be a good argument if Bush had been spending his own money to help Iraq, but his willingness to spend other peoples’ money doesn’t say much.
Collier says it is “surely irresponsible” to leave Somalia with no government. Yet most evidence I’ve seen says Somalia improved by most standard criteria such as life expectancy when it had no government. I don’t know how reliable that evidence is, but Collier’s apparent assumption that we don’t need to look at the evidence makes his opinion suspect.
The book’s biggest shortcoming is the absence of anything resembling footnotes. Collier implies this is too make the book more readable, but he could have put a section of notes at the end referencing individual pages without altering the main text in any way. Instead he only gives a fairly large list of papers he’s written. But I can’t tell without tracking down and reading a large fraction of them which of them if any support his controversial claims (e.g. that giving money to the poorest countries helps them a bit but that doubling it would reach a limit beyond which further money would be wasted).
But his advice is good enough that its value doesn’t depend much on those controversial claims being right. Following his advice to condition aid on results (e.g. sending money to countries when they stop wars, cutting it off if they have a coup or resume war) would provide incentives that would make aid beneficial.
I had previously suspected that large countries have tended to escape poverty more easily in the past few decades because “aid” organizations had enough money to prop up small corrupt governments but not enough to affect a government such as India’s. Collier presents a good alternative theory: being a large country pretty much guarantees access to the sea, and by increasing the number of neighbors, increases the chance of having a neighbor which is open to trade.
Another good tidbit is this point on Fair Trade: farmers “get charity as long as they stay producing the crops that have locked them into poverty.”

Nick Bostrom has a good paper on Astronomical Waste: The Opportunity Cost of Delayed Technological Development, which argues that under most reasonable ethical systems that aren’t completely selfish or very parochial, our philanthropic activities ought to be devoted primarily toward preventing disasters that would cause the extinction of intelligent life.
Some people who haven’t thought about the Fermi Paradox carefully may overestimate the probability that most of the universe is already occupied by intelligent life. Very high estimates for that probability would invalidate Bostrom’s conclusion, but I haven’t found any plausible arguments that would justify that high a probability.
I don’t want to completely dismiss Malthusian objections that life in the distant future will be barely worth living, but the risk of a Malthusian future would need to be well above 50 percent to substantially alter the optimal focus of philanthropy, and the strongest Malthusian arguments that I can imagine leave much more uncertainty than that. (If I thought I could alter the probability of a Malthusian future, maybe I should devote effort to that. But I don’t currently know where to start).
Thus the conclusion seems like it ought to be too obvious to need repeating, but it’s far enough from our normal experiences that most of us tend to pay inadequate attention to it. So I’m mentioning it in order to remind people (including myself) of the need to devote more of our time to thinking about risks such as those associated with AI or asteroid impacts.