[I mostly wrote this to clarify my thoughts. I’m unclear whether this will be valuable for readers. ]
I expect that within a decade, AI will be able to do 90% of current human jobs. I don’t mean that 90% of humans will be obsolete. I mean that the average worker could delegate 90% of their tasks to an AGI.
I feel confused about what this implies for the kind of AI long-term planning and strategizing that would enable an AI to create large-scale harm if it is poorly aligned.
Is the ability to achieve long-term goals hard for an AI to develop?
Disagreements related to what we value seem to explain maybe 10% of the disagreements over AI safety. This post will try to explain how I think about which values I care about perpetuating to the distant future.
Robin Hanson helped to clarify the choices in Which Of Your Origins Are You?:
The key hard question here is this: what aspects of the causal influences that lead to you do you now embrace, and which do you instead reject as “random” errors that you want to cut out? Consider two extremes.
At one extreme, one could endorse absolutely every random element that contributed to any prior choice or intuition.
At the other extreme, you might see yourself as primarily the result of natural selection, both of genes and of memes, and see your core non-random value as that of doing the best you can to continue to “win” at that game. … In this view, everything about you that won’t help your descendants be selected in the long run is a random error that you want to detect and reject.
In other words, the more unique criteria we have about what we want to preserve into the distant future, the less we should expect to succeed.
I’ve been hearing vague claims that automated theorem provers are able to, or will soon be able to prove things about complex software such as AIs.
Max Tegmark and Steve Omohundro have now published a paper, Provably Safe Systems : The Only Path To Controllable AGI, which convinces me that this is a plausible strategy to help with AI safety.
Book review: The Coming Wave: Technology, Power, and the Twenty-first Century’s Greatest Dilemma, by Mustafa Suleyman.
An author with substantial AI expertise has attempted to discuss AI in terms that the average book reader can understand.
The key message: AI is about to become possibly the most important event in human history.
Maybe 2% of readers will change their minds as a result of reading the book.
A large fraction of readers will come in expecting the book to be mostly hype. They won’t look closely enough to see why Suleyman is excited.
Context: looking for an alternative to a pause on AI development.
There’s some popular desire for software decisions to be explainable when used for decisions such as whether to grant someone a loan. That desire is not sufficient reason for possibly crippling AI progress. But in combination with other concerns about AI, it seems promising.
Much of this popular desire likely comes from people who have been (or expect to be) denied loans, and who want to scapegoat someone or something to avoid admitting that they look unsafe to lend to because they’ve made poor decisions. I normally want to avoid regulations that are supported by such motives.
Yet an explainability requirement shows some promise at reducing the risks from rogue AIs.
Book review: Outlive: The Science and Art of Longevity, by Peter Attia.
This year’s book on aging focuses mostly on healthspan rather than lifespan, in an effort to combat the tendency of people in the developed world to have a wasted decade around age 80.
Attia calls his approach Medicine 3.0. He wants people to pay a lot more attention to their lifestyle starting a couple of decades before problems such as diabetes and Alzheimer’s create obvious impacts.
He complains about Medicine 2.0 (i.e. mainstream medicine) treating disease as a binary phenomenon. There’s lots of evidence suggesting that age-related diseases develop slowly over periods of more than a decade.
He’s not aiming to cure aging. He aims to enjoy life until age 100 or 120.
I’m wondering how selection effects will influence the first serious attempt by an AGI to take over the world.
My question here is inspired by thoughts about people who say AGI couldn’t conquer the world because it will depend on humans to provide electricity, semiconductors, etc.
I participated last summer in Tetlock’s Existential Risk Persuasion Tournament (755(!) page paper here).
Superforecasters and “subject matter experts” engaged in a hybrid between a prediction market and debates, to predict catastrophic and existential risks this century.
Robin Hanson suggests, partly in response to calls for a pause in development of AGI, liability rules for risks related to AGI rapidly becoming powerful.
My intuitive reaction was to classify foom liability as equivalent to a near total ban on AGI.
Now that I’ve found time to think more carefully about it, I want to advocate foom liability as a modest improvement over any likely pause or ban on AGI research. In particular, I want the most ambitious AI labs worldwide to be required to have insurance against something like $10 billion to $100 billion worth of damages.
This is a tale of some differences between functional medicine and mainstream medicine in dealing with high cholesterol. Specifically, how I was treated at the California Center for Functional Medicine (CCFM) versus how I was treated at Kaiser.