Bayesian Investor Blog

Ramblings of a somewhat libertarian stock market speculator

AI Motivated by Compassion and Respect

Posted by Peter on December 11, 2007
Posted in: Artificial Intelligence. Tagged: existential risks, pascal's wager.

Tim Freeman has a paper which clarifies many of the issues that need to be solved for humans to coexist with a superhuman AI. It comes close to what we would need if we had unlimited computing power. I will try amplify on some of the criticisms of it from the sl4 mailing list.
It errs on the side of our current intuitions about what I consider to be subgoals, rather than trusting the AI’s reasoning to find good subgoals to meet primary human goal(s). Another way to phrase that would be that it fiddles with parameters to get special-case results that fit our intuitions rather than focusing on general purpose solutions that would be more likely to produce good results in conditions that we haven’t yet imagined.
For example, concern about whether the AI pays the grocer seems misplaced. If our current intuitions about property rights continue to be good guidelines for maximizing human utility in a world with a powerful AI, why would that AI not reach that conclusion by inferring human utility functions from observed behavior and modeling the effects of property rights on human utility? If not, then why shouldn’t we accept that the AI has decided on something better than property rights (assuming our other methods of verifying that the AI is optimizing human utility show no flaws)?
Is it because we lack decent methods of verifying the AI’s effects on phenomena such as happiness that are more directly related to our utility functions? If so, it would seem to imply that we have an inadequate understanding of what we mean by maximizing utility. I didn’t see a clear explanation of how the AI would infer utility functions from observing human behavior (maybe the source code, which I haven’t read, clarifies it), but that appears to be roughly how humans at their best make the equivalent moral judgments.
I see similar problems with designing the AI to produce the “correct” result with Pascal’s Wager. Tim says “If Heaven and Hell enter into a decision about buying apples, the outcome seems difficult to predict”. Since humans have a poor track record at thinking rationally about very small probabilities and phenomena such as Heaven that are hard to observe, I wouldn’t expect AI unpredictability in this area to be evidence of a problem. It seems more likely that humans are evaluating Pascal’s Wager incorrectly than that a rational AI which can infer most aspects of human utility functions from human behavior will evaluate it incorrectly.

Posts navigation

← Rebelling against our genes
The Execution Channel →
  • Recent Posts

    • The Ageless Brain
    • AI 2027 Thoughts
    • Should AIs be Encouraged to Cooperate?
    • Rain of Tariffs
    • Notes from the TRIIM-X Clinical Trial
    • AI Markets on Manifold
    • Retrospective on my Investing Advice
    • Medical Windfall Prizes
  • Recent Comments

    • The Ageless Brain | Bayesian Investor Blog on The End of Alzheimer’s
    • AI 2027 Thoughts | Bayesian Investor Blog on AI Fire Alarm Scenarios
    • Notes from the TRIIM-X Clinical Trial | Bayesian Investor Blog on True Age
    • Bruce Smith on Retrospective on my Investing Advice
    • Retrospective on my Investing Advice | Bayesian Investor Blog on Advice for Buy-and-Hold Investors
  • Tags

    aging amm autism best posts bias brain bubbles CFAR climate communication skills consciousness covid diet effective altruism empires equality ethics evolution existential risks genetics happiness history honesty industrial revolution information economics IQ kelvinism law macroeconomics meditation mind uploading MIRI neuroscience prediction markets prizes psychology rationality relationships risks seasteading status stock market crash transhumanism war willpower
  • Categories

    • Announcements [B] (6)
    • Book Reviews (281)
    • Economics (183)
      • Idea Futures (44)
      • Investing (82)
    • Life, the Universe, and Everything (153)
      • Fermi Paradox (6)
      • Health (111)
      • Humor (11)
    • Movies (2)
    • Politics (196)
      • China (18)
      • Freedom (19)
      • Mideast (14)
      • U.S. Politics (79)
    • Science and Technology (257)
      • Artificial Intelligence (89)
      • Miscellaneous (20)
      • Molecular Assemblers (Advanced Nanotech) (16)
      • The Flynn Effect (16)
      • The Human Mind (111)
      • Virtual Worlds (4)
    • Uncategorized (14)
Proudly powered by WordPress Theme: Parament by Automattic.