Bayesian Investor Blog

Ramblings of a somewhat libertarian stock market speculator

Steve Omohundro on AI Risks

Posted by Peter on March 25, 2008
Posted in: Artificial Intelligence. Tagged: existential risks.

Steve Omohundro has recently written a paper and given a talk (a video should become available soon) on AI ethics with arguments whose most important concerns resemble Eliezer Yudkowsky’s. I find Steve’s style more organized and more likely to convince mainstream researchers than Eliezer’s best attempt so far.
Steve avoids Eliezer’s suspicious claims about how fast AI will take off, and phrases his arguments in ways that are largely independent of the takeoff speed. But a sentence or two in the conclusion of his paper suggests that he is leaning toward solutions which assume multiple AIs will be able to safeguard against a single AI imposing its goals on the world. He doesn’t appear to have a good reason to consider this assumption reliable, but at least he doesn’t show the kind of disturbing certainty that Eliezer has about the first self-improving AI becoming powerful enough to take over the world.
Possibly the most important news in Steve’s talk was his statement that he had largely stopped working to create intelligent software due to his concerns about safely specifying goals for an AI. He indicated that one important insight that contributed to this change of mind came when Carl Shulman pointed out a flaw in Steve’s proposal for a utility function which included a goal of the AI shutting itself off after a specified time (the flaw involves a small chance of physics being different from apparent physics and how the AI will evaluate expected utilities resulting from that improbable physics).

Posts navigation

← Futarchy Mailing List
Greenspan’s Age of Turbulence →
  • Recent Posts

    • The Ageless Brain
    • AI 2027 Thoughts
    • Should AIs be Encouraged to Cooperate?
    • Rain of Tariffs
    • Notes from the TRIIM-X Clinical Trial
    • AI Markets on Manifold
    • Retrospective on my Investing Advice
    • Medical Windfall Prizes
  • Recent Comments

    • The Ageless Brain | Bayesian Investor Blog on The End of Alzheimer’s
    • AI 2027 Thoughts | Bayesian Investor Blog on AI Fire Alarm Scenarios
    • Notes from the TRIIM-X Clinical Trial | Bayesian Investor Blog on True Age
    • Bruce Smith on Retrospective on my Investing Advice
    • Retrospective on my Investing Advice | Bayesian Investor Blog on Advice for Buy-and-Hold Investors
  • Tags

    aging amm autism best posts bias brain bubbles CFAR climate communication skills consciousness covid diet effective altruism empires equality ethics evolution existential risks genetics happiness history honesty industrial revolution information economics IQ kelvinism law macroeconomics meditation mind uploading MIRI neuroscience prediction markets prizes psychology rationality relationships risks seasteading status stock market crash transhumanism war willpower
  • Categories

    • Announcements [B] (6)
    • Book Reviews (281)
    • Economics (183)
      • Idea Futures (44)
      • Investing (82)
    • Life, the Universe, and Everything (153)
      • Fermi Paradox (6)
      • Health (111)
      • Humor (11)
    • Movies (2)
    • Politics (196)
      • China (18)
      • Freedom (19)
      • Mideast (14)
      • U.S. Politics (79)
    • Science and Technology (257)
      • Artificial Intelligence (89)
      • Miscellaneous (20)
      • Molecular Assemblers (Advanced Nanotech) (16)
      • The Flynn Effect (16)
      • The Human Mind (111)
      • Virtual Worlds (4)
    • Uncategorized (14)
Proudly powered by WordPress Theme: Parament by Automattic.