Decision theory doesn’t prove that useful strong AIs will doom us all
Source ↗
👁 0
💬 0
Bottom-line up frontTraining for optimal behavior doesn't inevitably lead to act-utilitarian world optimizers ("WorldSUM agents"). People will prefer to deploy agents with more virtue-ethicsy / deontological approaches, for 2-3 reasons: 1) Traditional misalignment concerns2) Even if they have "the right values", we don't trust them to get the calculation right -- just like human subordinates.Similarly, many people including AI labs will prefer agents whose action space is bounded, because they
Comments (0)