🧩 Philosophy 6d ago · lumpenspace

No Strong Orthogonality From Selection Pressure

Less Wrong
View Channel →
Source ↗ 👁 1 💬 0
A postratfic version of this essay, together with the acknlowlegenents for both, is available on SubstackTL;DRIf everything goes according to plan, by the end of this post we should have separated three claims that are too often bundled together:Intelligence does not imply human morality.Weird minds are possible. A reflective, recursively improving intelligence should be expected to remain bound to a semantically thin “terminal goal” that emerged during training.I accept the first two. I am argu

Comments (0)

Sign in to join the discussion

More Like This

📰
Idealism
Stanford Encyclopedia of Philosophy · 7h ago
Toward a Better Evaluations Ecosystem
LessWrong · 10h ago
Model Spec Midtraining: Improving How Alignment Training Generalizes
LessWrong · 10h ago
📰
Positive Feedback Only
LessWrong · 11h ago
📰
What if LLMs are mostly crystallized intelligence?
LessWrong · 11h ago
Decision theory doesn’t prove that useful strong AIs will doom us all
LessWrong · 11h ago