“Problems I’ve Tried to Legibilize” by Wei Dai
Update: 2025-11-10
Description
Looking back, it appears that much of my intellectual output could be described as legibilizing work, or trying to make certain problems in AI risk more legible to myself and others. I've organized the relevant posts and comments into the following list, which can also serve as a partial guide to problems that may need to be further legibilized, especially beyond LW/rationalists, to AI researchers, funders, company leaders, government policymakers, their advisors (including future AI advisors), and the general public.
- Philosophical problems
- Probability theory
- Decision theory
- Beyond astronomical waste (possibility of influencing vastly larger universes beyond our own)
- Interaction between bargaining and logical uncertainty
- Metaethics
- Metaphilosophy: 1, 2
- Problems with specific philosophical and alignment ideas
- Utilitarianism: 1, 2
- Solomonoff induction
- "Provable" safety
- CEV
- Corrigibility
- IDA (and many scattered comments)
- UDASSA
- UDT
- Human-AI safety (x- and s-risks arising from the interaction between human nature and AI design)
- Value differences/conflicts between humans
- “Morality is scary” (human morality is often the result of status games amplifying random aspects of human value, with frightening results)
- [...]
---
First published:
November 9th, 2025
Source:
https://www.lesswrong.com/posts/7XGdkATAvCTvn4FGu/problems-i-ve-tried-to-legibilize
---
Narrated by TYPE III AUDIO.
Comments
In Channel



