Another neat alignment-motivated complexity theory conjecture from ARC! I am excited for more theory folk to work on alignment, and crisply defined conjectures are a great starting point.
Some thoughts on how this conjecture relates to the overall problem. 🧵
https://www.alignment.org/blog/a-computational-no-coincidence-principle
Some thoughts on how this conjecture relates to the overall problem. 🧵
https://www.alignment.org/blog/a-computational-no-coincidence-principle
Comments
1. Follow AI reasoning where we can
2. Admit we will sometimes fail to follow AI reasoning
3. Define a notion of "heuristic explanation" that holds even if (2) fails.
4. Use (3) to distinguish (1) and (2).