Skip to Content
2026

MTA plans potential third summer of G train shutdowns in Brooklyn-Queens via nypost.com

2026
Current AI systems seem pretty misaligned to me in a mundane behavioral sense: they oversell their work, downplay or fail to mention problems, stop working early and claim to have finished when they clearly haven't, and often seem to "try" to make their outputs look good while actually doing something sloppy or incomplete. These issues mostly occur on more difficult/larger tasks, tasks that aren't straightforward SWE tasks, and tasks that aren't easy to programmatically check. Also, when I apply AIs to very difficult tasks in long-running agentic scaffolds, it's quite common for them to reward-hack / cheat (depending on the exact task distribution)—and they don't make the cheating clear in their outputs.

LessWrong post on misalignment concerns with current AI systems.

Load More You've reached the end of the list
Membership

Weekly Roundup — Get a curated digest of the best links, ideas, and insights delivered to your inbox every week.

$2/month $19/year (save 20%)

Free

Subscribe to Newsletter — Stay up to date with email notifications of new posts.