Measuring no CoT math time horizon (single forward pass)
lesswrong.com·1d
🎲Deterministic Simulation
Preview
Report Post

Published on December 26, 2025 4:37 PM GMT

A key risk factor for scheming (and misalignment more generally) is opaque reasoning ability. One proxy for this is how good AIs are at solving math problems immediately without any chain-of-thought (CoT) (as in, in a single forward pass). I’ve measured this on a dataset of easy math problems and used this to estimate 50% reliability no-CoT time horizon using the same methodology introduced in Measuring AI Ability to Complete Long Tasks (the METR time horizon paper).

Important caveat: To get huma...

Similar Posts

Loading similar posts...