1a3orn
2025.10.13
70% relevant
The post and comments cite Meta’s CWM paper and the R1 language-consistency ablation to argue that pressuring models toward neat, legible CoT (or away from ‘gibberish’) can slightly degrade performance, paralleling findings that more/cleaner test‑time reasoning doesn’t always help and can even hurt.