Michael Hahn
m-hahn.bsky.social
Michael Hahn
@m-hahn.bsky.social
Prof at Saarland. NLP and machine learning.
Theory and interpretability of LLMs.
https://www.mhahn.info
5/8 Multiplication: LLMs notoriously flub middle digits of big-integer products. Those digits hinge on all input bits, so you get a linear CoT lower bound. Our best upper bound uses Schönhage–Strassen in O(n log n) steps— closing the log(n) gap is open.
May 5, 2025 at 12:25 PM
Chain-of-Thought (CoT) reasoning lets LLMs solve complex tasks, but long CoTs are expensive. How short can they be while still working? Our new ICML paper tackles this foundational question.
May 5, 2025 at 12:25 PM