evolvingstuff
banner
evolvingstuff.bsky.social
evolvingstuff
@evolvingstuff.bsky.social
I post about machine learning and occasionally some other stuff.
Reposted by evolvingstuff
MIT researchers revealed that decoder-only transformers can't learn inverse permutation tasks, challenging their expressive capacity. New methods, like using "scratch tokens," could enhance reasoning abilities in large language models. https://arxiv.org/abs/2509.24125
The Impossibility of Inverse Permutation Learning in Transformer Models
ArXiv link for The Impossibility of Inverse Permutation Learning in Transformer Models
arxiv.org
September 30, 2025 at 9:31 AM
Reposted by evolvingstuff
GLM-4.6 is out and things aren’t looking good for Sonnet 4.5

- improved tool calling
- improved token utilization
- improved writing

docs.z.ai/guides/llm/g...
September 30, 2025 at 10:33 AM