it feels somewhat antisocial but I don't want to publish that info
Claude Code Opus 4.5 has the juice, I have no other way to put it. Astonishing 1yr improvement
it still can't cd. scares me when it tries to git stash. I guess it's like how it can't play chess verbally? Can't do that kind of state well (yet?). But cd surely could be learned
Claude Code Opus 4.5 has the juice, I have no other way to put it. Astonishing 1yr improvement
it still can't cd. scares me when it tries to git stash. I guess it's like how it can't play chess verbally? Can't do that kind of state well (yet?). But cd surely could be learned
there are many failure modes, like it still can't be trusted with cd:
> Oh, I'm in the wrong directory. Let me use absolute paths:
but it's very good. Hyperbolic words good
there are many failure modes, like it still can't be trusted with cd:
> Oh, I'm in the wrong directory. Let me use absolute paths:
but it's very good. Hyperbolic words good
The steady improvement of intent-inference and faithful instruction following has been a fun ride. Implicit style conformance too!
Sonnet 4.5 doesn't have the depth or subtlety of Opus 4.1, but for coding, I rarely reach for Opus now
The steady improvement of intent-inference and faithful instruction following has been a fun ride. Implicit style conformance too!
Sonnet 4.5 doesn't have the depth or subtlety of Opus 4.1, but for coding, I rarely reach for Opus now
it feels somewhat antisocial but I don't want to publish that info
it feels somewhat antisocial but I don't want to publish that info
me saying 'more like "robocrap"' is like 'zoom in and search'
I didn't like the taste of the first place, but it was an easy hop from there to something that resonated
claude.ai/share/8336be...
me saying 'more like "robocrap"' is like 'zoom in and search'
I didn't like the taste of the first place, but it was an easy hop from there to something that resonated
claude.ai/share/8336be...
just a potty humor example, but it's clear LLMs can synthesize new legible info, which many or most users already know, but it seems widely doubted still
just a potty humor example, but it's clear LLMs can synthesize new legible info, which many or most users already know, but it seems widely doubted still
bsky.app/profile/ryan...
bsky.app/profile/ryan...
"next token predictor" acknowledges linear time but I don't think it's as profoundly limiting as it sounds, especially for machine-compressed time at many hz
"next token predictor" acknowledges linear time but I don't think it's as profoundly limiting as it sounds, especially for machine-compressed time at many hz