I obviously review all the code I ship, but if do I miss things, I basically have to present the reasoning as my own before fixing it. It'd be poor form to just blame the LLM for it
I obviously review all the code I ship, but if do I miss things, I basically have to present the reasoning as my own before fixing it. It'd be poor form to just blame the LLM for it
Every large language model has an unembedding matrix that turns hidden states (model thoughts) into tokens (output text). This matrix associates a vector with every token in the vocabulary. The idea is you compare the last hidden state with the vectors to pick the right token.
Every large language model has an unembedding matrix that turns hidden states (model thoughts) into tokens (output text). This matrix associates a vector with every token in the vocabulary. The idea is you compare the last hidden state with the vectors to pick the right token.
openrxiv.org/enabling-rev...
openrxiv.org/enabling-rev...
#VRChat
#VRChat
the difference: I do not use Discover
the difference: I do not use Discover
From Patch 7.4 onward, you'll be able to equip any glamours regardless of your class/job! ✨
Mix and match your favorite outfits and slay on the battlefield! 💅
From Patch 7.4 onward, you'll be able to equip any glamours regardless of your class/job! ✨
Mix and match your favorite outfits and slay on the battlefield! 💅
GPT-5 has lower per-token cost than Sonnet 4.5, but costs more in the benchmark because GPT-5 agents are more "token hungry".
GPT-5 has lower per-token cost than Sonnet 4.5, but costs more in the benchmark because GPT-5 agents are more "token hungry".
The one interesting idea I see here is "what if we gave agents DIDs" but I'm not entirely sure what the implications of that would be
The one interesting idea I see here is "what if we gave agents DIDs" but I'm not entirely sure what the implications of that would be
white box: their method, using SAE
black box: an LLMs own judgement via text outputs
it’s a lot cheaper because they only need one set of activations, i.e. only one output token + a single random forest inference