Before: 7y of C++/JS/VR/AR/ML at Varjo, Yle, Reaktor, Automattic ...
After dark: synthesizers & 3D gfx
My guess is that most progress in "agents" will be driven more by human developed LLM-friendly APIs rather than improvement of generalization capabilities in LLMs. No exponential speed-ups there.
My guess is that most progress in "agents" will be driven more by human developed LLM-friendly APIs rather than improvement of generalization capabilities in LLMs. No exponential speed-ups there.
I just meant that language is not the end-all tool for everything. Sure, LLMs can be trained to use tools like calculators and browsers like us. But so far we need to develop those tools, and train the LLMs to use them.
I just meant that language is not the end-all tool for everything. Sure, LLMs can be trained to use tools like calculators and browsers like us. But so far we need to develop those tools, and train the LLMs to use them.
Either via 1000 people annotating data for a year, or a bunch of scientists coming up with neat self-supervised losses for it 😆
Either via 1000 people annotating data for a year, or a bunch of scientists coming up with neat self-supervised losses for it 😆
IMO the (multimodal) LLM paradigm to handle everything in a single model will not scale. Language is a bad abstraction for 1) math (LLMs can't multiply) 2) physical things (where is my cleaning robot?)
End of the sigmoid for data/compute.
IMO the (multimodal) LLM paradigm to handle everything in a single model will not scale. Language is a bad abstraction for 1) math (LLMs can't multiply) 2) physical things (where is my cleaning robot?)
End of the sigmoid for data/compute.
That'll be a difficult read, having limited background in dynamics/control/RL, but it's on the TODO.
Coming from ML, Neural ODEs got me hooked on dynamics and state spaces. Also variational math x optimal control is 🔥
Now learning basics from the book by Brunton/Kutz.
That'll be a difficult read, having limited background in dynamics/control/RL, but it's on the TODO.
Coming from ML, Neural ODEs got me hooked on dynamics and state spaces. Also variational math x optimal control is 🔥
Now learning basics from the book by Brunton/Kutz.
has a nice overview of the papers.
For example
- The 1943 perceptron paper (neural nets)
- Landauer's principle (reversible computing)
- Info theory (Shannon's og paper)
- State space models (Kalman)
...Turing's AI, Nash equilibrium...
has a nice overview of the papers.
For example
- The 1943 perceptron paper (neural nets)
- Landauer's principle (reversible computing)
- Info theory (Shannon's og paper)
- State space models (Kalman)
...Turing's AI, Nash equilibrium...
2) The tokens are (usually) not even full words to begin with. 2/2
2) The tokens are (usually) not even full words to begin with. 2/2
Meanwhile, every time I check X I find a ton of interesting stuff, unfortunately also mixed with a lot of toxic bullshit as well.
Meanwhile, every time I check X I find a ton of interesting stuff, unfortunately also mixed with a lot of toxic bullshit as well.
You have had unique angles for the physics stuff. While anyone with a brain can see, that even though OpenAI does very cool research, they are over-hyping every single release.
You have had unique angles for the physics stuff. While anyone with a brain can see, that even though OpenAI does very cool research, they are over-hyping every single release.
But somehow the fact that there is no consistency/identifiability guarantees even with infinite data makes me afraid of VI 😅
2/2
But somehow the fact that there is no consistency/identifiability guarantees even with infinite data makes me afraid of VI 😅
2/2
In practice, would you, e.g., try adding GMM components to boost ELBO? You’d need to keep everything else fixed for comparability, right?
1/2
In practice, would you, e.g., try adding GMM components to boost ELBO? You’d need to keep everything else fixed for comparability, right?
1/2