Until December 7th, you can send us your suggestions, and we will use them to shape the program.
icml.cc/Conferences/...
Until December 7th, you can send us your suggestions, and we will use them to shape the program.
icml.cc/Conferences/...
Lots of progress in RL research over last 10 years, but too much performance-driven => overfitting to benchmarks (like the ALE).
1⃣ Let's advance science of RL
2⃣ Let's be explicit about how benchmarks map to formalism
1/X
Lots of progress in RL research over last 10 years, but too much performance-driven => overfitting to benchmarks (like the ALE).
1⃣ Let's advance science of RL
2⃣ Let's be explicit about how benchmarks map to formalism
1/X
Endless thanks to my amazing co-authors @claireve.bsky.social and @keggensperger.bsky.social
📄 Read it on arXiv: arxiv.org/abs/2505.05226
(1/3)
Endless thanks to my amazing co-authors @claireve.bsky.social and @keggensperger.bsky.social
📄 Read it on arXiv: arxiv.org/abs/2505.05226
(1/3)
I finally gave in and made a nice blog post about my most recent paper. This was a surprising amount of work, so please be nice and go read it!
I finally gave in and made a nice blog post about my most recent paper. This was a surprising amount of work, so please be nice and go read it!
He made a great thread about it, come chat with us about POMDP theory :)
He made a great thread about it, come chat with us about POMDP theory :)
⭐ Avery Ma
⭐ Claas Voelcker (cvoelcker.bsky.social)
⭐ Tyler Kastner
Meet them to talk about Model-based RL, Distributional RL, and Jailbreaking LLMs.
⭐ Avery Ma
⭐ Claas Voelcker (cvoelcker.bsky.social)
⭐ Tyler Kastner
Meet them to talk about Model-based RL, Distributional RL, and Jailbreaking LLMs.
🧠🤖 #MLSky
"AI systems will not acquire the flexibility and adaptability of human intelligence until they can actually learn like humans do, shining brightly with their own light rather than observing a shadow from ours."
🧠🤖
sergeylevine.substack.com/p/language-m...
🧠🤖 #MLSky
Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?
TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases – without extra loss terms and predictors!
🧵 (1/10)
Can we simultaneously learn transformation-invariant and transformation-equivariant representations with self-supervised learning?
TL;DR Yes! This is possible via simple predictive learning & architectural inductive biases – without extra loss terms and predictors!
🧵 (1/10)
arxiv.org/abs/2504.12501
arxiv.org/abs/2504.12501
🔎 For a deeper dive into the theory:
blog.foersterlab.com/fixing-td-pa...
blog.foersterlab.com/fixing-td-pa...
See you in Singapore! 🇸🇬
This mostly covers France and Germany for now but I’m willing to extend it. I build on @ellis.eu resources and my own knowledge of these systems.
This mostly covers France and Germany for now but I’m willing to extend it. I build on @ellis.eu resources and my own knowledge of these systems.
Next:
* Crafting more blog content into future topics,
* DPO+ chapter,
* Meeting with publishers to get wheels turning on physical copies,
* Cleaning & cohesiveness
rlhfbook.com
Next:
* Crafting more blog content into future topics,
* DPO+ chapter,
* Meeting with publishers to get wheels turning on physical copies,
* Cleaning & cohesiveness
rlhfbook.com
Get your application in early to be a student or teaching assistant for this year’s courses!
Applications are due Sunday, March 23.
Apply & learn more: neuromatch.io/courses/
#mlsky #compneurosky #ai #climatesolutions #ScienceEdu 🧪
Get your application in early to be a student or teaching assistant for this year’s courses!
Applications are due Sunday, March 23.
Apply & learn more: neuromatch.io/courses/
#mlsky #compneurosky #ai #climatesolutions #ScienceEdu 🧪
2018 ResNet: A more accurate model is trainable in a 1/2 hour on a single GPU.
What stops this from happening for LLMs?
What a world we're in where this well-trodden pattern rocks financial markets and escalates geopolitical conflict.
2018 ResNet: A more accurate model is trainable in a 1/2 hour on a single GPU.
What stops this from happening for LLMs?
Talks on SlideShare are really hard to track!
Talks on SlideShare are really hard to track!
if i were stuck on a desert island i'd hope to have something better to read than #RL papers... but anyway, here's a thread with my choices, maybe you can read them on your flight to @neuripsconf.bsky.social #NeurIPS2024 .
Enjoy!
if i were stuck on a desert island i'd hope to have something better to read than #RL papers... but anyway, here's a thread with my choices, maybe you can read them on your flight to @neuripsconf.bsky.social #NeurIPS2024 .
Enjoy!
go.bsky.app/3WPHcHg
go.bsky.app/3WPHcHg