Reinforcement learning / Neural Networks Plasticity / Neural Network Representations / AI4Science
Most RL methods’ performance saturate at ~5 layers. In this work led by Kevin Wang, we crack the right configuration for scaling Contrastive RL and go beyond 1000 layers NNs! Deep NNs unlock emergent behaviors and other cool properties. Check out Kevin’s thread!
Webpage+Paper+Code: wang-kevin3290.github.io/scaling-crl/
Most RL methods’ performance saturate at ~5 layers. In this work led by Kevin Wang, we crack the right configuration for scaling Contrastive RL and go beyond 1000 layers NNs! Deep NNs unlock emergent behaviors and other cool properties. Check out Kevin’s thread!
Apparently, you achieve 🚨state-of-the-art🚨 model merging results! 🔥
✨ Introducing “No Task Left Behind: Isotropic Model Merging with Common and Task-Specific Subspaces”
Apparently, you achieve 🚨state-of-the-art🚨 model merging results! 🔥
✨ Introducing “No Task Left Behind: Isotropic Model Merging with Common and Task-Specific Subspaces”
🎉 Our paper "Learning Graph Representation of Agent Diffusers (LGR-AD)" has been accepted as a full paper at #AAMAS (A*) International Conference on Autonomous Agents and Multiagent Systems!
#diffusion #graphs #agentsystem
@ideas-ncbr.bsky.social #WarszawUniversityOfTechnology
🎉 Our paper "Learning Graph Representation of Agent Diffusers (LGR-AD)" has been accepted as a full paper at #AAMAS (A*) International Conference on Autonomous Agents and Multiagent Systems!
#diffusion #graphs #agentsystem
@ideas-ncbr.bsky.social #WarszawUniversityOfTechnology
📍 Poster #6302
📅 West Ballroom A-D
🕚 Friday, 11:00-14:00
Join us to discuss with Michał Nauman and me. Let’s talk SOTA in RL! 💪
🧵👇
📍 Poster #6302
📅 West Ballroom A-D
🕚 Friday, 11:00-14:00
Join us to discuss with Michał Nauman and me. Let’s talk SOTA in RL! 💪
🧵👇
Excellent news from NeurIPS. Two papers in, including a spotlight.
1. Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
2. Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
Excellent news from NeurIPS. Two papers in, including a spotlight.
1. Repurposing Language Models into Embedding Models: Finding the Compute-Optimal Recipe
2. Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control
📜Let's focus on the BRO algorithm, introduced in 🔦spotlight paper by Michał Nauman, @mateuszostaszewski.bsky.social, Krzysztof Jankowski, @piotrmilos.bsky.social, Marek Cygan, to find out why the second dog runs better ➡️ ideas-ncbr.pl/en/bro-algor...
📜Let's focus on the BRO algorithm, introduced in 🔦spotlight paper by Michał Nauman, @mateuszostaszewski.bsky.social, Krzysztof Jankowski, @piotrmilos.bsky.social, Marek Cygan, to find out why the second dog runs better ➡️ ideas-ncbr.pl/en/bro-algor...