albertgong.bsky.social
@albertgong.bsky.social
Reposted
Your data is low-rank, so stop wasting compute! In our new paper on low-rank thinning, we share one weird trick to speed up Transformer inference, SGD training, and hypothesis testing at scale. Come by ICML poster W-1012 Tuesday at 4:30!
New guarantees for approximating attention, accelerating SGD, and testing sample quality in near-linear time
July 14, 2025 at 6:29 PM