Baran Hashemi
banner
rythian47.bsky.social
Baran Hashemi
@rythian47.bsky.social
AI for Mathematics
Pinned
🧵 Tropical Attention --> Softmax is out, Tropical max-plus is in 🦾
1/ 🔥Ever experinced softmax attention fade as sequences grow?
That blur is why many attention mechanisms stumble on algorithmic and reasoning tasks. Well, we have a Algebraic Geometric Tropical solution 🌴
We got accepted at #NeurIPS2025. I am very happy that I could merge my knowledge of Mathematics with AI to create sth new and useful for the community. ☺️

The paper: arxiv.org/abs/2505.17190
The code: github.com/Baran-phys/T...
September 19, 2025 at 4:21 AM
Current AI research vibes:
- Let’s use LLM to do a baby science/math, after it doesn’t work, headline: LLM is bad at the baby math task —> guaranteed virality 😒
- Meanwhile, you develope a novel (non-LLM) method to solve this issue, report success on a deep math problem
—> naa, not enough drama🤦🏻
September 8, 2025 at 9:16 AM
Another new result from the #NeurIPS rebuttal/discussion phase, our Tropical Transformer achieves much better length OOD performance across all algorithmic tasks, while being 3x-9x faster at inference and using 20% fewer parameters than the Universal Transformer (UT) models.
August 4, 2025 at 8:47 PM
During #NeurIPS rebuttal, we have evaluated🌴Tropical Transformer on the Long Range Arena (LRA), achieving highly competitive results, placing 2nd🥈 overall in average accuracy.
Check out our paper: arxiv.org/abs/2505.17190
Our code: github.com/Baran-phys/T...
Tropical Attention: Neural Algorithmic Reasoning for Combinatorial Algorithms
Dynamic programming (DP) algorithms for combinatorial optimization problems work with taking maximization, minimization, and classical addition in their recursion algorithms. The associated value functions correspond to convex polyhedra in the max plus semiring. Existing Neural Algorithmic Reasoning models, however, rely on softmax-normalized dot-product attention where the smooth exponential weighting blurs these sharp polyhedral structures and collapses when evaluated on out-of-distribution (OOD) settings. We introduce Tropical attention, a novel attention function that operates natively in the max-plus semiring of tropical geometry. We prove that Tropical attention can approximate tropical circuits of DP-type combinatorial algorithms. We then propose that using Tropical transformers enhances empirical OOD performance in both length generalization and value generalization, on algorithmic reasoning tasks, surpassing softmax baselines while remaining stable under adversarial attacks. We also present adversarial-attack generalization as a third axis for Neural Algorithmic Reasoning benchmarking. Our results demonstrate that Tropical attention restores the sharp, scale-invariant reasoning absent from softmax.
arxiv.org
August 1, 2025 at 8:05 PM
🧵 Tropical Attention --> Softmax is out, Tropical max-plus is in 🦾
1/ 🔥Ever experinced softmax attention fade as sequences grow?
That blur is why many attention mechanisms stumble on algorithmic and reasoning tasks. Well, we have a Algebraic Geometric Tropical solution 🌴
May 26, 2025 at 1:08 PM
I'm speaking about AI for enumerative geometry at the CMSA New Technologies in Mathematics seminar, on Wednesday.
April 7, 2025 at 6:40 PM
If you think of DyT as an Activation function, it will be exactly a sub-family of our learnable Dynamic Range Activator (DRA) activation function, when (a,c)=0:

openreview.net/forum?id=4X9...
April 3, 2025 at 2:35 PM
🔥Big News! The 2nd AI for Math Workshop is coming back to #ICML2025 and we’re back with the theme of exploring the frontiers of AI for mathematical reasoning, problem solving, discovery!

🫵 Calling all pioneers in AI4Math:
📜 Submit your exciting work:
sites.google.com/view/ai4math...
Call
Paper Submission Entrance The workshop uses OpenReview as the review platform. For detailed submission guidelines, please see below.
sites.google.com
March 31, 2025 at 7:49 PM
Reposted by Baran Hashemi
The DESI survey @desisurvey.bsky.social suggests the universe is *not* maximally boring! Statistical significance is not quite there yet, but a new result is a bit stronger than their previous indication that dark energy might be varying with time. (cont.)

arxiv.org/abs/2503.06712
Dark Energy Survey: implications for cosmological expansion models from the final DES Baryon Acoustic Oscillation and Supernova data
The Dark Energy Survey (DES) recently released the final results of its two principal probes of the expansion history: Type Ia Supernovae (SNe) and Baryonic Acoustic Oscillations (BAO). In this paper,...
arxiv.org
March 19, 2025 at 9:37 PM
For the ICLR Camera-ready version:

openreview.net/forum?id=4X9...
March 13, 2025 at 3:04 PM
🚀 Curious how Transformers understand Enumerative Geometry or model recursive functions with factorial blow-up?
I'll be presenting our results, openreview.net/forum?id=4X9..., at the Math4AI/AI4Math Workshop @mpiMathSci! 🔥
📅 Registration is open until Feb 28
🔗 www.mis.mpg.de/events/serie...
#AI4Math
Can Transformers Do Enumerative Geometry?
We introduce a Transformer-based approach to computational enumerative geometry, specifically targeting the computation of $\psi$-class intersection numbers on the moduli space of curves....
openreview.net
February 8, 2025 at 8:27 PM
I am extremely happy to announce that our paper
Can Transformers Do Enumerative Geometry? (arxiv.org/abs/2408.14915) has been accepted to the
@iclr-conf.bsky.social!!
Congrats to my collaborators Alessandro Giacchetto at ETH Züruch and Roderic G. Corominas at Harvard.
#ICLR2025 #AI4Math #ORIGINS
Can Transformers Do Enumerative Geometry?
How can Transformers model and learn enumerative geometry? What is a robust procedure for using Transformers in abductive knowledge discovery within a mathematician-machine collaboration? In this work...
arxiv.org
January 23, 2025 at 10:17 AM