Dhruv Rawat
banner
imdhruv.bsky.social
Dhruv Rawat
@imdhruv.bsky.social
distributed systems, networks and ml
prev: yugabyte, cs@bitspilani
4. Low-Latency Interconnect: Reducing latency with directions like:

- high-connectivity topologies (tree, dragonfly), that require fewer hops
- in-network acceleration of communication collectives (broadcast, all-reduce) used by LLMs
- AI chip optimization
- codesigning reliability and interconnect
January 25, 2026 at 9:18 AM
3. 3D memory-logic stacking: Stacking memory and logic layers vertically (TSVs) to get a wide-and-dense memory interface for high bandwidth at low power. This is a version of PNM.
January 25, 2026 at 9:11 AM
2. Processing-Near-Memory (PNM): Moving computation closer to where data is stored (but separate dies) to overcome bandwidth limitations. They clearly distinguish it from Processing-in-Memory (PIM), in which the processor and memory are on the same die.
January 25, 2026 at 9:09 AM
1. High Bandwidth Flash (HBF): Developing flash storage that offers 10x the memory capacity while maintaining bandwidth comparable to HBM. This enables new capabilities for LLM inference, such as 10x weight memory, 10x context memory, a smaller inference system, and greater resource capacity.
January 25, 2026 at 9:04 AM