Trending topics
#
Bonk Eco continues to show strength amid $USELESS rally
#
Pump.fun to raise $1B token sale, traders speculating on airdrop
#
Boop.Fun leading the way with a new launchpad on Solana.

Gavin Baker
Managing Partner & CIO, @atreidesmgmt. Husband, @l3eckyy. No investment advice, views my own. https://t.co/pFe9KmNu9U
Nvidia is buying Groq for two reasons imo.
1) Inference is disaggregating into prefill and decode. SRAM architectures have unique advantages in decode for workloads where performance is primarily a function of memory bandwidth. Rubin CPX, Rubin and the putative “Rubin SRAM” variant derived from Groq should give Nvidia the ability to mix and match chips to create the optimal balance of performance vs. cost for each workload. Rubin CPX is optimized for massive context windows during prefill as a result of super high memory capacity with its relatively low bandwidth GDDR DRAM. Rubin is the workhorse for training and high density, batched inference workloads with its HBM DRAM striking a balance between memory bandwidth and capacity. The Groq-derived "Rubin SRAM" is optimized for ultra-low latency agentic reasoning inference workloads as a result of SRAM’s extremely high memory bandwidth at the cost of lower memory capacity. In the latter case, either CPX or the normal Rubin will likely be used for prefill.
2) It has been clear for a long time that SRAM architectures can hit token per second metrics much higher than GPUs, TPUs or any ASIC that we have yet seen. Extremely low latency per individual user at the expense of throughput per dollar. It was less clear 18 months ago whether end users were willing to pay for this speed (SRAM more expensive per token due to much smaller batch sizes). It is now abundantly clear from Cerebras and Groq’s recent results that users are willing to pay for speed.
Increases my confidence that all ASICs except TPU, AI5 and Trainium will eventually be canceled. Good luck competing with the 3 Rubin variants and multiple associated networking chips. Although it does sound like OpenAI’s ASIC will be surprisingly good (much better than the Meta and Microsoft ASICs).
Let’s see what AMD does. Intel already moving in this direction (they have a prefill optimized SKU and purchased SambaNova, which was the weakest SRAM competitor). Kinda funny that Meta bought Rivos.
And Cerebras, where I am biased, is now in a very interesting and highly strategic position as the last (per public knowledge) independent SRAM player that was ahead of Groq on all public benchmarks. Groq’s “many chip” rack architecture, however, was much easier to integrate with Nvidia’s networking stack and perhaps even within a single rack while Cerebras’s WSE almost has to be an independent rack.
233
Deeply amused by all the confident commentary that datacenters in space do not work from a physics and engineering perspective.
Elon operates two of the largest coherent GPU clusters in the world, SpaceX is responsible for over 90% of mass to orbit and SpaceX operates the largest satellite constellation in the solar system. More than 10 years later, no other company or country can consistently land and reuse orbital rockets.
He publicly stated that the “lowest cost way to do AI compute will be with solar powered satellites.”
Maybe, just maybe, his “pencil and paper analysis of the physics or the economics at play” is superior to yours. There might have even been more than just a “pencil and paper analysis” of the subject done by some of the best engineers in the world. Perhaps they have thought of a cooling solution that has not occurred to the galaxy brain accounts here even after they took several minutes to carefully think about the problem.
The CEO of Google also agrees that data centers in space will be “normal” within a decade.
If you are not currently operating a large AI datacenter, a large satellite cluster and have not landed a rocket, maybe be a little less quick to confidently assume that Elon and Google are *both* wrong on this topic.
Especially when there is a working, albeit very small, datacenter in space *today* - Starcloud’s orbital setup just successfully trained an LLM. Great name btw.
Yes, I am biased on these topics and as ever, time will tell.
305
Top
Ranking
Favorites

