Nvidia Kicks Off the Next Generation of AI with Rubin

Rubin promises cheaper AI as commenters roast hype, decode jargon, and ask who pays

TLDR: NVIDIA’s Rubin platform promises dramatically cheaper AI and fewer chips for training, while big players like Microsoft line up. Commenters joke about buzzwords, cringe at the emoji hype, and debate whether claimed savings can overcome longer GPU lifecycles and finance pushback — a hype vs. spreadsheet showdown.

NVIDIA just dropped “Rubin,” a six‑chip mega platform that it says slashes AI costs — think up to 10x cheaper text generation and fewer graphics cards needed for training. The internet’s reaction? A chaotic mix of hype, side-eye, and CFO anxiety. One early hero posted the technical deep dive here, while others immediately latched onto the buzzwords. The phrase “extreme co‑design” had folks Googling whether NVIDIA was talking about secure code signing; turns out it just means the parts were built to work together — CPU, GPU, networking, and more — like an AI pit crew.

The biggest drama: can “10x cheaper per token” — cheaper words from AI — beat companies’ new plan to keep their pricey GPUs for six years? One commenter basically asked if finance teams will let anyone upgrade so soon, igniting a clash between optimists (cheaper AI now!) and pragmatists (we’re still paying off the last racks). Meanwhile, the vibes took a meme-y turn when someone called Elon’s emoji‑laden promo “the most cringe” of the week, and a deadpan “Riveting” became the thread’s eye-roll.

Amid the fireworks, there’s real news: Microsoft is building massive “Fairwater” superfactories with Rubin systems, CoreWeave’s jumping in, and Red Hat’s shipping a tuned software stack. Translation: the big leagues are betting on Rubin — but the comments want receipts, not just starry names and cosmic branding.

Key Points

  • NVIDIA launched the Rubin platform with six co-designed chips to function as one AI supercomputer.
  • Rubin claims up to 10x reduction in inference token cost and 4x fewer GPUs for MoE training versus Blackwell.
  • New innovations include latest NVLink, Transformer Engine, Confidential Computing, RAS Engine, and the Vera CPU.
  • Spectrum-X Ethernet Photonics systems are stated to deliver 5x improved power efficiency and uptime.
  • Ecosystem announcements include Microsoft Fairwater AI superfactories, CoreWeave availability, and expanded Red Hat collaboration.

Hottest takes

"it took a couple searches to figure out that 'extreme codesign' wasn't actually code-signing" — Groxx
"Elon's emoji-filled blurb for that press release is the most cringe things I've seen this week." — metalliqaz
"does that play well or not well with the recently updated GPU deprecation schedules" — codyb
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.