Has Google solved two of AI's oldest problems?

Miracle model or hype? Users gush over 'one‑prompt apps' while skeptics shout clickbait

TLDR: Google’s AI Studio is reportedly serving A/B replies that hint at a new model doing shockingly good handwriting transcription and reasoning. The crowd is split: excited historians cheer, while skeptics cry clickbait, random‑seed vibes, and Betteridge’s law, demanding proof this isn’t just another hype blip.

Google’s AI Studio is quietly showing some users two answers and asking them to pick a winner—classic pre‑launch A/B testing—and rumor has it this might be the next Gemini model. One tester claims it spit out full operating systems, 3D tools, and even emulators from single prompts, then aced tricky handwritten history docs with what felt like human‑level reasoning. The comments? A battlefield. Scholars and hobbyists are starry‑eyed: one historian cheered that tackling old Spanish archives is “a nightmare,” so if this works, it’s a lifesaver. But the skeptics arrive with air horns. A top reply calls the write‑up “clickbait,” advising readers to skip straight to “The Ultimate Test,” while another sneers that it’s “just another academic… beguiled by ‘AI’.” A pragmatic voice says the A/B felt like the same model with a different random seed—translation: maybe nothing’s new at all. Jokes flew too: someone dropped Betteridge’s law, implying any headline phrased as a question answers itself. The vibe is pure internet: half “this changes everything,” half “show me reproducible receipts.” If true, this could transform archives, research, and any job where seeing and thinking must fuse. If not, it’s just another hype cycle in a very crowded week.

Key Points

  • Users in Google’s AI Studio encountered A/B result prompts, suggesting pre-release model testing.
  • Speculation centers on a new Gemini model (possibly Gemini-3) being evaluated.
  • Reports claim the model can generate complex software (e.g., OS clones, emulators, productivity tools) from single prompts.
  • The author’s test on handwritten historical documents yielded near-expert transcription and unprompted expert-level reasoning.
  • The article argues that effective handwriting transcription requires both visual accuracy and contextual/logical reasoning, and suggests the observed model may advance on both fronts.

Hottest takes

“holy cow reading 16th century handwritten Spanish and translating it at the same time is a nightmare” — throwup238
“Skip to the section headed ‘The Ultimate Test’ for the resolution of the clickbait” — netsharc
“It is strange that others can never reproduce such amazing feats.” — bgwalter
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.