Ask HN: Claude Opus performance affected by time of day?

Is Claude getting sleepy at noon? Users say yes, Anthropic says nope

TLDR: Users suspect AI quality dips during busy hours and after outages, possibly due to fallback models, while Anthropic insists performance doesn’t change by time of day. The crowd wants benchmarks, jokes about “tired” bots, and debates trust vs. transparency—because reliability matters when your work depends on these tools.

The thread kicked off with a simple question—does Claude Opus perform worse at certain times?—and immediately turned into a spicy group chat about AI mood swings. One user swore that Google’s Gemini “falls apart” right after big announcements, with iffy answers, lost context, and UI meltdowns. Another claimed peak hours might trigger fallback models to lighter, cheaper brains, especially around West Coast midday. Cue drama: is this shady or just smart load balancing? The plot thickened as folks pointed at service hiccups and outages: one commenter said Claude 4.5 felt “very Sonnet” (a lighter model) for hours after issues, before “mysteriously regaining its wits.” Meanwhile, the ethics brigade arrived—an Anthropic fan cited the company’s own postmortem and said they don’t degrade by time of day, linking receipts: Anthropic’s postmortem. The community split into camps: the conspiracy testers promising hourly benchmarks, the trusters who think Anthropic keeps it clean, and the comedians who blame the bot being “tired” or “hungover.” Bonus spark: complaints that Claude’s “long conversation reminder” tells you to go to bed instead of answering. Verdict? No hard proof—just vibes, graphs-to-be, and some peak-hour paranoia.

Key Points

  • The author reports Gemini’s quality varies, with notable degradation during Google public announcements that likely trigger high usage.
  • Failure modes include lower-quality responses, random errors, loss of context even in new chats, and UI-level issues like incomplete answers.
  • A suspected cause is dynamic load balancing under high demand, potentially involving fallback to lighter models or resource limits per prompt.
  • For Claude, Opus 4.5 is usually reliable but shows changes around service interruptions; post-resolution results can resemble Sonnet and be more error-prone before stabilizing.
  • The article states providers don’t silently switch to cheaper models and notes Anthropic’s “long conversation reminder,” which may distract from answering.

Hottest takes

“secretly routing traffic to a worse model is a bit shady” — janalsncm
“then a bit later, Claude mysteriously regains its wits” — bayarearefugee
“Anthropic very strongly claim that they don’t degrade model performance by time of day” — oncallthrow
Made with <3 by @siedrix and @shesho from CDMX. Powered by Forge&Hive.