Rate Limits, Peak Hours, and the Netflix Home Page | LUMP Depot

Last week I argued you should be willing to pay something like ten times what you're paying now for the same LLM help and still call it a win, or your coding workflow doesn't really pencil out long term. I still believe that frame. I was a little naive about one variable: time of day.

That post was about price. This one is about when you're allowed to spend what you already bought.

Peak hours eat wall-clock time

Anthropic has been public about tightening how five-hour session limits behave during peak demand. The short version from their side: weekly caps stay the same, but during busy windows your session burns faster than "five hours on the clock" would suggest, especially if you're running token-heavy background work. The Register walked through it with the usual bite.

Anthropic described peak in Pacific time (think early morning through late morning on the west coast). Where I live that's roughly 8:00 AM–2:00 PM Eastern on weekdays: normal business hours, when you're actually trying to get work done. That's when a "five hour" allowance stops feeling like five hours of real use.

So we're not only pricing the future. We're scheduling it.

The cutbacks are spreading

This isn't just one vendor tweaking dials. OpenAI had already been pulling back on things that were expensive to run and light on paying users (hello, Sora-shaped line items). Google has been tightening which premium models free users get in tools like Antigravity. Same pattern: demand up, margin thin, knobs turn.

Meanwhile power isn't getting cheaper. You've seen the photos: diesel banks outside data centers, grid stories that sound like war logistics. Energy is a real input. So is food when the next price wave hits. The squeeze is not theoretical.

When does it stop making sense?

Here's the question I keep coming back to. If the model is great but you can't rely on it when you actually want to work, and the escape hatch is hundreds of dollars a month for "enough" headroom, at what point does that fail the average business user?

Right now people are still vibe-coding ideas as fast as they can type. My own experience says that phase has a shelf life. You burn through the backlog of "what if we tried…" and then you're maintaining. At some velocity, there simply aren't that many net-new features users are asking for. Code changes are mostly bugs (which the models are cheerfully good at creating) and feature requests that slow down once the product is good enough.

Then that $200 tab starts to feel like my Netflix home page. I'm paying. I've already watched everything I cared about. The rows are full; nothing pulls me in. I think code gets there faster than people expect.

Will the whole industry level out on email drafting, content moderation, and blog posts? Is that enough to justify the run rate? I don't have a tidy answer. I'm asking where the curve flattens.

One more worry

Separately: I have a hard time believing all the datacenter buildout happens as drawn on the slide decks. A hundred billion dollars of concrete and GPUs is a fat target before you even get to the political layer. Maybe I'm wrong. It still factors into how durable I think this moment is.

tl;dr: limits are moving from "how much money" to "what time of day" and "which tier." Stack that on power and food and feature saturation, and the cheap-token party starts to look like a scheduling problem with a subscription attached. I'm still using the tools. I'm just updating the model of what "enough" means.