8/10/2025

A Guide to the New GPT-5 Rate Limits: What You Need to Know

Alright, let's talk about GPT-5. If you've been anywhere near the tech world lately, you've probably heard the buzz. OpenAI dropped its latest model, & it’s been a whirlwind of excitement, a little bit of confusion, & a whole lot of questions about what it all means for us, the actual users.
One of the biggest topics of conversation? The new rate limits. Honestly, it's probably the most practical question on everyone's mind. How much can we actually use this thing? Whether you're a free user just messing around, a Plus subscriber who relies on it for daily tasks, or a business building the next big thing on their API, the limits matter. A LOT.
So, I've spent some serious time digging into the official announcements, developer docs, & the chatter from early users to put together a comprehensive guide for you. We'll break down the rate limits for every tier, explore the different flavors of GPT-5 (because it’s not just one model anymore), & talk about what this all means for your workflow & your wallet.

The Big Picture: A Whole New System

First thing you need to understand is that GPT-5 isn't just a simple upgrade from GPT-4o. OpenAI has fundamentally changed how we interact with their models. They've moved away from giving us a dropdown menu of different models like GPT-4o, o3, etc., & instead introduced a "unified system".
Here's the gist: When you type a prompt into ChatGPT, a "real-time router" or "autoswitcher" instantly decides which version of GPT-5 is best for the job. Is it a simple question? You'll get a super-fast response from a zippy, efficient model. Is it a complex coding problem or a request to "think hard about this"? The system will route your request to a more powerful, deeper reasoning model.
The goal was to make it seamless. One model name, no more guessing which one to use. However, the launch was, as OpenAI's CEO Sam Altman put it, a bit "bumpy". The autoswitcher was apparently broken for a bit, which made GPT-5 seem "way dumber" than it actually was. This led to some initial user frustration, which we'll get into, but it's a key piece of context. They're working on making it more transparent which model is answering your query & even giving users more control to manually trigger the "thinking" mode.
Now, let's get into the nitty-gritty of the limits for each user type.

For the Casual User: The Free Tier

If you're using ChatGPT for free, you still get access to the power of GPT-5, which is pretty awesome. But, as you'd expect, there are some pretty tight restrictions.
  • Standard GPT-5: Free users are limited to 10 requests every 5 hours. Another source mentioned 10 messages per hour, so there might be some variation as they fine-tune things.
  • GPT-5 Thinking: You get just one message per day that can access the deeper "Thinking" model.
Once you hit your limit, you're not completely cut off. Your queries will automatically be handled by a "mini" version of the model, which is less powerful but still capable. This is a big change from before, where you'd just have to wait. Now, you get continuous, albeit downgraded, access.
This is a pretty clear strategy to give everyone a taste of GPT-5's capabilities while encouraging heavy users to upgrade. For quick questions, brainstorming, or casual writing help, the free tier is still incredibly useful.

For the Power User: The ChatGPT Plus Subscription

This is where things get really interesting. For $20 a month, ChatGPT Plus users get a SIGNIFICANTLY higher cap, making it a must-have for anyone who integrates AI into their daily workflow.
Here’s the breakdown for Plus users:
  • Standard GPT-5: The initial limit was set at 80 messages every 3 hours.
  • BIG Update: After the initial launch feedback (and a bit of an uproar), Sam Altman announced they would be doubling the rate limits for Plus users to 160 messages per 3 hours, though this change might take a few days to roll out to everyone.
  • GPT-5 Thinking: Plus users also get a separate, generous limit for the more powerful reasoning model: 200 messages per week.
A key benefit for Plus subscribers is the ability to manually select the GPT-5 Thinking model from the picker, giving you direct access to its deeper reasoning capabilities without having to rely on the autoswitcher. When you hit your 3-hour limit on the standard model, you'll also default to the mini version, but your use of the Thinking model is counted against its own weekly limit.
Plus users also get a larger 32,000-token context window, compared to the 8,000-token window for free users. This means you can have much longer, more detailed conversations without the model losing track of what you discussed earlier.
Oh, & in another major concession after the bumpy launch, OpenAI is restoring access to older models like GPT-4o for Plus customers, recognizing that many users had built specific workflows around them.

For the Pros & Teams: The $200 Tier & Beyond

For professionals, startups, & larger companies, there are the ChatGPT Pro ($200/month) & Team ($25/user/month) plans. For these users, the story is much simpler:
  • Unrestricted access to GPT-5 Standard, Thinking, & the exclusive GPT-5 Pro model.
Of course, "unrestricted" is subject to fair use policies & abuse guardrails, but for all practical purposes, these high-tier users don't have to worry about message caps. The Pro tier also gets access to the "GPT-5 Pro" model, which is described as using "parallel test time compute" for even more power, though details on that are still a bit sparse.
This tier is really for those who are building their business processes around AI. Think content creation agencies, marketing teams running massive campaigns, or developers building complex applications.
Speaking of businesses, this is where having a reliable, 24/7 AI presence becomes critical. Many companies are looking at how to leverage this power for their own customer interactions. This is where tools like Arsturn come into the picture. Imagine taking the power of a GPT-level model but training it exclusively on your business data. Arsturn helps businesses create custom AI chatbots that provide instant customer support, answer questions, & engage with website visitors 24/7. It’s like having a super-smart, tireless employee who knows your business inside & out, available on your website all the time.

Under the Hood: The API & The Cost of "Thinking"

For developers & businesses building their own applications, the rate limits are less about message counts & more about tokens & pricing. And here, OpenAI has made some AGGRESSIVELY competitive moves.
First, the API isn't a single "GPT-5". It's broken down into three model sizes to give developers flexibility:
  • gpt-5 (regular)
  • gpt-5-mini
  • gpt-5-nano
Here’s a look at the groundbreaking pricing, per 1 million tokens:
ModelInput PriceOutput Price
GPT-5$1.25$10.00
GPT-5 Mini$0.25$2.00
GPT-5 Nano$0.05$0.40
To put this in perspective, the top-tier GPT-5 model has an input cost that is half of its predecessor, GPT-4o. This is a huge deal & is expected to ignite a price war among AI providers like Google & Anthropic.
But there's a crucial new detail: "invisible reasoning tokens."
When you ask the API to perform a complex task, it might "think" before giving you the final answer. This thinking process generates tokens that aren't part of the final output you see, but you still pay for them. These are counted as output tokens. This means that for complex queries, your output token cost could be higher than you'd expect.
To give developers control over this, OpenAI introduced a few new parameters in the API:
  • 1 reasoning_effort
    : You can set this to
    1 minimal
    ,
    1 low
    ,
    1 medium
    , or
    1 high
    . Setting it to
    1 minimal
    turns off most of the deep reasoning, resulting in a much faster, lower-latency response, which is great for simple tasks like data extraction or formatting.
  • 1 verbosity
    : You can set this to
    1 low
    ,
    1 medium
    , or
    1 high
    to control how detailed the model's replies are without having to tweak your prompt.
There's also a massive 90% discount on cached input tokens, which is a game-changer for chat applications where you're sending the same conversation history over & over again.
The models also boast an impressive input limit of 272,000 tokens & an output limit of 128,000 tokens (which includes those reasoning tokens).

What This Means for Businesses & Developers

The new pricing & model structure has massive implications. The lower input costs make it more economically viable to build applications that process large amounts of text. Think analyzing customer feedback, summarizing long documents, or powering internal knowledge bases.
This is exactly the kind of environment where a platform like Arsturn shines. As businesses look to leverage these powerful new models, they need an easy way to harness them. Arsturn helps businesses build no-code AI chatbots trained on their own data to boost conversions & provide personalized customer experiences. You don't need to be an AI expert to build a bot that can handle lead generation, answer complex product questions, or guide users through your website. It’s about making this advanced tech accessible & actionable for any business. The new, more affordable API pricing makes these kinds of custom solutions even more attractive.
For developers, GPT-5 is being hailed as a phenomenal coding collaborator. It's showing massive improvements on benchmarks like SWE-bench & Aider, meaning it's better at fixing bugs, writing clean code, & even handling end-to-end project builds. It's already being rolled out to paid GitHub Copilot plans, giving developers direct access to its power within their existing workflow. The ability to handle long chains of tool calls without getting lost is a huge step forward for creating more sophisticated AI agents.

So, What's the Bottom Line?

The launch of GPT-5 & its new rate limits is a major milestone. Here are the key takeaways:
  1. It's a System, Not a Model: The new autoswitching router aims to give you the best performance for your specific query, though it had a rocky start.
  2. Generous Limits for Subscribers: ChatGPT Plus users get a substantial number of messages, especially with the doubled rate limits, making the $20/month subscription a high-value proposition for power users.
  3. API Pricing is Disruptive: The aggressive new pricing for API access is set to make powerful AI more accessible than ever, fueling a new wave of innovation.
  4. Control is Key: The new API parameters for
    1 reasoning_effort
    &
    1 verbosity
    give developers fine-grained control over performance, latency, & cost.
  5. Business Applications are Expanding: The combination of lower costs & higher capabilities makes custom AI solutions more practical. For customer engagement, lead generation, & support, building a tailored chatbot is becoming a no-brainer.
The dust is still settling from the GPT-5 launch, & OpenAI is clearly listening to user feedback & making adjustments on the fly. It’s a dynamic and exciting time in the world of AI.
Hope this was helpful in breaking down what all these new changes mean for you. It's a lot to take in, but the big picture is that these powerful tools are becoming more capable & more accessible. Let me know what you think

Copyright © Arsturn 2025