- ChatGPT-5’s increased token usage raises costs for publishers
- Different models needed for various publishing tasks
- Speed and efficiency now key to successful AI integration in media
When ChatGPT-5 launched, I was ready to double down on my loyalty to OpenAI. Tokens looked cheaper on paper, and, as a publisher-facing tech business, NoahWire consumes AI at scale to create, curate and distribute content. Every improvement in the models means an improvement in what we can deliver. But within weeks that optimism turned into hard questions about cost, speed and reliability.
On launch day, Sam Altman, OpenAI’s ceo, proudly said API usage had “doubled overnight”. That puzzled me. Publishers, for example, don’t suddenly double the number of calls they make to an API just because a model gets smarter – their daily requirements are tied to editorial workflows.
What I soon realised was that while usage hadn’t changed, costs had. Token burn shot up. Even Altman, who I admire and often think of as the closest thing we have to a Messiah in this industry, seemed to have missed this. He believed people were simply using the system more, when in fact the model itself was consuming more tokens to do the same work.
Paying more for the same job
GPT-5 tokens may be cheaper per unit, but the model is verbose and “over thinks” even simple editorial tasks. A summarisation or headline job that GPT-4 completed neatly now consumes far more tokens. Worse, it runs slower.
For publishers who integrate AI into live pipelines, where feeds, newsletters, CMS integrations and alerts demand split-second accuracy, unpredictability is as damaging as under-performance. Even “over-performance” can cause problems when you’re building precisely tuned stacks.
Then there are the online costs. Models like Gemini tempt publishers with a free allowance, but once you’re running tens of thousands of queries a day – the norm for live feeds – the bills explode. OpenAI’s trick is subtler: GPT-5 looks cheaper per unit, but by bloating responses and tokens it drives up the true cost in production. Hidden charges like these can quietly wreck a publishing P&L.
Rethinking model choices
This has pushed us, and the publishers we work with, to rethink how tasks are assigned across models:
* ChatGPT-5 – Still the best for polished, multilingual long-form drafts. Excellent for features or in-depth explainers, but too slow and costly for daily pipelines
* GPT-4 – Cheaper, faster and more predictable. Less “brilliant” than GPT-5, but a reliable workhorse for summaries, headlines and bulletins
- Gemini – Strong technically, but the cost of online calls makes it impractical for publishers running live news or alerts
- DeepSeek – Initially impressive, but quickly became unreliable in tests, drifting mid-task. There’s also unease about how it sourced its training data. I invest in China myself, but in AI publishing it can feel like swimming in shark-infested waters.
What this means for publishers
ChatGPT-5 was supposed to simplify publishing operations. Instead, it has pushed publishers – and many others beyond publishing – into becoming experts in model economics, workflow design and cost optimisation. Entire weeks have been lost re-engineering systems simply to stabilise them.
Perhaps this is the new reality of AI publishing. The winners won’t be those who pledge loyalty to one model, but those who treat AI as a toolkit, choosing the right model for each editorial job and constantly auditing cost against output.
For publishers, that means high-quality features may still be written with ChatGPT-5, but live newsrooms, newsletters and bulletins demand the speed and economy of simpler models. The true cost of AI publishing isn’t just about tokens – it’s about online calls, latency and efficiency. Those who grasp this will publish faster, cheaper and better. Those who don’t will keep getting caught out.
Ivan Massow is founder and ceo of NoahWire
- https://www.tomsguide.com/ai/what-is-chat-gpt-5 – This article provides an overview of ChatGPT-5’s features, including its adaptability, reasoning capabilities, and user experience enhancements, corroborating the claim that improvements in the models lead to better content delivery for publishers.
- https://www.tomsguide.com/news/live/openai-chatgpt-5-live-blog – This source details the launch of ChatGPT-5, highlighting its advancements in intelligence, safety, reasoning, and personalization, supporting the assertion that the model’s improvements impact publishers’ operations.
- https://www.tomsguide.com/ai/chatgpt/i-tested-chatgpt-vs-gemini-2-5-pro-with-these-3-prompts-and-it-shows-what-gpt-5-needs-to-do – This article compares ChatGPT-5 with Gemini 2.5 Pro, illustrating areas where GPT-5 needs improvement, aligning with the observation that the model’s verbosity and overthinking can lead to increased token consumption and slower performance.
- https://www.tomsguide.com/ai/chatgpt/i-tested-chatgpt-vs-gemini-2-5-pro-with-these-3-prompts-and-it-shows-what-gpt-5-needs-to-do – This source discusses the impact of GPT-5’s performance on publishers, noting that while the model offers advanced capabilities, it may not be ideal for daily pipelines due to its slower response times and higher token usage.
- https://www.tomsguide.com/ai/chatgpt/i-tested-chatgpt-vs-gemini-2-5-pro-with-these-3-prompts-and-it-shows-what-gpt-5-needs-to-do – This article highlights the challenges publishers face with GPT-5’s increased token consumption and slower performance, emphasizing the need for careful consideration of model choices in live newsrooms and alert systems.
- https://www.tomsguide.com/ai/chatgpt/i-tested-chatgpt-vs-gemini-2-5-pro-with-these-3-prompts-and-it-shows-what-gpt-5-needs-to-do – This source underscores the importance of understanding the true costs of AI publishing, including online calls, latency, and efficiency, to ensure faster, cheaper, and better publishing outcomes.


