Home
/
Latest news
/
Research developments
/

Investigating open ai's prompt caching: key findings

OpenAI's Prompt Caching Behavior | Surprising Findings Across Model Generations

By

Fatima El-Hawari

Nov 28, 2025, 02:43 AM

3 minutes needed to read

A visual representation showing data flow and efficiency in OpenAI's prompt caching process with model generations.

A developer recently tested OpenAI's prompt caching mechanism across its latest AI models and uncovered undocumented behaviors. The results reveal significant differences in cache performance that could impact cost and efficiency for businesses using AI for various applications.

Insights from the Test

The developer created a chatbot to monitor network devices, integrating 10 tools and utilizing a system prompt of about 1,400 tokens. They ran tests on three model generations: gpt-4o-mini, gpt-5-mini, and gpt-5. During testing, they logged essential metrics like prompt_tokens, cached_tokens, latency, and cost per call.

Key Findings

  1. Caching Functionality: The caching feature performs effectively once the prefix exceeds 1,024 tokens. The gpt-4o-mini achieved an 80% cache hit rate, while both gpt-5-mini and gpt-5 scored 90%.

    "Cache warming provides a significant advantage for multi-model operations."

    • Developer Insight

  2. Compressed Tool Definitions: When the developer added additional tool definitions, the increase in token count was minimalโ€”only 56 tokens instead of the expected 400-500. This indicates that OpenAI has optimized the structure heavily.

  3. Cross-Model Cache Sharing: In an unexpected turn, the developer confirmed that cache information was shared across model generations. For instance, after a cold start with gpt-4o-mini, the gpt-5-mini successfully hit the cache on its first call. This behavior isn't documented anywhere by OpenAI, presenting a tactical advantage for users.

    • "The ability to warm caches across models can save significant costs, especially for businesses managing numerous cold starts."

    • Developer Analysis

Practical Implications

This behavior can lead to substantial savings. For example, running a primary model with 10,000 tokens in daily usage could cost businesses $4,562 annually without cache-sharing. However, using the less expensive gpt-5-mini for initial calls could reduce that cost to $182, resulting in savings of over $4,380 a year.

Community Reactions

The developer's findings sparked conversations in forums.

  • Positive Sentiment: Many users expressed interest in the cost-saving aspect and the hidden functionalities of OpenAI's models.

  • Mixed Views: Some participants found the undocumented nature of the behavior concerning, emphasizing the need for clearer documentation.

  • Curious Observations: A few users remarked on how these findings could change the approach to building AI-based applications, allowing for a more intelligent and cost-effective use of resources.

Key Takeaways

  • ๐Ÿš€ Caching Efficiency: 80-90% hit rate with reduced costs.

  • ๐Ÿ“‰ Potential Savings: Implementing multi-model caching can save thousands per year.

  • โ— Need for Documentation: Undocumented features raise concerns about transparency in AI models.

Predictions on Future Usage

There's a strong possibility that more businesses will adopt multi-model caching strategies as they learn about these findings. Experts estimate that companies looking to optimize costs and performance will likely implement this practice within the next year. As word spreads through forums and discussions, we might see a shift in how developers approach AI model usage, with significant interest in exploring undocumented features, leading to further scrutiny of OpenAIโ€™s practices. The transparency of these systems may also prompt a call for better documentation from OpenAI, with an estimated 70% of developers considering clearer guidelines pivotal for trust and adoption moving forward.

A Historical Echo of Transformation

In the early 2000s, the introduction of open-source software significantly altered tech development. Just like today's AI landscape, developers saw a new path for innovation and collaboration, largely due to hidden efficiencies. Millions took the plunge into creating adaptable software models, similar to the way developers are now learning to optimize AI models. The unexpected sharing of resources mirrored how coders leveraged community assets for broader access and performance. Itโ€™s a reminder that when resources are unveiled, they can ignite a revolution in interconnected thinking and practice that reshapes industries entirely.