- Input images are also cached
- Used tools (like search or code) are also cached
- If you use structured output schema, that is also cached!
- Cache is cleared from the OpenAI servers after 5-10 minutes of no requests (although it can last up to 1 hour)
- Input images are also cached
- Used tools (like search or code) are also cached
- If you use structured output schema, that is also cached!
- Cache is cleared from the OpenAI servers after 5-10 minutes of no requests (although it can last up to 1 hour)
- Less latency for big prompts (OpenAI says up to 80% less latency!!)
- Reduced costs for the input tokens (since a big part will be cached, OpenAI says up to 50% reduced costs)
- And caching is free, so do it!
- Less latency for big prompts (OpenAI says up to 80% less latency!!)
- Reduced costs for the input tokens (since a big part will be cached, OpenAI says up to 50% reduced costs)
- And caching is free, so do it!
The system prompt will define the basics of every article, so it is my "static" content, it goes at the beginning of the final prompt.
The system prompt will define the basics of every article, so it is my "static" content, it goes at the beginning of the final prompt.
Let me give an example (sad that X does not allow code formatting yet...):
Let me give an example (sad that X does not allow code formatting yet...):