Anthropic prompt caching with cache_control: code samples, up to 90% off cached input tokens, plus OpenAI and Google caching and the edge cases that break it.
Loading full article...