Control your Generative AI costs with the Vertex API’s context caching

Me and my colleague Nim Jayawardena wrote a post about use of context caching when working with Gen AI models. The post explores how to use Vertex AI’s context caching to reduce the cost of using Gemini models with large, repeated contexts. It shows criteria for using context caching together with the code samples. Since the platform does not support posting content with multiple authors, you can find the post in leoy.blog or in the Nim’s Medium blog.

1 Like