The Token Optimizer: Automating Prompt Caching Breakpoints in Python Microservices to Slash LLM Costs