Groq introduces automatic prompt caching for gpt-oss-120b
AI Impact Summary
Groq has released automatic prompt caching for the openai/gpt-oss-120b model, offering significant cost savings (50%) and reduced latency. This capability automatically optimizes prompt reuse, eliminating the need for manual configuration and improving performance without requiring any code changes. This enhancement allows users to immediately benefit from reduced operational costs and faster response times.
Affected Systems
- Date
- Date not specified
- Change type
- capability
- Severity
- medium