Groq
inference_host
79 signals tracked
OpenAI Adds Citation and Reasoning Options for Chat Completions
Added optional citation_options and include_reasoning parameters to provide more contextual control over model responses, including citation inclusion and reasoning visibility.
Date not specified
MediumCapabilityAPI Parameter Deprecations: function_call, max_tokens, etc. require migration
Multiple API parameters deprecated including: function_call, functions, max_tokens, exclude_domains, include_domains. Developers should migrate to new alternatives like tools, max_completion_tokens, and search_settings.
Date not specified
HighDeprecationGroqCloud Status provides real-time system health monitoring
Date not specified
InfoCapabilityGroq adds citation options for chat completion
Groq now supports optional citation configuration via citation_options parameter. Can be set to 'enabled' or 'disabled', allowing more control over information retrieval and citation in responses.
Date not specified
MediumCapabilityGroq Chat Completion: Deprecation of API Parameters - Update to 'tools' and 'max_completion_tokens'
Several API parameters are now deprecated, including: function_call, functions, max_tokens, exclude_domains, include_domains. Developers should update to newer equivalents like tools, max_completion_tokens, and search_settings.
Date not specified
HighDeprecationGroqCloud Status: Real-time System Health Monitoring
Date not specified
InfoCapabilityNew Chat Completion Configuration Options Added
New parameters introduced including citation_options, compound_custom, tool_choice, documents, and include_reasoning, expanding configuration flexibility for chat interactions.
Date not specified
MediumCapabilityGroq Chat Completions: Deprecation of function_call, functions, max_tokens, exclude_domains, include_domains
Deprecated parameters include: function_call, functions, max_tokens, exclude_domains, include_domains. Developers should migrate to newer equivalents like tools, max_completion_tokens, and search_settings.
Date not specified
HighDeprecationGroq adds automatic prompt caching for openai/gpt-oss-120b
Groq has introduced automatic prompt caching for the openai/gpt-oss-120b model, providing 50% cost savings on cached input tokens, lower latency, and higher effective rate limits. The feature requires zero setup and automatically activates when requests share common prefixes with recent requests. Users will automatically benefit from this optimization without any required action.
Date not specified
MediumCapabilityGroq releases GPT-OSS-Safeguard 20B - open weight reasoning model
Groq launched GPT-OSS-Safeguard 20B, their first open weight reasoning model designed for Trust & Safety content moderation tasks. The model features 131K token context window, prompt caching for cost savings, and supports customizable policy-based classification. This enables organizations to implement bring-your-own-policy content moderation with structured reasoning capabilities.
Date not specified
MediumCapabilityGroq SDK updates: Prompt caching and citation support added
Python SDK updated to v0.33.0 and TypeScript SDK to v0.34.0 with enhanced prompt caching and new annotation/citation features for chat completions. These are additive improvements that don't require immediate action but may benefit applications using chat functionality.
Date not specified
MediumCapabilityGroq adds Google Workspace integration via MCP Connectors
Groq introduced pre-built MCP Connectors that provide zero-configuration integration with Google Workspace services including Gmail, Google Calendar, and Google Drive. The connectors use OAuth 2.0 authentication and are compatible with existing OpenAI Responses API workflows. This is a new feature addition that enhances integration capabilities without requiring any changes to existing implementations.
Date not specified
MediumCapabilityGroq releases GPT-OSS-Safeguard 20B - new open weight reasoning model
Groq launched GPT-OSS-Safeguard 20B, their first open weight reasoning model designed for Trust & Safety content moderation tasks. The model features a 131K token context window, custom policy support, and structured reasoning capabilities for automated content classification. This is a new feature release that expands Groq's model offerings for safety applications.
Date not specified
MediumCapabilityGroq introduces automatic prompt caching for gpt-oss-120b
Groq has introduced automatic prompt caching for the openai/gpt-oss-120b model, providing 50% cost reduction on cached input tokens and improved performance through lower latency and higher effective rate limits. The feature requires zero setup and automatically activates when requests share common prefixes with recent requests. This is a beneficial enhancement that users can immediately leverage without any configuration changes.
Date not specified
MediumCapabilityGroq SDK updates: Prompt caching and annotation support added
Python SDK updated to v0.33.0 and TypeScript SDK to v0.34.0 with enhanced prompt caching capabilities and new annotation/citation support for chat completion messages. These are additive features that improve functionality without requiring immediate action from existing users.
Date not specified
MediumCapabilityGroq adds automatic prompt caching to openai/gpt-oss-20b
Groq has introduced automatic prompt caching for the openai/gpt-oss-20b model, providing 50% cost reduction on cached input tokens and improved response times through computation reuse. The feature requires zero setup and automatically activates when requests share common prefixes with recent requests. Users will automatically benefit from this optimization without any required action.
Date not specified
MediumCapabilityGroq adds Remote Model Context Protocol (MCP) server integration in Beta
Groq has launched Beta support for Remote MCP server integration on GroqCloud, enabling AI models to connect to thousands of external tools through Anthropic's open MCP standard. The implementation is fully compatible with OpenAI's APIs, allowing developers to migrate from OpenAI to Groq without code changes while benefiting from faster execution and lower costs. This is a new feature addition that enhances existing capabilities without requiring immediate action.
Date not specified
InfoCapabilityGroq adds Kimi K2-0905 model with 256K context window
Groq has launched Moonshot AI's Kimi K2-0905 model on GroqCloud, featuring the largest context window available (256K tokens) and prompt caching capabilities for up to 50% cost savings. The model offers enhanced agentic coding capabilities and improved frontend development performance at competitive pricing of $1.50/M tokens blended. This is a new model addition that expands available capabilities without requiring any changes to existing implementations.
Date not specified
MediumCapabilityGroq launches Compound model with enhanced accuracy and agentic tools
Groq released their new Compound model built on GPT-OSS-120B and Llama, delivering significantly improved performance with built-in server-side tools including web search, code execution, and browser automation. The model achieves general availability with increased rate limits and outperforms competing systems on key benchmarks. No immediate action required - this is a new capability enhancement for users.
Date not specified
MediumCapabilityGroq SDK updated with OpenAI compatibility and Compound tools
Python SDK updated to v0.31.1 and TypeScript SDK to v0.32.0 with better OpenAI message type compatibility and bug fixes. Added support for new Groq Compound tools including Wolfram Alpha, Browser Automation, and Visit Website functionality. These are enhancement updates that improve functionality without requiring immediate action.
Date not specified
MediumCapability
Get alerts for Groq
Never miss a breaking change. SignalBreak monitors Groq and dozens of other AI providers in real time.
Sign up free — no credit card required