GitHub Copilot bring your own key (BYOK) enhancements
Source: GitHub Changelog
What’s new
New provider options
You can now connect API keys from AWS Bedrock, Google AI Studio, and any OpenAI‑compatible provider. These options join Anthropic, Microsoft Foundry, OpenAI, and xAI as supported BYOK choices.
Support for the Responses API
BYOK now supports models that use the Responses API. This enables structured outputs and richer multimodal interactions.
Ability to set a maximum context window
Admins can define the maximum context window for BYOK models, helping balance cost, performance, and response quality.
Streaming responses for faster interaction
Streaming output from connected models lets Copilot display responses as they are generated, rather than waiting for completion.
Start using the new BYOK features today
These new capabilities are available now in public preview for GitHub Enterprise and Business customers. Connect your LLM provider’s API key in your enterprise or organization settings and start using your models in Copilot Chat and supported IDEs. Learn more by visiting our BYOK documentation.
Help us shape the future
We are just getting started and your feedback will guide what comes next. Join the discussion within the GitHub Community to share feedback and connect with other developers.