GitHub Copilot can use a variety of AI models. This article explains how these models are hosted and served.
OpenAI models
Used for:
- GPT-4.1
- GPT-5 mini
- GPT-5
- o3
- o4-mini
GPT-4.1 is hosted by GitHub's Azure tenant when used in GitHub Copilot.
GPT-5, GPT-5 mini, o3 and o4-mini models are hosted by OpenAI and GitHub's Azure tenant. OpenAI makes the following data commitment: We [OpenAI] do not train our models on your business data by default. GitHub maintains a zero data retention agreement with OpenAI.
When using OpenAI's models, input requests and output responses continue to run through GitHub Copilot's content filters for public code matching, when applied, along with those for harmful or offensive content.
Anthropic models
Used for:
- Claude Opus 4.1
- Claude Opus 4
- Claude Sonnet 3.5
- Claude Sonnet 3.7
- Claude Sonnet 3.7 Thinking
- Claude Sonnet 4
Claude Opus 4.1 is hosted by Anthropic PBC. Claude Opus 4 and Claude Sonnet 4 are hosted by Anthropic PBC and Google Cloud Platform. Claude Sonnet 3.7 is hosted by Amazon Web Services, Anthropic PBC, and Google Cloud Platform. Claude Sonnet 3.5 is hosted exclusively by Amazon Web Services. GitHub has provider agreements in place to ensure data is not used for training. Additional details for each provider are included below:
- Amazon Bedrock: Amazon makes the following data commitments: Amazon Bedrock doesn't store or log your prompts and completions. Amazon Bedrock doesn't use your prompts and completions to train any AWS models and doesn't distribute them to third parties.
- Anthropic PBC: GitHub maintains a zero data retention agreement with Anthropic.
- Google Cloud: Google commits to not training on GitHub data as part of their service terms. GitHub is additionally not subject to prompt logging for abuse monitoring.
To provide better service quality and reduce latency, GitHub uses prompt caching. You can read more about prompt caching on Anthropic PBC, Amazon Bedrock, and Google Cloud.
When using Claude, input prompts and output completions continue to run through GitHub Copilot's content filters for public code matching, when applied, along with those for harmful or offensive content.
Google models
Used for:
- Gemini 2.5 Pro
- Gemini 2.0 Flash
GitHub Copilot uses Gemini 2.0 Flash and Gemini 2.5 Pro hosted on Google Cloud Platform (GCP). When using Gemini models, prompts and metadata are sent to GCP, which makes the following data commitment: Gemini doesn't use your prompts, or its responses, as data to train its models.
To provide better service quality and reduce latency, GitHub uses prompt caching.
When using Gemini models, input prompts and output completions continue to run through GitHub Copilot's content filters for public code matching, when applied, along with those for harmful or offensive content.
xAI models
Complimentary access for Grok Code Fast 1 is available until 2 p.m. PDT on Tuesday, September 2, 2025. Regular pricing applies after that point.
Grok Code Fast 1 is hosted by xAI when used in GitHub Copilot.
xAI operates Grok Code Fast 1 in GitHub Copilot under a zero data retention API policy. This means xAI commits that user content (both inputs sent to the model and outputs generated by the model):
Will not be:
- Logged for any purpose, including human review
- Saved to disk or retained in any form, including as metadata
- Accessible by xAI personnel
Will only:
- Exist temporarily in RAM for the minimum time required to process and respond to each request
- Be immediately deleted from memory once the response is delivered
When using xAI, input prompts and output completions continue to run through GitHub Copilot's content filters for public code matching, when applied, along with those for harmful or offensive content.