Skip to main content

Supported LLM Providers

Goose is compatible with a wide range of LLM providers, allowing you to choose and integrate your preferred model.

Model Selection

Goose relies heavily on tool calling capabilities and currently works best with Anthropic's Claude 3.5 Sonnet and OpenAI's GPT-4o (2024-11-20) model. Berkeley Function-Calling Leaderboard can be a good guide for selecting models.

Available Providers

ProviderDescriptionParameters
Amazon BedrockOffers a variety of foundation models, including Claude, Jurassic-2, and others. AWS environment variables must be set in advance, not configured through goose configureAWS_PROFILE, or AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY, AWS_REGION, ...
Amazon SageMaker TGIRun Text Generation Inference models through Amazon SageMaker endpoints. AWS credentials must be configured in advance.SAGEMAKER_ENDPOINT_NAME, AWS_REGION (optional), AWS_PROFILE (optional)
AnthropicOffers Claude, an advanced AI model for natural language tasks.ANTHROPIC_API_KEY, ANTHROPIC_HOST (optional)
Azure OpenAIAccess Azure-hosted OpenAI models, including GPT-4 and GPT-3.5. Supports both API key and Azure credential chain authentication.AZURE_OPENAI_ENDPOINT, AZURE_OPENAI_DEPLOYMENT_NAME, AZURE_OPENAI_API_KEY (optional)
DatabricksUnified data analytics and AI platform for building and deploying models.DATABRICKS_HOST, DATABRICKS_TOKEN
Docker Model RunnerLocal models running in Docker Desktop or Docker CE with OpenAI-compatible API endpoints. Because this provider runs locally, you must first download a model.OPENAI_HOST, OPENAI_BASE_PATH
GeminiAdvanced LLMs by Google with multimodal capabilities (text, images).GOOGLE_API_KEY
GCP Vertex AIGoogle Cloud's Vertex AI platform, supporting Gemini and Claude models. Credentials must be configured in advance.GCP_PROJECT_ID, GCP_LOCATION and optionally GCP_MAX_RATE_LIMIT_RETRIES (5), GCP_MAX_OVERLOADED_RETRIES (5), GCP_INITIAL_RETRY_INTERVAL_MS (5000), GCP_BACKOFF_MULTIPLIER (2.0), GCP_MAX_RETRY_INTERVAL_MS (320_000).
GitHub CopilotAccess to GitHub Copilot's chat models including gpt-4o, o1, o3-mini, and Claude models. Uses device code authentication flow for secure access.Uses GitHub device code authentication flow (no API key needed)
GroqHigh-performance inference hardware and tools for LLMs.GROQ_API_KEY
OllamaLocal model runner supporting Qwen, Llama, DeepSeek, and other open-source models. Because this provider runs locally, you must first download and run a model.OLLAMA_HOST
RamalamaLocal model using native OCI container runtimes, CNCF tools, and supporting models as OCI artifacts. Ramalama API an compatible alternative to Ollama and can be used with the Goose Ollama provider. Supports Qwen, Llama, DeepSeek, and other open-source models. Because this provider runs locally, you must first download and run a model.OLLAMA_HOST
OpenAIProvides gpt-4o, o1, and other advanced language models. Also supports OpenAI-compatible endpoints (e.g., self-hosted LLaMA, vLLM, KServe). o1-mini and o1-preview are not supported because Goose uses tool calling.OPENAI_API_KEY, OPENAI_HOST (optional), OPENAI_ORGANIZATION (optional), OPENAI_PROJECT (optional), OPENAI_CUSTOM_HEADERS (optional)
OpenRouterAPI gateway for unified access to various models with features like rate-limiting management.OPENROUTER_API_KEY
SnowflakeAccess the latest models using Snowflake Cortex services, including Claude models. Requires a Snowflake account and programmatic access token (PAT).SNOWFLAKE_HOST, SNOWFLAKE_TOKEN
Venice AIProvides access to open source models like Llama, Mistral, and Qwen while prioritizing user privacy. Requires an account and an API key.VENICE_API_KEY, VENICE_HOST (optional), VENICE_BASE_PATH (optional), VENICE_MODELS_PATH (optional)
xAIAccess to xAI's Grok models including grok-3, grok-3-mini, and grok-3-fast with 131,072 token context window.XAI_API_KEY, XAI_HOST (optional)

CLI Providers

Goose also supports special "pass-through" providers that work with existing CLI tools, allowing you to use your subscriptions instead of paying per token:

ProviderDescriptionRequirements
Claude Code (claude-code)Uses Anthropic's Claude CLI tool with your Claude Code subscription. Provides access to Claude with 200K context limit.Claude CLI installed and authenticated, active Claude Code subscription
Gemini CLI (gemini-cli)Uses Google's Gemini CLI tool with your Google AI subscription. Provides access to Gemini with 1M context limit.Gemini CLI installed and authenticated
CLI Providers

CLI providers are cost-effective alternatives that use your existing subscriptions. They work differently from API providers as they execute CLI commands and integrate with the tools' native capabilities. See the CLI Providers guide for detailed setup instructions.

Configure Provider

To configure your chosen provider or see available options, run goose configure in the CLI or visit the Settings page in the Goose Desktop.

To update your LLM provider and API key:

  1. Click the button in the top-left to open the sidebar
  2. Click the Settings button on the sidebar
  3. Click the Models tab
  4. Click Configure Providers
  5. Click Configure on the LLM provider to update
  6. Add additional configurations (API key, host, etc) then press submit

To change provider model

  1. Click the button in the top-left to open the sidebar
  2. Click the Settings button on the sidebar
  3. Click the Models tab
  4. Click Switch models
  5. Select a Provider from drop down menu
  6. Select a model from drop down menu
  7. Press Select Model

You can explore more models by selecting a provider name under Browse by Provider. A link will appear, directing you to the provider's website. Once you've found the model you want, return to step 6 and paste the model name.

Using Custom OpenAI Endpoints

Goose supports using custom OpenAI-compatible endpoints, which is particularly useful for:

  • Self-hosted LLMs (e.g., LLaMA, Mistral) using vLLM or KServe
  • Private OpenAI-compatible API servers
  • Enterprise deployments requiring data governance and security compliance
  • OpenAI API proxies or gateways

Configuration Parameters

ParameterRequiredDescription
OPENAI_API_KEYYesAuthentication key for the API
OPENAI_HOSTNoCustom endpoint URL (defaults to api.openai.com)
OPENAI_ORGANIZATIONNoOrganization ID for usage tracking and governance
OPENAI_PROJECTNoProject identifier for resource management
OPENAI_CUSTOM_HEADERSNoAdditional headers to include in the request. Can be set via environment variable, configuration file, or CLI, in the format HEADER_A=VALUE_A,HEADER_B=VALUE_B.

Example Configurations

If you're running LLaMA or other models using vLLM with OpenAI compatibility:

OPENAI_HOST=https://your-vllm-endpoint.internal
OPENAI_API_KEY=your-internal-api-key

Setup Instructions

  1. Click the button in the top-left to open the sidebar
  2. Click the Settings button on the sidebar
  3. Next to Models, click the browse link
  4. Click the configure link in the upper right corner
  5. Press the + button next to OpenAI
  6. Fill in your configuration details:
    • API Key (required)
    • Host URL (for custom endpoints)
    • Organization ID (for usage tracking)
    • Project (for resource management)
  7. Press submit
Enterprise Deployment

For enterprise deployments, you can pre-configure these values using environment variables or configuration files to ensure consistent governance across your organization.

Using Goose for Free

Goose is a free and open source AI agent that you can start using right away, but not all supported LLM Providers provide a free tier.

Below, we outline a couple of free options and how to get started with them.

Limitations

These free options are a great way to get started with Goose and explore its capabilities. However, you may need to upgrade your LLM for better performance.

Groq

Groq provides free access to open source models with high-speed inference. To use Groq with Goose, you need an API key from Groq Console.

Groq offers several open source models that support tool calling:

  • moonshotai/kimi-k2-instruct - Mixture-of-Experts model with 1 trillion parameters, optimized for agentic intelligence and tool use
  • qwen/qwen3-32b - 32.8 billion parameter model with advanced reasoning and multilingual capabilities
  • gemma2-9b-it - Google's Gemma 2 model with instruction tuning
  • llama-3.3-70b-versatile - Meta's Llama 3.3 model for versatile applications

To set up Groq with Goose, follow these steps:

To update your LLM provider and API key:

  1. Click the button in the top-left to open the sidebar.
  2. Click the Settings button on the sidebar.
  3. Click the Models tab.
  4. Click Configure Providers
  5. Choose Groq as provider from the list.
  6. Click Configure, enter your API key, and click Submit.

Google Gemini

Google Gemini provides a free tier. To start using the Gemini API with Goose, you need an API Key from Google AI studio.

To set up Google Gemini with Goose, follow these steps:

To update your LLM provider and API key:

  1. Click the button in the top-left to open the sidebar.
  2. Click the Settings button on the sidebar.
  3. Click the Models tab.
  4. Click Configure Providers
  5. Choose Google Gemini as provider from the list.
  6. Click Configure, enter your API key, and click Submit.

Local LLMs

Goose is a local AI agent, and by using a local LLM, you keep your data private, maintain full control over your environment, and can work entirely offline without relying on cloud access. However, please note that local LLMs require a bit more set up before you can use one of them with Goose.

Limited Support for models without tool calling

Goose extensively uses tool calling, so models without it can only do chat completion. If using models without tool calling, all Goose extensions must be disabled.

Here are some local providers we support:

  1. Download Ollama.
  2. In a terminal, run any model supporting tool-calling

Example:

ollama run qwen2.5
  1. In a separate terminal window, configure with Goose:
goose configure
  1. Choose to Configure Providers
┌   goose-configure 

◆ What would you like to configure?
│ ● Configure Providers (Change provider or update credentials)
│ ○ Toggle Extensions
│ ○ Add Extension

  1. Choose Ollama as the model provider
┌   goose-configure 

◇ What would you like to configure?
│ Configure Providers

◆ Which model provider should we use?
│ ○ Anthropic
│ ○ Databricks
│ ○ Google Gemini
│ ○ Groq
│ ● Ollama (Local open source models)
│ ○ OpenAI
│ ○ OpenRouter

  1. Enter the host where your model is running
Endpoint

For Ollama, if you don't provide a host, we set it to localhost:11434. When constructing the URL, we prepend http:// if the scheme is not http or https. If you're running Ollama on a different server, you'll have to set OLLAMA_HOST=http://{host}:{port}.

┌   goose-configure 

◇ What would you like to configure?
│ Configure Providers

◇ Which model provider should we use?
│ Ollama

◆ Provider Ollama requires OLLAMA_HOST, please enter a value
│ http://localhost:11434

  1. Enter the model you have running
┌   goose-configure 

◇ What would you like to configure?
│ Configure Providers

◇ Which model provider should we use?
│ Ollama

◇ Provider Ollama requires OLLAMA_HOST, please enter a value
│ http://localhost:11434

◇ Enter a model from that provider:
│ qwen2.5

◇ Welcome! You're all set to explore and utilize my capabilities. Let's get started on solving your problems together!

└ Configuration saved successfully
Context Length

If you notice that Goose is having trouble using extensions or is ignoring .goosehints, it is likely that the model's default context length of 4096 tokens is too low. Set the OLLAMA_CONTEXT_LENGTH environment variable to a higher value.

Azure OpenAI Credential Chain

Goose supports two authentication methods for Azure OpenAI:

  1. API Key Authentication - Uses the AZURE_OPENAI_API_KEY for direct authentication
  2. Azure Credential Chain - Uses Azure CLI credentials automatically without requiring an API key

To use the Azure Credential Chain:

  • Ensure you're logged in with az login
  • Have appropriate Azure role assignments for the Azure OpenAI service
  • Configure with goose configure and select Azure OpenAI, leaving the API key field empty

This method simplifies authentication and enhances security for enterprise environments.


If you have any questions or need help with a specific provider, feel free to reach out to us on Discord or on the Goose repo.