Handling LLM Rate Limits
Rate limiting is the process of restricting the number of requests a user or application can send to an LLM API within a specific timeframe. LLM providers enforce this with the purpose of managing resources and preventing abuse.
Since Goose is working very quickly to implement your tasks, you may need to manage rate limits imposed by the provider. If you frequently hit rate limits, consider upgrading your LLM plan to access higher tier limits or using OpenRouter.
Using OpenRouter
OpenRouter provides a unified interface for LLMs that allows you to select and switch between different providers automatically - all under a single billing plan. With OpenRouter, you can utilize free models or purchase credits for paid models.
- Go to openrouter.ai and create an account.
- Once verified, create your API key.
- Goose CLI
- Goose Desktop
- Run the Goose configuration command:
goose configure
- Select
Configure Providers
from the menu. - Follow the prompts to choose OpenRouter as your provider and enter your OpenRouter API key when prompted.
- Click on the three dots in the top-right corner.
- Select
Settings
from the menu. - Click on "Browse" in the
Models
section. - Click on
Configure
- Select
OpenRouter
from the list of available providers. - Enter your OpenRouter API key in the dialog that appears.
Now Goose will send your requests through OpenRouter which will automatically switch models when necessary to avoid interruptions due to rate limiting.