Skip to content

AI Gateway rate limits

LLM providers apply limits to concurrency, requests per minute, input tokens per minute, and output tokens per minute, which are configured in the AI Gateway (see the Inference limits documentation)

If our usage is approaching or exceeding these limits, we may need to request an increase. See provider-specific instructions below.

You can see the current limits set for the GitLab account in https://console.anthropic.com/settings/limits.

To request a rate limit increase, contact the Anthropic team via the #ext-anthropic channel.

If you do not have access to the GitLab Anthropic account, please file an Access Request.