“Model Not Supported” (400 Error)
Why this happens: The model you’re trying to use hasn’t been enabled for your organization or tool setup.
How to fix:
Log in to the dashboard
Go to your Unbound dashboard.
Open the Models page
Click Models in the left sidebar.
Check model availability
Verify the model you want is listed and available for your organization.
Contact your admin
If the model is listed but you still get the error, contact your organization administrator — they may need to enable it in your configuration.
Only organization administrators can enable or disable models. If you’re a regular member, ask your admin to make changes.
How to Switch Models
Different tools have different model switching mechanisms:
- Unbound Chat — Use the model selector dropdown at the top of the chat interface
- Claude Code CLI — The model is determined by your Unbound configuration. To request a different default model, contact your admin
- Roo Code — Open Roo Code Settings (
Ctrl/Cmd + Shift + P → “Roo Code: Open Settings”) and select a different model from the dropdown
- Cursor — Model selection is configured within Cursor’s AI settings panel
Requesting New Models (e.g., Opus 4.6, GPT-5)
New models must be:
- Added to Unbound’s platform by the Unbound team (if not yet supported)
- Enabled for your organization by your admin
If a model you need isn’t available in the Models page:
- Contact support to check if it’s on the roadmap
- If it’s a recently released model, it may take a few days to be added
Token Limit Exceeded Errors
”Claude’s Response Exceeded the Output Token Maximum”
This error occurs when Claude generates a response larger than the configured maximum output tokens.
Fix for Claude Code CLI Users
Set the environment variable to increase the limit:
export CLAUDE_CODE_MAX_OUTPUT_TOKENS=64000
Add this to your shell profile (~/.bashrc or ~/.zshrc) and restart your terminal.
Fix for API Users
Include the max_tokens parameter in your API request with a higher value:
response = client.messages.create(
model="claude-opus-4-5-20250514",
max_tokens=8192, # Increase this value
messages=[{"role": "user", "content": "Your prompt"}]
)
General Tips to Avoid Token Limits
- Break large tasks into smaller prompts — instead of “rewrite the entire file”, ask for one function at a time
- Ask for concise responses — add “be brief” or “concise output” to your prompts
- For code generation — ask for one file at a time instead of entire projects
- Use streaming if your tool supports it — this allows partial responses even if the total would exceed limits
Overloaded Errors (529)
Error: 529 {"type":"error","error":{"type":"overloaded_error","message":"Overloaded"}}
This is a temporary upstream error from the AI provider (e.g., Anthropic). It means the model is experiencing high traffic.
How to handle:
- Wait a few minutes and retry — these are usually brief
- Try a different model — switch to an alternative (e.g., from Claude Opus to Claude Sonnet)
- If persistent, contact support
Still Having Issues?
Contact support with:
- The exact error message (screenshot preferred)
- The model name you’re trying to use
- Which tool you’re using (Claude Code, Roo Code, Cursor, Chat, API)