Supported Models#
Lens Agents is designed to be model-agnostic. The platform provides tools, connectivity, and governance; the model provides intelligence. The set of integrated providers expands as customer engagements require.
Currently integrated providers#
As of the current release:
| Provider | Models | How integration works |
|---|---|---|
| Anthropic | Claude Opus, Sonnet, Haiku | Direct Anthropic API |
| AWS Bedrock | Claude, Llama, Mistral, and other Bedrock-hosted models | Regional deployment, IAM-based access |
Additional providers — including Azure OpenAI, OpenAI direct, Google, and self-hosted endpoints — are configured to customer requirements during evaluation engagements. Tell us during onboarding which providers your deployment requires.
How models work by agent type#
Desktop AI tools#
Desktop tools like Claude Desktop, Cursor, ChatGPT, and Copilot use their own built-in model. Lens Agents has no involvement in the model layer — it provides tools and connectivity via MCP. The user's existing subscription or API key powers the intelligence. This means desktop tools can use any model their vendor supports, independently of what Lens Agents integrates directly.
External agents#
External agents (LangChain, CrewAI, Claude Agent SDK, custom frameworks) bring their own model configuration. The agent developer chooses the provider, model, and parameters. Lens Agents provides the tool execution environment through MCP — the model choice is outside the platform's scope.
Managed agents#
Managed agents route model requests through the platform's LLM proxy. The proxy supports the providers listed above and provides:
- Usage extraction — token counts (input, output, cache read, cache write) and cost per request
- Spending enforcement — budget checked before every LLM call. Exceeded budgets return a 429 error.
- Audit logging — every model request is recorded with model name, token usage, cost, and prompt metadata
- Prompt caching — stable context sections are cached across invocations to reduce latency and cost
The default model for managed agents is Claude (via AWS Bedrock). The classification model for alert suppression and channel routing is Claude Haiku.
Spending controls#
Spending controls apply to managed agent LLM usage. They do not apply to desktop or external agents, because those agents use their own model provider and billing.
Spending limits can be set at three levels:
| Level | What it controls |
|---|---|
| Organization | Total LLM spend across all managed agents |
| Team | LLM spend for all managed agents in a team |
| Agent | LLM spend for a single managed agent |
When a limit is exceeded, the agent's next LLM request is rejected. Heartbeat monitoring skips gracefully and does not count as an error. Scheduled tasks notify the user that the budget has been exceeded.
Cost visibility#
LLM costs are tracked per request and visible in:
- Cost Explorer — time-series cost breakdown by agent, team, or organization
- Audit trail — LLM proxy events include token counts and cost per request
- Agent overview — each agent's detail page shows 30-day LLM cost
Model configuration for managed agents#
| Setting | Default |
|---|---|
| Primary model | Claude (via AWS Bedrock) |
| Classification model | Claude Haiku |
| Temperature | 0.3 |
| Max output tokens | 16,000 per response |
| Step limit | 100 tool-use steps per invocation |
| Tool output truncation | 30,000 characters per tool call |
Prompt caching is applied automatically. Stable prompt sections (identity, workspace files, memory summary) are cached. Dynamic sections (timestamp, heartbeat context, recent messages) are not.
Related#
- Spending controls — budget enforcement for LLM usage
- Cost Explorer — cost visualization
- How Lens Agents works — architecture and request flow