Enable javascript in your browser for better experience. Need to know to enable it? Go here.


Model Catalog — Register any model via OpenAI-compatible API: vendor-hosted (OpenAI, Anthropic, etc.) or self-hosted (Nemotron, Llama, Mistral, etc.). Keys managed securely. Metadata (context windows, capabilities, internal/external classification) tracked per model.

 

MCP Server Registry — Register Model Context Protocol servers (internal or external) with automatic tool discovery. Agents see only tools they're authorized to access.

 

Agent Registry — Define agents using Oracle's AgentSpec DSL (JSON). Platform compiles to the configured runtime. Agents are version-controlled with full lineage tracking.

 

Policy Engine — Every action against the platform passes through Cerbos-based policy enforcement. Policies define which models & tools agents can access, how internal data flows to external providers and what actions are authorized across the agent graph. All rules are configurable via YAML, enabling flexible, environment-specific governance without hardcoding logic.

 

Observability — Every agent run instrumented end-to-end via OpenTelemetry. Full audit logs with literal input/output content at every step.

 

Memory Layer — Based on MemGPT research. Treats LLM as operating system—agent decides when to access memory. Enables small models (Claude Haiku, GPT-4o mini) to handle deep, multi-turn research by automatically managing context window economics. Creates knowledge graphs during execution.

 

Resource Governance — Set limits at any level: per model, per team, per month. Control token budgets, tool call limits, delegation depth. Prevent runaway costs and infinite agent recursion.

 

Profiling & Tuning — Configurable profiling integration provides flame-graph style execution timelines. See where time is spent, identify bottlenecks, auto-tune hyperparameters against evaluations.