diff --git a/registries/toolhive/servers/mcp-optimizer/server.json b/registries/toolhive/servers/mcp-optimizer/server.json index 676efb88..af0486e5 100644 --- a/registries/toolhive/servers/mcp-optimizer/server.json +++ b/registries/toolhive/servers/mcp-optimizer/server.json @@ -3,9 +3,9 @@ "_meta": { "io.modelcontextprotocol.registry/publisher-provided": { "io.github.stacklok": { - "ghcr.io/stackloklabs/mcp-optimizer:0.2.5": { + "ghcr.io/stackloklabs/mcp-optimizer:0.2.6": { "metadata": { - "last_updated": "2026-03-10T10:15:28Z", + "last_updated": "2026-03-10T14:53:37Z", "stars": 10 }, "overview": "## MCP Optimizer\n\nMCP Optimizer is an intelligent intermediary MCP server that acts as a unified gateway in front of all ToolHive-managed MCP servers. Instead of configuring your AI client with every individual MCP server, you point it at MCP Optimizer and it handles semantic tool discovery and routing automatically. It exposes a single MCP endpoint that aggregates tools from all running servers, intelligently routing each LLM request to the most appropriate tool regardless of which underlying server provides it.\n\nKey capabilities include group-based filtering to scope tool discovery to specific ToolHive groups (e.g. production vs staging environments), connection resilience with configurable exponential backoff retry logic, and support for both Docker and Kubernetes runtime modes. MCP Optimizer is particularly useful for managing large numbers of MCP tools, addressing the common problem of LLMs being overwhelmed by too many available tools.", @@ -143,7 +143,7 @@ "name": "RUNTIME_MODE" } ], - "identifier": "ghcr.io/stackloklabs/mcp-optimizer:0.2.5", + "identifier": "ghcr.io/stackloklabs/mcp-optimizer:0.2.6", "registryType": "oci", "transport": { "type": "streamable-http",