@@ -31,8 +31,7 @@ Server (vMCP) and an EmbeddingServer for semantic tool search.
3131- How to create an MCPGroup with multiple backend MCP servers
3232- How to deploy an EmbeddingServer for semantic search
3333- How to create a VirtualMCPServer with the optimizer enabled
34- - How to connect your AI client to the optimized endpoint and verify it exposes
35- only ` find_tool ` and ` call_tool `
34+ - How to connect your AI client to the optimized endpoint
3635
3736## About MCP Optimizer
3837
@@ -90,15 +89,13 @@ Before starting this tutorial, make sure you have:
9089- An MCP client (Visual Studio Code with GitHub Copilot is used in this
9190 tutorial)
9291
93- :::warning [ ARM64 compatibility ]
92+ :::tip [ ARM64 support ]
9493
95- The default text embeddings inference (TEI) images depend on Intel MKL, which is
96- x86_64-only. Native ARM64 support has been merged upstream but is not yet
97- included in a published release. If you are using Apple Silicon or any other
98- ARM64 nodes (including kind on macOS), you can run the amd64 image under
99- emulation as a workaround. See the
94+ The default TEI image is x86_64-only. If you are running on ARM64 nodes (for
95+ example, Apple Silicon with kind), set the ` image ` field in your EmbeddingServer
96+ to use the ARM64 image. See
10097[ EmbeddingServer resource] ( ../guides-vmcp/optimizer.mdx#embeddingserver-resource )
101- section for the required steps, including a Docker Desktop configuration change .
98+ for details .
10299
103100:::
104101
@@ -245,15 +242,15 @@ metadata:
245242 namespace : toolhive-system
246243spec :
247244 # highlight-start
245+ groupRef :
246+ name : optimizer-demo
248247 embeddingServerRef :
249248 name : optimizer-embedding
250249 # highlight-end
251250 incomingAuth :
252251 type : anonymous
253252 serviceType : ClusterIP
254253 config :
255- groupRef :
256- name : optimizer-demo
257254 aggregation :
258255 conflictResolution : prefix
259256 conflictResolutionConfig :
@@ -350,6 +347,14 @@ To check your token savings, send this prompt to your AI client:
350347
351348- "How many tokens did I save using MCP Optimizer?"
352349
350+ :::note
351+
352+ With only two backend MCP servers and a small number of tools, the optimizer may
353+ report minimal or no token savings. The benefit becomes more significant as you
354+ add more backends and tools to your MCPGroup.
355+
356+ :::
357+
353358## Clean up
354359
355360Remove the local workload and delete the Kubernetes resources when you're done:
@@ -384,7 +389,7 @@ kind delete cluster --name toolhive
384389## Related information
385390
386391- [ Optimize tool discovery] ( ../guides-vmcp/optimizer.mdx ) - full parameter
387- reference, high availability, and ARM64 workaround details
392+ reference, high availability, and ARM64 support details
388393- [ Optimizing LLM context] ( ../concepts/tool-optimization.mdx ) - background on
389394 tool filtering and context pollution
390395- [ Virtual MCP Server overview] ( ../concepts/vmcp.mdx ) - conceptual overview of
0 commit comments