Connect a local OpenAI-compatible endpoint and chat with tools.
Configure OpenAI-compatible /v1 endpoints. Examples: llama.cpp server, vLLM, LM Studio, Ollama (via /v1), OpenAI, etc.
/v1
MCP servers run on this machine (stdio). They expose tools the model can call mid-response.