Ollama Local API Server for LLM Inference | MCP Architecture