Gcore
Enables large language model assistants to interact with the Gcore Cloud API.
Acerca de
The Gcore Model Context Protocol (MCP) server acts as a crucial bridge, allowing Large Language Model (LLM) assistants to seamlessly interact with and manage Gcore Cloud resources. It provides a highly configurable interface, empowering LLMs to perform operations such as creating virtual machines, managing storage, configuring networks, and overseeing AI/ML services, all through a unified and intelligent automation layer. This tool simplifies cloud resource management by making it accessible and programmable via AI-driven interfaces.
Características Principales
- Unified configuration for tool selection via a single environment variable (GCORE_TOOLS)
- Flexible tool filtering with predefined toolsets, wildcard patterns, and combined modes
- Support for comprehensive Gcore Cloud API operations, including instances, networks, storage, and AI/ML
- Seamless integration capabilities with LLM development environments like Cursor IDE
- Lightweight execution options for temporary runs and persistent installation via `uv`
- 0 GitHub stars
Casos de Uso
- Automating Gcore Cloud resource provisioning and management through LLM assistants
- Integrating Gcore Cloud services into AI development workflows for intelligent automation
- Developing AI-powered applications that can dynamically interact with and control cloud infrastructure