Search for:
Why this server?
Integrates Cursor AI with Vite Dev server, allowing AI agents to modify code and observe live updates through the Hot Module Replacement system in real-time.
Why this server?
Server that enhances the capabilities of the Cline coding agent. It provides intelligent code suggestions, reduces hallucinations, and documents the knowledge base by leveraging your project's documentation and detecting the technologies used in your codebase.
Why this server?
A Model Context Protocol server that integrates Figma design files with AI coding tools like Cursor, Windsurf, and Cline, allowing AI tools to access and understand Figma design data for generating more accurate code.
Why this server?
A Model Context Protocol server that enables LLMs to read, search, and analyze code files with advanced caching and real-time file watching capabilities.
Why this server?
Leverages Vim's native text editing commands and workflows, which Claude already understands, to create a lightweight code assistance layer.
Why this server?
A Python server implementing the Model Context Protocol to provide customizable prompt templates, resources, and tools that enhance LLM interactions in the continue.dev environment.
Why this server?
A secure MCP server for executing whitelisted shell commands with resource and timeout controls, designed for integration with Claude and other MCP-compatible LLMs.
Why this server?
A line-oriented text file editor. Optimized for LLM tools with efficient partial file access to minimize token usage.
Why this server?
A Model Context Protocol server that helps large language models index, search, and analyze code repositories with minimal setup