Search for:
Why this server?
Integrates with Figma's API, allowing interaction with Figma files, comments, components, projects, and webhook management.
Why this server?
Enables seamless integration with Figma via the Model Context Protocol, allowing LLM applications to access, manipulate, and track Figma files, components, and variables.
Why this server?
A Model Context Protocol server that connects AI tools and LLMs to Figma designs, enabling them to extract design data, analyze design systems, and generate development documentation.
Why this server?
A Model Context Protocol server for interacting with the Figma API that handles large Figma files efficiently through memory-aware chunking and pagination capabilities.
Why this server?
Connects Gemini 2.5 Pro to Claude Code, enabling users to generate detailed implementation plans based on their codebase and receive feedback on code changes. Could use figma files to generate code.
Why this server?
A Model Context Protocol server that integrates Figma design files with AI coding tools like Cursor, Windsurf, and Cline, allowing AI tools to access and understand Figma design data for generating more accurate code.
Why this server?
Enables AI agents to interact with web browsers using natural language, featuring automated browsing, form filling, vision-based element detection, and structured JSON responses for systematic browser control. Useful to access Figma web.
Why this server?
Automates the creation of standardized documentation by extracting information from source files and applying templates, with integration capabilities for GitHub, Google Drive, and Perplexity AI. May work on extracting information for Figma.
Why this server?
Enables Cursor AI to interact with Figma designs, allowing users to read design information and programmatically modify elements through natural language commands.