Integrations
Provides a Node.js application interface for integrating the MCP server into custom applications, with support for context management and caching features
Gemini Context MCP Server
A powerful MCP (Model Context Protocol) server implementation that leverages Gemini's capabilities for context management and caching. This server maximizes the value of Gemini's 2M token context window while providing tools for efficient caching of large contexts.
🚀 Features
Context Management
- Up to 2M token context window support - Leverage Gemini's extensive context capabilities
- Session-based conversations - Maintain conversational state across multiple interactions
- Smart context tracking - Add, retrieve, and search context with metadata
- Semantic search - Find relevant context using semantic similarity
- Automatic context cleanup - Sessions and context expire automatically
API Caching
- Large prompt caching - Efficiently reuse large system prompts and instructions
- Cost optimization - Reduce token usage costs for frequently used contexts
- TTL management - Control cache expiration times
- Automatic cleanup - Expired caches are removed automatically
🏁 Quick Start
Prerequisites
- Node.js 18+ installed
- Gemini API key (Get one here)
Installation
Basic Usage
MCP Client Integration
This MCP server can be integrated with various MCP-compatible clients:
- Claude Desktop - Add as an MCP server in Claude settings
- Cursor - Configure in Cursor's AI/MCP settings
- VS Code - Use with MCP-compatible extensions
For detailed integration instructions with each client, see the MCP Client Configuration Guide in the MCP documentation.
Quick Client Setup
Use our simplified client installation commands:
Each command sets up the appropriate configuration files and provides instructions for completing the integration.
💻 Usage Examples
For Beginners
Directly using the server:
- Start the server:Copy
- Interact using the provided test scripts:Copy
Using in your Node.js application:
For Power Users
Using custom configurations:
Using the caching system for cost optimization:
🔌 Using with MCP Tools (like Cursor)
This server implements the Model Context Protocol (MCP), making it compatible with tools like Cursor or other AI-enhanced development environments.
Available MCP Tools
- Context Management Tools:
generate_text
- Generate text with contextget_context
- Get current context for a sessionclear_context
- Clear session contextadd_context
- Add specific context entriessearch_context
- Find relevant context semantically
- Caching Tools:
mcp_gemini_context_create_cache
- Create a cache for large contextsmcp_gemini_context_generate_with_cache
- Generate with cached contextmcp_gemini_context_list_caches
- List all available cachesmcp_gemini_context_update_cache_ttl
- Update cache TTLmcp_gemini_context_delete_cache
- Delete a cache
Connecting with Cursor
When used with Cursor, you can connect via the MCP configuration:
For detailed usage instructions for MCP tools, see README-MCP.md.
⚙️ Configuration Options
Environment Variables
Create a .env
file with these options:
🧪 Development
📚 Further Reading
- For MCP-specific usage, see README-MCP.md
- Explore the manifest in mcp-manifest.json to understand available tools
- Check example scripts in the repository for usage patterns
📋 Future Improvements
- Database persistence for context and caches
- Cache size management and eviction policies
- Vector-based semantic search
- Analytics and metrics tracking
- Integration with vector stores
- Batch operations for context management
- Hybrid caching strategies
- Automatic prompt optimization
📄 License
MIT
This server cannot be installed
remote-capable server
The server can be hosted and run remotely because it primarily relies on remote services or has no dependency on the local environment.
An MCP server implementation that maximizes Gemini's 2M token context window with tools for efficient context management and caching across multiple AI client applications.
Related MCP Servers
- -securityAlicense-qualityModel Context Protocol (MCP) server implementation that enables Claude Desktop to interact with Google's Gemini AI models.Last updated -53TypeScriptMIT License
- AsecurityFlicenseAqualityA Model Context Protocol (MCP) server that optimizes token usage by caching data during language model interactions, compatible with any language model and MCP client.Last updated -4JavaScript
- -securityAlicense-qualityAn MCP server implementation that standardizes how AI applications access tools and context, providing a central hub that manages tool discovery, execution, and context management with a simplified configuration system.Last updated -9PythonMIT License
- -security-license-qualityAn MCP server implementation that allows using Google's Gemini AI models (specifically Gemini 1.5 Pro) through Claude or other MCP clients via the Model Context Protocol.Last updated -1JavaScript