Skip to main content
Glama

MCP PostgreSQL Operations

MCP Server for PostgreSQL Operations and Monitoring

Overview

MCP-PostgreSQL-Ops is a professional MCP server for PostgreSQL database operations, monitoring, and management. Supports PostgreSQL 12-17 with comprehensive database analysis, performance monitoring, and intelligent maintenance recommendations through natural language queries. Most features work independently, but advanced query analysis capabilities are enhanced when pg_stat_statements and (optionally) pg_stat_monitor extensions are installed.


Features

  • Zero Configuration: Works with PostgreSQL 12-17 out-of-the-box with automatic version detection.
  • Natural Language: Ask questions like "Show me slow queries" or "Analyze table bloat."
  • Production Safe: Read-only operations, RDS/Aurora compatible with regular user permissions.
  • Extension Enhanced: Optional pg_stat_statements and pg_stat_monitor for advanced query analytics.
  • Comprehensive Database Monitoring: Performance analysis, bloat detection, and maintenance recommendations.
  • Smart Query Analysis: Slow query identification with pg_stat_statements and pg_stat_monitor integration.
  • Schema & Relationship Discovery: Database structure exploration with detailed relationship mapping.
  • VACUUM & Autovacuum Intelligence: Real-time maintenance monitoring and effectiveness analysis.
  • Multi-Database Operations: Seamless cross-database analysis and monitoring.
  • Enterprise-Ready: Safe read-only operations with RDS/Aurora compatibility.
  • Developer-Friendly: Simple codebase for easy customization and tool extension.

🔧 Advanced Capabilities

  • Version-aware I/O statistics (enhanced on PostgreSQL 16+).
  • Real-time connection and lock monitoring.
  • Background process and checkpoint analysis.
  • Replication status and WAL monitoring.
  • Database capacity and bloat analysis.

Tool Usage Examples

📸 More Examples with Screenshots →


MCP-PostgreSQL-Ops Usage Screenshot


MCP-PostgreSQL-Ops Usage Screenshot


⭐ Quickstart (5 minutes)

Note: The postgresql container included in docker-compose.yml is intended for quickstart testing purposes only. You can connect to your own PostgreSQL instance by adjusting the environment variables as needed.

If you want to use your own PostgreSQL instance instead of the built-in test container:

  • Update the target PostgreSQL connection information in your .env file (see POSTGRES_HOST, POSTGRES_PORT, POSTGRES_USER, POSTGRES_PASSWORD, POSTGRES_DB).
  • In docker-compose.yml, comment out (disable) the postgres and postgres-init-extensions containers to avoid starting the built-in test database.

Flow Diagram of Quickstart/Tutorial

Flow Diagram of Quickstart/Tutorial

1. Environment Setup

Note: While superuser privileges provide access to all databases and system information, the MCP server also works with regular user permissions for basic monitoring tasks.

git clone https://github.com/call518/MCP-PostgreSQL-Ops.git cd MCP-PostgreSQL-Ops ### Check and modify .env file cp .env.example .env vim .env
### No need to modify defaults, but if using your own PostgreSQL server, edit below: POSTGRES_HOST=host.docker.internal POSTGRES_PORT=15432 # External port for host access (mapped to internal 5432) POSTGRES_USER=postgres POSTGRES_PASSWORD=changeme!@34 POSTGRES_DB=ecommerce # Default connection DB. Superusers can access all DBs.

Note: PGDATA=/data/db is preconfigured for the Percona PostgreSQL Docker image, which requires this specific path for proper write permissions.

2. Start Demo Containers

# Start all containers including built-in PostgreSQL for testing docker-compose up -d # Alternative: If using your own PostgreSQL instance # Comment out postgres and postgres-init-extensions services in docker-compose.yml # Then use the custom configuration: # docker-compose -f docker-compose.custom-db.yml up -d

⏰ Wait for Environment Setup: The initial environment setup takes a few minutes as containers are started in sequence:

  1. PostgreSQL container starts first with database initialization
  2. PostgreSQL Extensions container installs extensions and creates comprehensive test data (~83K records)
  3. MCP Server and MCPO Proxy containers start after PostgreSQL is ready
  4. OpenWebUI container starts last and may take additional time to load the web interface

💡 Tip: Wait 2-3 minutes after running docker-compose up -d before accessing OpenWebUI to ensure all services are fully initialized.

🔍 Check Container Status (Optional):

# Monitor container startup progress docker-compose logs -f # Check if all containers are running docker-compose ps # Verify PostgreSQL is ready docker-compose logs postgres | grep "ready to accept connections"

3. Access to OpenWebUI

http://localhost:3003/

  • The list of MCP tool features provided by swagger can be found in the MCPO API Docs URL.
    • e.g: http://localhost:8003/docs

4. Registering the Tool in OpenWebUI

  1. logging in to OpenWebUI with an admin account
  2. go to "Settings" → "Tools" from the top menu.
  3. Enter the postgresql-ops Tool address (e.g., http://localhost:8003/postgresql-ops) to connect MCP Tools.
  4. Setup Ollama or OpenAI.

5. Complete!

Congratulations! Your MCP PostgreSQL Operations server is now ready for use. You can start exploring your databases with natural language queries.

🚀 Try These Example Queries:
  • "Show me the current active connections"
  • "What are the slowest queries in the system?"
  • "Analyze table bloat across all databases"
  • "Show me database size information"
  • "What tables need VACUUM maintenance?"
📖 Next Steps:

(NOTE) Sample Test Data Overview

The create-test-data.sql script is executed by the postgres-init-extensions container (defined in docker-compose.yml) on first startup, automatically generating comprehensive test databases for MCP tool testing:

DatabasePurposeSchema & TablesScale
ecommerceE-commerce systempublic: categories, products, customers, orders, order_items10 categories, 500 products, 100 customers, 200 orders, 400 order items
analyticsAnalytics & reportingpublic: page_views, sales_summary1,000 page views, 30 sales summaries
inventoryWarehouse managementpublic: suppliers, inventory_items, purchase_orders10 suppliers, 100 items, 50 purchase orders
hr_systemHR managementpublic: departments, employees, payroll5 departments, 50 employees, 150 payroll records

Test users created: app_readonly, app_readwrite, analytics_user, backup_user

Optimized for testing: Intentional table bloat, various indexes (used/unused), time-series data, complex relationships


Tool Compatibility Matrix

Automatic Adaptation: All tools work transparently across supported versions - no configuration needed!

🟢 Extension-Independent Tools (No Extensions Required)

Tool NameExtensions RequiredPG 12PG 13PG 14PG 15PG 16PG 17System Views/Tables Used
get_server_info❌ Noneversion(), pg_extension
get_active_connections❌ Nonepg_stat_activity
get_postgresql_config❌ Nonepg_settings
get_database_list❌ Nonepg_database
get_table_list❌ Noneinformation_schema.tables
get_table_schema_info❌ Noneinformation_schema.*, pg_indexes
get_database_schema_info❌ Nonepg_namespace, pg_class, pg_proc
get_table_relationships❌ Noneinformation_schema.* (constraints)
get_user_list❌ Nonepg_user, pg_roles
get_index_usage_stats❌ Nonepg_stat_user_indexes
get_database_size_info❌ Nonepg_database_size()
get_table_size_info❌ Nonepg_total_relation_size()
get_vacuum_analyze_stats❌ Nonepg_stat_user_tables
get_current_database_info❌ Nonepg_database, current_database()
get_table_bloat_analysis❌ Nonepg_stat_user_tables
get_database_bloat_overview❌ Nonepg_stat_user_tables
get_autovacuum_status❌ Nonepg_stat_user_tables
get_autovacuum_activity❌ Nonepg_stat_user_tables
get_running_vacuum_operations❌ Nonepg_stat_activity
get_vacuum_effectiveness_analysis❌ Nonepg_stat_user_tables
get_table_bloat_analysis❌ Nonepg_stat_user_tables
get_database_bloat_overview❌ Nonepg_stat_user_tables
get_lock_monitoring❌ Nonepg_locks, pg_stat_activity
get_wal_status❌ Nonepg_current_wal_lsn()
get_database_stats❌ Nonepg_stat_database
get_table_io_stats❌ Nonepg_statio_user_tables
get_index_io_stats❌ Nonepg_statio_user_indexes
get_database_conflicts_stats❌ Nonepg_stat_database_conflicts

🚀 Version-Aware Tools (Auto-Adapting)

Tool NameExtensions RequiredPG 12PG 13PG 14PG 15PG 16PG 17Special Features
get_io_stats❌ None✅ Basic✅ Basic✅ Basic✅ BasicEnhancedEnhancedPG16+: pg_stat_io support
get_bgwriter_stats❌ NoneSpecialPG15: Separate checkpointer stats
get_replication_status❌ None✅ CompatibleEnhancedEnhancedEnhancedEnhancedEnhancedPG13+: wal_status, safe_wal_size; PG16+: enhanced WAL receiver
get_all_tables_stats❌ None✅ CompatibleEnhancedEnhancedEnhancedEnhancedEnhancedPG13+: n_ins_since_vacuum tracking for vacuum maintenance optimization
get_user_functions_stats⚙️ Config RequiredRequires track_functions=pl

🟡 Extension-Dependent Tools (Extensions Required)

Tool NameRequired ExtensionPG 12PG 13PG 14PG 15PG 16PG 17Notes
get_pg_stat_statements_top_queriespg_stat_statementsCompatibleEnhancedEnhancedEnhancedEnhancedEnhancedPG12: total_timetotal_exec_time; PG13+: native total_exec_time
get_pg_stat_monitor_recent_queriespg_stat_monitorCompatibleEnhancedEnhancedEnhancedEnhancedEnhancedPG12: total_timetotal_exec_time; PG13+: native total_exec_time

📋 PostgreSQL 18 Support: PostgreSQL 18 is currently in beta phase and not yet supported by Percona Distribution PostgreSQL. Support will be added once PostgreSQL 18 reaches stable release and distribution support becomes available.


Usage Examples

Claude Desktop Integration

(Recommended) Add to your Claude Desktop configuration file:

{ "mcpServers": { "postgresql-ops": { "command": "uvx", "args": ["--python", "3.11", "mcp-postgresql-ops"], "env": { "POSTGRES_HOST": "127.0.0.1", "POSTGRES_PORT": "15432", "POSTGRES_USER": "postgres", "POSTGRES_PASSWORD": "changeme!@34", "POSTGRES_DB": "ecommerce" } } } }

"Show all active connections in a clear and readable html table format." Claude Desktop Integration

"Show all relationships for customers table in ecommerce database as a Mermaid diagram." Claude Desktop Integration


Installation

# Install the package pip install mcp-postgresql-ops # Or with uv (faster) uv add mcp-postgresql-ops # Verify installation mcp-postgresql-ops --help

From Source

# Clone the repository git clone https://github.com/call518/MCP-PostgreSQL-Ops.git cd MCP-PostgreSQL-Ops # Install with uv (recommended) uv sync uv run mcp-postgresql-ops --help # Or with pip pip install -e . mcp-postgresql-ops --help

MCP Configuration

Claude Desktop Configuration

(Optional) Run with Local Source:

{ "mcpServers": { "postgresql-ops": { "command": "uv", "args": ["run", "python", "-m", "mcp_postgresql_ops"], "env": { "POSTGRES_HOST": "127.0.0.1", "POSTGRES_PORT": "15432", "POSTGRES_USER": "postgres", "POSTGRES_PASSWORD": "changeme!@34", "POSTGRES_DB": "ecommerce" } } } }

Run MCP-Server as Standalon

# Stdio mode uvx --python 3.11 mcp-postgresql-ops \ --type stdio # HTTP mode uvx --python 3.11 mcp-postgresql-ops --type streamable-http \ --host 127.0.0.1 \ --port 8000 \ --log-level DEBUG

(Option) Configure Multiple PostgreSQL Instances

{ "mcpServers": { "Postgresql-A": { "command": "uvx", "args": ["--python", "3.11", "mcp-postgresql-ops"], "env": { "POSTGRES_HOST": "a.foo.com", "POSTGRES_PORT": "5432", "POSTGRES_USER": "postgres", "POSTGRES_PASSWORD": "postgres", "POSTGRES_DB": "postgres" } }, "Postgresql-B": { "command": "uvx", "args": ["--python", "3.11", "mcp-postgresql-ops"], "env": { "POSTGRES_HOST": "b.bar.com", "POSTGRES_PORT": "5432", "POSTGRES_USER": "postgres", "POSTGRES_PASSWORD": "postgres", "POSTGRES_DB": "postgres" } } } }
/w Local Source
# Method 1: Module execution (for development, requires PYTHONPATH) PYTHONPATH=/path/to/MCP-PostgreSQL-Ops/src python -m mcp_postgresql_ops \ --type stdio # Method 2: Direct script (after uv installation in project directory) uv run mcp-postgresql-ops \ --type stdio # Method 3: Installed package script (after pip/uv install) mcp-postgresql-ops \ --type stdio # HTTP mode examples: # Development mode PYTHONPATH=/path/to/MCP-PostgreSQL-Ops/src python -m mcp_postgresql_ops \ --type streamable-http \ --host 127.0.0.1 \ --port 8000 \ --log-level DEBUG # Production mode (after installation) mcp-postgresql-ops \ --type streamable-http \ --host 127.0.0.1 \ --port 8000 \ --log-level DEBUG

CLI Arguments

  • --type: Transport type (stdio or streamable-http) - Default: stdio
  • --host: Host address for HTTP transport - Default: 127.0.0.1
  • --port: Port number for HTTP transport - Default: 8000
  • --auth-enable: Enable Bearer token authentication for streamable-http mode - Default: false
  • --secret-key: Secret key for Bearer token authentication (required when auth enabled)
  • --log-level: Logging level (DEBUG, INFO, WARNING, ERROR, CRITICAL) - Default: INFO

Environment Variables

VariableDescriptionDefaultProject Default
PYTHONPATHPython module search path (only needed for development mode)-/app/src
MCP_LOG_LEVELServer logging verbosity (DEBUG, INFO, WARNING, ERROR)INFOINFO
FASTMCP_TYPEMCP transport protocol (stdio for CLI, streamable-http for web)stdiostreamable-http
FASTMCP_HOSTHTTP server bind address (0.0.0.0 for all interfaces)127.0.0.10.0.0.0
FASTMCP_PORTHTTP server port for MCP communication80008000
REMOTE_AUTH_ENABLEEnable Bearer token authentication for streamable-http mode (Default: false if undefined/null/empty)falsefalse
REMOTE_SECRET_KEYSecret key for Bearer token authentication (required when auth enabled)-your-secret-key-here
PGSQL_VERSIONPostgreSQL major version for Docker image selection1717
PGDATAPostgreSQL data directory inside Docker container (Do not modify)/var/lib/postgresql/data/data/db
POSTGRES_HOSTPostgreSQL server hostname or IP address127.0.0.1host.docker.internal
POSTGRES_PORTPostgreSQL server port number543215432
POSTGRES_USERPostgreSQL connection username (needs read permissions)postgrespostgres
POSTGRES_PASSWORDPostgreSQL user password (supports special characters)changeme!@34changeme!@34
POSTGRES_DBDefault database name for connectionstestdbecommerce
POSTGRES_MAX_CONNECTIONSPostgreSQL max_connections configuration parameter200200
DOCKER_EXTERNAL_PORT_OPENWEBUIHost port mapping for Open WebUI container80803003
DOCKER_EXTERNAL_PORT_MCP_SERVERHost port mapping for MCP server container808018003
DOCKER_EXTERNAL_PORT_MCPO_PROXYHost port mapping for MCPO proxy container80008003
DOCKER_INTERNAL_PORT_POSTGRESQLPostgreSQL container internal port54325432

Note: POSTGRES_DB serves as the default target database for operations when no specific database is specified. In Docker environments, if set to a non-default name, this database will be automatically created during initial PostgreSQL startup.

Port Configuration: The built-in PostgreSQL container uses port mapping 15432:5432 where:

  • POSTGRES_PORT=15432: External port for host access and MCP server connections
  • DOCKER_INTERNAL_PORT_POSTGRESQL=5432: Internal container port (PostgreSQL default)
  • When using external PostgreSQL servers, set POSTGRES_PORT to match your server's actual port

Prerequisites

Required PostgreSQL Extensions

For more details, see the ## Tool Compatibility Matrix

Note: Most MCP tools work without any PostgreSQL extensions. section below. Some advanced performance analysis tools require the following extensions:

-- Query performance statistics (required only for get_pg_stat_statements_top_queries) CREATE EXTENSION IF NOT EXISTS pg_stat_statements; -- Advanced monitoring (optional, used by get_pg_stat_monitor_recent_queries) CREATE EXTENSION IF NOT EXISTS pg_stat_monitor;

Quick Setup: For new PostgreSQL installations, add to postgresql.conf:

shared_preload_libraries = 'pg_stat_statements'

Then restart PostgreSQL and run the CREATE EXTENSION commands above.

  • pg_stat_statements is required only for slow query analysis tools.
  • pg_stat_monitor is optional and used for real-time query monitoring.
  • All other tools work without these extensions.

Minimum Requirements

  • PostgreSQL 12+ (tested with PostgreSQL 17)
  • Python 3.11
  • Network access to PostgreSQL server
  • Read permissions on system catalogs

Required PostgreSQL Configuration

⚠️ Statistics Collection Settings: Some MCP tools require specific PostgreSQL configuration parameters to collect statistics. Choose one of the following configuration methods:

Tools affected by these settings:

  • get_user_functions_stats: Requires track_functions = pl or track_functions = all
  • get_table_io_stats & get_index_io_stats: More accurate timing with track_io_timing = on
  • get_database_stats: Enhanced I/O timing with track_io_timing = on

Verification: After applying any method, verify the settings:

SELECT name, setting, context FROM pg_settings WHERE name IN ('track_activities', 'track_counts', 'track_io_timing', 'track_functions') ORDER BY name; name | setting | context ------------------+---------+----------- track_activities | on | superuser track_counts | on | superuser track_functions | pl | superuser track_io_timing | on | superuser (4 rows)

Add the following to your postgresql.conf:

# Basic statistics collection (usually enabled by default) track_activities = on track_counts = on # Required for function statistics tools track_functions = pl # Enables PL/pgSQL function statistics collection # Optional but recommended for accurate I/O timing track_io_timing = on # Enables I/O timing statistics collection

Then restart PostgreSQL server.

Method 2: PostgreSQL Startup Parameters

For Docker or command-line PostgreSQL startup:

# Docker example docker run -d \ -e POSTGRES_PASSWORD=mypassword \ postgres:17 \ -c track_activities=on \ -c track_counts=on \ -c track_functions=pl \ -c track_io_timing=on # Direct postgres command postgres -D /data \ -c track_activities=on \ -c track_counts=on \ -c track_functions=pl \ -c track_io_timing=on
Method 3: Dynamic Configuration (AWS RDS, Azure, GCP, Managed Services)

For managed PostgreSQL services where you cannot modify postgresql.conf, use SQL commands to change settings dynamically:

-- Enable basic statistics collection (usually enabled by default) ALTER SYSTEM SET track_activities = 'on'; ALTER SYSTEM SET track_counts = 'on'; -- Enable function statistics collection (requires superuser privileges) ALTER SYSTEM SET track_functions = 'pl'; -- Enable I/O timing statistics (optional but recommended) ALTER SYSTEM SET track_io_timing = 'on'; -- Reload configuration without restart (run separately) SELECT pg_reload_conf();

Alternative for session-level testing:

-- Set for current session only (temporary) SET track_activities = 'on'; SET track_counts = 'on'; SET track_functions = 'pl'; SET track_io_timing = 'on';

Note: When using command-line tools, run each SQL statement separately to avoid transaction block errors.


RDS/Aurora Compatibility

  • This server is read-only and works with regular roles on RDS/Aurora. For advanced analysis enable pg_stat_statements; pg_stat_monitor is not available on managed engines.
  • On RDS/Aurora, prefer DB Parameter Group over ALTER SYSTEM for persistent settings.
    -- Verify preload setting SHOW shared_preload_libraries; -- Enable extension in target DB CREATE EXTENSION IF NOT EXISTS pg_stat_statements; -- Recommended visibility for monitoring GRANT pg_read_all_stats TO <app_user>;

Example Queries

🟢 Extension-Independent Tools (Always Available)

  • get_server_info
    • "Show PostgreSQL server version and extension status."
    • "Check if pg_stat_statements is installed."
  • get_active_connections
    • "Show all active connections."
    • "List current sessions with database and user."
  • get_postgresql_config
    • "Show all PostgreSQL configuration parameters."
    • "Find all memory-related configuration settings."
  • get_database_list
    • "List all databases and their sizes."
    • "Show database list with owner information."
  • get_table_list
    • "List all tables in the ecommerce database."
    • "Show table sizes in the public schema."
  • get_table_schema_info
    • "Show detailed schema information for the customers table in ecommerce database."
    • "Get column details and constraints for products table in ecommerce database."
    • "Analyze table structure with indexes and foreign keys for orders table in sales schema of ecommerce database."
    • "Show schema overview for all tables in public schema of inventory database."
    • 📋 Features: Column types, constraints, indexes, foreign keys, table metadata
    • ⚠️ Required: database_name parameter must be specified
  • get_database_schema_info
    • "Show all schemas in ecommerce database with their contents."
    • "Get detailed information about sales schema in ecommerce database."
    • "Analyze schema structure and permissions for inventory database."
    • "Show schema overview with table counts and sizes for hr_system database."
    • 📋 Features: Schema owners, permissions, object counts, sizes, contents
    • ⚠️ Required: database_name parameter must be specified
  • get_table_relationships
    • "Show all relationships for customers table in ecommerce database."
    • "Analyze foreign key relationships for orders table in sales schema of ecommerce database."
    • "Get database-wide relationship overview for ecommerce database."
    • "Find all tables that reference products table in ecommerce database."
    • "Show cross-schema relationships in inventory database."
    • 📋 Features: Foreign key relationships (inbound/outbound), cross-schema dependencies, constraint details
    • ⚠️ Required: database_name parameter must be specified
    • 💡 Usage: Leave table_name empty for database-wide relationship analysis
  • get_user_list
    • "List all database users and their roles."
    • "Show user permissions for a specific database."
  • get_index_usage_stats
    • "Analyze index usage efficiency."
    • "Find unused indexes in the current database."
  • get_database_size_info
    • "Show database capacity analysis."
    • "Find the largest databases by size."
  • get_table_size_info
    • "Show table and index size analysis."
    • "Find largest tables in a specific schema."
  • get_vacuum_analyze_stats
    • "Show recent VACUUM and ANALYZE operations."
    • "List tables needing VACUUM."
  • get_current_database_info
    • "What database am I connected to?"
    • "Show current database information and connection details."
    • "Display database encoding, collation, and size information."
    • 📋 Features: Database name, encoding, collation, size, connection limits
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
  • get_table_bloat_analysis
    • "Analyze table bloat in the current database."
    • "Show dead tuple ratios and bloat estimates for user_logs table pattern."
    • "Find tables with high bloat that need VACUUM maintenance."
    • "Analyze bloat in specific schema with minimum 100 dead tuples."
    • 📋 Features: Dead tuple ratios, bloat size estimates, VACUUM recommendations, pattern filtering
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
    • 💡 Usage: Extension-Independent approach using pg_stat_user_tables
  • get_database_bloat_overview
    • "Show database-wide bloat summary by schema."
    • "Get high-level view of storage efficiency across all schemas."
    • "Identify schemas requiring maintenance attention."
    • 📋 Features: Schema-level aggregation, total bloat estimates, maintenance status
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
  • get_autovacuum_status
    • "Check autovacuum configuration and trigger conditions."
    • "Show tables needing immediate autovacuum attention."
    • "Analyze autovacuum threshold percentages for public schema."
    • "Find tables approaching autovacuum trigger points."
    • 📋 Features: Trigger threshold analysis, urgency classification, configuration status
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
    • 💡 Usage: Extension-Independent autovacuum monitoring using pg_stat_user_tables
  • get_autovacuum_activity
    • "Show autovacuum activity patterns for the last 48 hours."
    • "Monitor autovacuum execution frequency and timing."
    • "Find tables with irregular autovacuum patterns."
    • "Analyze recent autovacuum and autoanalyze history."
    • 📋 Features: Activity patterns, execution frequency, timing analysis
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
    • 💡 Usage: Historical autovacuum pattern analysis
  • get_running_vacuum_operations
    • "Show currently running VACUUM and ANALYZE operations."
    • "Monitor active maintenance operations and their progress."
    • "Check if any VACUUM operations are blocking queries."
    • "Find long-running maintenance operations."
    • 📋 Features: Real-time operation status, elapsed time, impact level, process details
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
    • 💡 Usage: Real-time maintenance monitoring using pg_stat_activity
  • get_vacuum_effectiveness_analysis
    • "Analyze VACUUM effectiveness and maintenance patterns."
    • "Compare manual VACUUM vs autovacuum efficiency."
    • "Find tables with suboptimal maintenance patterns."
    • "Check VACUUM frequency vs table activity ratios."
    • 📋 Features: Maintenance pattern analysis, effectiveness assessment, DML-to-VACUUM ratios
    • 🔧 PostgreSQL 12-17: Fully compatible, no extensions required
    • 💡 Usage: Strategic VACUUM analysis using existing statistics
  • get_table_bloat_analysis
    • "Analyze table bloat in the public schema."
    • "Show tables with high dead tuple ratios in ecommerce database."
    • "Find tables requiring VACUUM maintenance."
    • "Check bloat for tables with more than 5000 dead tuples."
    • 📋 Features: Dead tuple ratios, estimated bloat size, VACUUM recommendations
    • ⚠️ Required: Specify database_name for cross-database analysis
  • get_database_bloat_overview
    • "Show database-wide bloat summary by schema."
    • "Get bloat overview for inventory database."
    • "Identify schemas with highest bloat ratios."
    • "Database maintenance planning with bloat statistics."
    • 📋 Features: Schema-level aggregation, maintenance priorities, size recommendations
  • get_lock_monitoring
    • "Show all current locks and blocked sessions."
    • "Show only blocked sessions with granted=false filter."
    • "Monitor locks by specific user with username filter."
    • "Check exclusive locks with mode filter."
  • get_wal_status
    • "Show WAL status and archiving information."
    • "Monitor WAL generation and current LSN position."
  • get_replication_status
    • "Check replication connections and lag status."
    • "Monitor replication slots and WAL receiver status."
  • get_database_stats
    • "Show comprehensive database performance metrics."
    • "Analyze transaction commit ratios and I/O statistics."
    • "Monitor buffer cache hit ratios and temporary file usage."
  • get_bgwriter_stats
    • "Analyze checkpoint performance and timing."
    • "Show me checkpoint performance."
    • "Show background writer efficiency statistics."
    • "Monitor buffer allocation and fsync patterns."
  • get_user_functions_stats
    • "Analyze user-defined function performance."
    • "Show function call counts and execution times."
    • "Identify performance bottlenecks in custom functions."
    • ⚠️ Requires: track_functions = pl in postgresql.conf
  • get_table_io_stats
    • "Analyze table I/O performance and buffer hit ratios."
    • "Identify tables with poor buffer cache performance."
    • "Monitor TOAST table I/O statistics."
    • 💡 Enhanced with: track_io_timing = on for accurate timing
  • get_index_io_stats
    • "Show index I/O performance and buffer efficiency."
    • "Identify indexes causing excessive disk I/O."
    • "Monitor index cache-friendliness patterns."
    • 💡 Enhanced with: track_io_timing = on for accurate timing
  • get_database_conflicts_stats
    • "Check replication conflicts on standby servers."
    • "Analyze conflict types and resolution statistics."
    • "Monitor standby server query cancellation patterns."
    • "Monitor WAL generation and current LSN position."
  • get_replication_status
    • "Check replication connections and lag status."
    • "Monitor replication slots and WAL receiver status."

🚀 Version-Aware Tools (Auto-Adapting)

  • get_io_stats (New!)
    • "Show comprehensive I/O statistics." (PostgreSQL 16+ provides detailed breakdown)
    • "Analyze I/O statistics."
    • "Analyze buffer cache efficiency and I/O timing."
    • "Monitor I/O patterns by backend type and context."
    • 📈 PG16+: Full pg_stat_io with timing, backend types, and contexts
    • 📊 PG12-15: Basic pg_statio_* fallback with buffer hit ratios
  • get_bgwriter_stats (Enhanced!)
    • "Show background writer and checkpoint performance."
    • 📈 PG15: Separate checkpointer and bgwriter statistics (unique feature)
    • 📊 PG12-14, 16+: Combined bgwriter stats (includes checkpointer data)
  • get_server_info (Enhanced!)
    • "Show server version and compatibility features."
    • "Check server compatibility."
    • "Check what MCP tools are available on this PostgreSQL version."
    • "Displays feature availability matrix and upgrade recommendations."
  • get_all_tables_stats (Enhanced!)
    • "Show comprehensive statistics for all tables." (version-compatible for PG12-17)
    • "Include system tables with include_system=true parameter."
    • "Analyze table access patterns and maintenance needs."
    • 📈 PG13+: Tracks insertions since vacuum (n_ins_since_vacuum) for optimal maintenance scheduling
    • 📊 PG12: Compatible mode with NULL for unsupported columns

🟡 Extension-Dependent Tools

  • get_pg_stat_statements_top_queries (Requires pg_stat_statements)
    • "Show top 10 slowest queries."
    • "Analyze slow queries in the inventory database."
    • 📈 Version-Compatible: PG12 uses total_timetotal_exec_time mapping; PG13+ uses native columns
    • 💡 Cross-Version: Automatically adapts query structure for PostgreSQL 12-17 compatibility
  • get_pg_stat_monitor_recent_queries (Optional, uses pg_stat_monitor)
    • "Show recent queries in real time."
    • "Monitor query activity for the last 5 minutes."
    • 📈 Version-Compatible: PG12 uses total_timetotal_exec_time mapping; PG13+ uses native columns
    • 💡 Cross-Version: Automatically adapts query structure for PostgreSQL 12-17 compatibility

💡 Pro Tip: All tools support multi-database operations using the database_name parameter. This allows PostgreSQL superusers to analyze and monitor multiple databases from a single MCP server instance.


Troubleshooting

Connection Issues

  1. Check PostgreSQL server status
  2. Verify connection parameters in .env file
  3. Ensure network connectivity
  4. Check user permissions

Extension Errors

  1. Run get_server_info to check extension status
  2. Install missing extensions:
    CREATE EXTENSION pg_stat_statements; CREATE EXTENSION pg_stat_monitor;
  3. Restart PostgreSQL if needed

Configuration Issues

  1. "No data found" for function statistics: Check track_functions setting
    SHOW track_functions; -- Should be 'pl' or 'all'
    Quick fix for managed services (AWS RDS, etc.):
    ALTER SYSTEM SET track_functions = 'pl'; SELECT pg_reload_conf();
  2. Missing I/O timing data: Enable timing collection
    SHOW track_io_timing; -- Should be 'on'
    Quick fix:
    ALTER SYSTEM SET track_io_timing = 'on'; SELECT pg_reload_conf();
  3. Apply configuration changes:
    • Self-managed: Add settings to postgresql.conf and restart server
    • Managed services: Use ALTER SYSTEM SET + SELECT pg_reload_conf()
    • Temporary testing: Use SET parameter = value for current session
    • Generate some database activity to populate statistics

Performance Issues

  1. Use limit parameters to reduce result size
  2. Run monitoring during off-peak hours
  3. Check database load before running analysis

Version Compatibility Issues

For more details, see the ## Tool Compatibility Matrix

  1. Run compatibility check first:
    # "Use get_server_info to check version and available features"
  2. Understanding feature availability:
    • PostgreSQL 16-17: All features available
    • PostgreSQL 15+: Separate checkpointer stats
    • PostgreSQL 14+: Parallel query tracking
    • PostgreSQL 12-13: Core functionality only
  3. If a tool shows "Not Available":
    • Feature requires newer PostgreSQL version
    • Tool will automatically use best available alternative
    • Consider upgrading PostgreSQL for enhanced monitoring

Development

Testing & Development

# Clone and setup for development git clone https://github.com/call518/MCP-PostgreSQL-Ops.git cd MCP-PostgreSQL-Ops uv sync # Test with MCP Inspector (loads .env automatically) ./scripts/run-mcp-inspector-local.sh # Direct execution methods: # 1. Using uv run (recommended for development) uv run mcp-postgresql-ops --log-level DEBUG # 2. Module execution (requires PYTHONPATH) PYTHONPATH=src python -m mcp_postgresql_ops --log-level DEBUG # 3. After installation mcp-postgresql-ops --log-level DEBUG # Test version compatibility (requires different PostgreSQL versions) # Modify POSTGRES_HOST in .env to point to different versions # Run tests (if you add any) uv run pytest

Version Compatibility Testing

The MCP server automatically adapts to PostgreSQL versions 12-17. To test across versions:

  1. Set up test databases: Different PostgreSQL versions (12, 14, 15, 16, 17)
  2. Run compatibility tests: Point to each version and verify tool behavior
  3. Check feature detection: Ensure proper version detection and feature availability
  4. Verify fallback behavior: Confirm graceful degradation on older versions

Security Notes

  • All tools are read-only - no data modification capabilities
  • Sensitive information (passwords) are masked in outputs
  • No direct SQL execution - only predefined queries
  • Follows principle of least privilege

Contributing

🤝 Got ideas? Found bugs? Want to add cool features?

We're always excited to welcome new contributors! Whether you're fixing a typo, adding a new monitoring tool, or improving documentation - every contribution makes this project better.

Ways to contribute:

  • 🐛 Report issues or bugs
  • 💡 Suggest new PostgreSQL monitoring features
  • 📝 Improve documentation
  • 🚀 Submit pull requests
  • ⭐ Star the repo if you find it useful!

Pro tip: The codebase is designed to be super friendly for adding new tools. Check out the existing @mcp.tool() functions in mcp_main.py.


MCPO Swagger Docs

[MCPO Swagger URL] http://localhost:8003/postgresql-ops/docs

MCPO Swagger APIs


🔐 Security & Authentication

Bearer Token Authentication

For streamable-http mode, this MCP server supports Bearer token authentication to secure remote access. This is especially important when running the server in production environments.

Default Policy: REMOTE_AUTH_ENABLE defaults to false if undefined, null, or empty. This ensures backward compatibility and prevents startup errors when the variable is not set.

Configuration

Enable Authentication:

# In .env file REMOTE_AUTH_ENABLE=true REMOTE_SECRET_KEY=your-secure-secret-key-here

Or via CLI:

# Module method python -m mcp_postgresql_ops --type streamable-http --auth-enable --secret-key your-secure-secret-key-here # Script method mcp-postgresql-ops --type streamable-http --auth-enable --secret-key your-secure-secret-key-here
Security Levels
  1. stdio mode (Default): Local-only access, no authentication needed
  2. streamable-http + REMOTE_AUTH_ENABLE=false: Remote access without authentication ⚠️ NOT RECOMMENDED for production
  3. streamable-http + REMOTE_AUTH_ENABLE=true: Remote access with Bearer token authentication ✅ RECOMMENDED for production
Client Configuration

When authentication is enabled, MCP clients must include the Bearer token in the Authorization header:

{ "mcpServers": { "postgresql-ops": { "type": "streamable-http", "url": "http://your-server:8000/mcp", "headers": { "Authorization": "Bearer your-secure-secret-key-here" } } } }
Security Best Practices
  • Always enable authentication when using streamable-http mode in production
  • Use strong, randomly generated secret keys (32+ characters recommended)
  • Use HTTPS when possible (configure reverse proxy with SSL/TLS)
  • Restrict network access using firewalls or network policies
  • Rotate secret keys regularly for enhanced security
  • Monitor access logs for unauthorized access attempts
Error Handling

When authentication fails, the server returns:

  • 401 Unauthorized for missing or invalid tokens
  • Detailed error messages in JSON format for debugging

License

Freely use, modify, and distribute under the MIT License.


⭐ Other Projects

Other MCP servers by the same author:

Deploy Server
-
security - not tested
A
license - permissive license
-
quality - not tested

hybrid server

The server is able to function both locally and remotely, depending on the configuration or use case.

Enables comprehensive PostgreSQL database monitoring, analysis, and management through natural language queries. Provides performance insights, bloat analysis, vacuum monitoring, and intelligent maintenance recommendations across PostgreSQL versions 12-17.

  1. Overview
    1. Features
      1. 🔧 Advanced Capabilities
    2. Tool Usage Examples
      1. 📸 More Examples with Screenshots →
    3. ⭐ Quickstart (5 minutes)
      1. Flow Diagram of Quickstart/Tutorial
      2. 1. Environment Setup
      3. 2. Start Demo Containers
      4. 3. Access to OpenWebUI
      5. 4. Registering the Tool in OpenWebUI
      6. 5. Complete!
    4. (NOTE) Sample Test Data Overview
      1. Tool Compatibility Matrix
        1. 🟢 Extension-Independent Tools (No Extensions Required)
        2. 🚀 Version-Aware Tools (Auto-Adapting)
        3. 🟡 Extension-Dependent Tools (Extensions Required)
      2. Usage Examples
        1. Claude Desktop Integration
      3. Installation
        1. From PyPI (Recommended)
        2. From Source
      4. MCP Configuration
        1. Claude Desktop Configuration
        2. Run MCP-Server as Standalon
        3. (Option) Configure Multiple PostgreSQL Instances
      5. CLI Arguments
        1. Environment Variables
          1. Prerequisites
            1. Required PostgreSQL Extensions
            2. Minimum Requirements
            3. Required PostgreSQL Configuration
          2. RDS/Aurora Compatibility
            1. Example Queries
              1. 🟢 Extension-Independent Tools (Always Available)
              2. 🚀 Version-Aware Tools (Auto-Adapting)
              3. 🟡 Extension-Dependent Tools
            2. Troubleshooting
              1. Connection Issues
              2. Extension Errors
              3. Configuration Issues
              4. Performance Issues
              5. Version Compatibility Issues
            3. Development
              1. Testing & Development
              2. Version Compatibility Testing
            4. Security Notes
              1. Contributing
                1. MCPO Swagger Docs
                  1. 🔐 Security & Authentication
                    1. Bearer Token Authentication
                  2. License
                    1. ⭐ Other Projects

                      Related MCP Servers

                      • A
                        security
                        A
                        license
                        A
                        quality
                        Facilitates management and optimization of PostgreSQL databases, offering analysis, setup guidance, and debugging, while ensuring secure and efficient database operations.
                        Last updated -
                        3
                        1
                        16
                        AGPL 3.0
                        • Linux
                        • Apple
                      • A
                        security
                        A
                        license
                        A
                        quality
                        Enables AI models to interact with PostgreSQL databases through a standardized interface, supporting operations like queries, table manipulation, and schema inspection.
                        Last updated -
                        6
                        382
                        6
                        MIT License
                      • -
                        security
                        F
                        license
                        -
                        quality
                        Provides read-only access to PostgreSQL databases, enabling LLMs to inspect database schemas and execute read-only SQL queries.
                        Last updated -
                        20,167
                        • Apple
                      • -
                        security
                        A
                        license
                        -
                        quality
                        An extended PostgreSQL management and analysis server that provides database professionals with tools for schema management, query optimization, performance monitoring, and health analysis through a collection of specialized functions.
                        Last updated -
                        3
                        MIT License

                      View all related MCP servers

                      MCP directory API

                      We provide all the information about MCP servers via our MCP API.

                      curl -X GET 'https://glama.ai/api/mcp/v1/servers/call518/MCP-PostgreSQL-Ops'

                      If you have feedback or need assistance with the MCP directory API, please join our Discord server