Skip to main content
Glama
128,620 tools. Last updated 2026-05-06 03:22

"Exploring Computer Usage" matching MCP tools:

  • Read-only. Returns your current APIHub credit balance (in microdollars and USD), total lifetime spending (microdollars and USD), and total completed request count. Requires a valid API key. Use before apihub_call or apihub_call_external to confirm sufficient funds for a paid request, or periodically to audit usage. Does not modify state, send payments, or call upstream APIs; for top-ups use apihub_topup.
    Connector
  • Runs a free one-off security scan of the given domain and returns its grade (A–F), scan timestamp, and up to three top-priority issues with a permalink to the full report on siteguardian.io. Use this when the user asks for a quick security check of a domain that is NOT yet under SiteGuardian monitoring, or when they want a fresh assessment before subscribing. Results are cached for two hours, so repeated calls about the same domain return the same snapshot and mark it with cached=True. Do NOT use this for domains already under monitoring by the user — call get_domain_status instead for the account-scoped view with framework tags. Do NOT use this to batch-scan many domains as a competitive-intelligence tool; per-source-IP and per-target rate limits bound usage. This tool does not require authentication.
    Connector
  • Read-only. Returns your current APIHub credit balance (in microdollars and USD), total lifetime spending (microdollars and USD), and total completed request count. Requires a valid API key. Use before apihub_call or apihub_call_external to confirm sufficient funds for a paid request, or periodically to audit usage. Does not modify state, send payments, or call upstream APIs; for top-ups use apihub_topup.
    Connector
  • Create an alert rule to monitor CPU, memory, or disk usage. When the metric crosses the threshold, a notification is sent via email and/or webhook. Max 10 rules per site. Requires: API key with write scope. Args: slug: Site identifier metric: "cpu", "memory", or "disk" (percentage-based) threshold: Threshold value 0-100 (e.g. 90 for 90%) operator: "gt" (greater than) or "lt" (less than). Default: "gt" severity: "warning" or "critical". Default: "warning" cooldown_minutes: Min minutes between repeated alerts. Default: 30 notify_email: Send email notification. Default: true notify_webhook: Optional webhook URL for POST notifications Returns: {"id": "uuid", "metric": "disk", "threshold": 90, ...}
    Connector
  • Fetch and convert a Microsoft Learn documentation webpage to markdown format. This tool retrieves the latest complete content of Microsoft documentation webpages including Azure, .NET, Microsoft 365, and other Microsoft technologies. ## When to Use This Tool - When search results provide incomplete information or truncated content - When you need complete step-by-step procedures or tutorials - When you need troubleshooting sections, prerequisites, or detailed explanations - When search results reference a specific page that seems highly relevant - For comprehensive guides that require full context ## Usage Pattern Use this tool AFTER microsoft_docs_search when you identify specific high-value pages that need complete content. The search tool gives you an overview; this tool gives you the complete picture. ## URL Requirements - The URL must be a valid HTML documentation webpage from the microsoft.com domain - Binary files (PDF, DOCX, images, etc.) are not supported ## Output Format markdown with headings, code blocks, tables, and links preserved.
    Connector
  • List available MCP tools and get detailed help. Use this tool to discover what tools are available and how to use them. Call without parameters to see all tools, or provide a tool name to get detailed help including parameters, examples, and related tools. Args: tool_name: Optional name of a specific tool to get detailed help for. Example: "search_funders", "get_funder_profile" Returns: If called without parameters: - server_name: Name of the MCP server - server_version: Current version - total_tools: Number of available tools - tier: Current access tier (free) - rate_limit: Rate limit information - tools: List of available tools with names, descriptions, and examples If called with tool_name: - tool: Detailed tool information including: - name: Tool name - description: What the tool does - parameters: List of parameters with types, descriptions, and examples - examples: Example usage - related_tools: Tools that work well together with this one Examples: list_tools() # See all available tools list_tools(tool_name="search_funders") # Get detailed help for search_funders list_tools(tool_name="get_funder_profile") # Get help for get_funder_profile
    Connector

Matching MCP Servers

  • A
    license
    A
    quality
    C
    maintenance
    An MCP server that retrieves GitHub Copilot usage metrics and seat assignment data across Enterprise, Organization, and Team levels. It allows users to monitor code completions, chat activity, and active user counts through integrated tools.
    Last updated
    5
    20
    MIT

Matching MCP Connectors

  • Baselight’s MCP server lets you seamlessly integrate your favourite applications with the Baselight platform. By connecting to the MCP server, you can browse, discover, and query 70,000+ datasets and 450+ billion rows directly from your preferred environment—whether you’re building, analysing, or exploring.

  • HiveCompute MCP Server — decentralized inference router for AI agents

  • BROWSING / DISCOVERY search — cities, neighbourhoods, or mixed venues near a location. Use this when the user is exploring a REGION rather than looking for a specific category. Supports population filtering ('cities > 100k'), distance/population sorting, and layer filtering (locality / neighbourhood / venue / address / street). For specific POI categories (gas, food, charging, etc.), use `search_places` instead.
    Connector
  • Get the caller's referral earnings, milestones, and free-call status. Requires Authorization: Bearer <api_key>, has no usage charge, and returns the current discount ledger without creating a new analysis. Example: call after a referral campaign to inspect earned credits. Use this when you need balances and milestones. Use get_referral_code instead when you only need the shareable token.
    Connector
  • Report the calling account's plan, key usage, and limits. Use this to introspect what the caller is allowed to do. Agents that hit rate limits or key-count caps can call this to explain the limit to the human and suggest upgrading if needed. Args: api_key: GeodesicAI API key (starts with gai_) Returns: plan: The user's current plan — one of pilot, trial, tier1, tier2, beta, enterprise plan_label: Human-readable plan name (e.g. "Personal Access", "Small Business") account_key_count: Number of account-level API keys currently issued account_key_limit: Maximum account keys allowed on this plan blueprint_count: Number of Blueprints owned by this user blueprint_limit: Maximum Blueprints allowed on this plan email: The user's email address (for reference in support) user_id: Stable user identifier trial_days_remaining: Days left on trial, if plan == "trial"; else null
    Connector
  • Find working SOURCE CODE examples from 37 indexed Senzing GitHub repositories. Indexes only source code files (.py, .java, .cs, .rs) and READMEs — NOT build files (Cargo.toml, pom.xml), data files (.jsonl, .csv), or project configuration. For sample data, use get_sample_data instead. Covers Python, Java, C#, and Rust SDK usage patterns including initialization, record ingestion, entity search, redo processing, and configuration. Also includes message queue consumers, REST API examples, and performance testing. Supports three modes: (1) Search: query for examples across all repos, (2) File listing: set repo and list_files=true to see all indexed source files in a repo, (3) File retrieval: set repo and file_path to get full source code. Use max_lines to limit large files. Returns GitHub raw URLs for file retrieval — fetch to read the source code.
    Connector
  • WORKFLOW: Step 1 of 4 - Start infrastructure design conversation Open an InsideOut V2 session and receive the assistant's intro message. The response contains a clean message from Riley (the infrastructure advisor) - display it to the user. ⚠️ Riley will ask questions - forward these to the user, DO NOT answer on their behalf. CRITICAL: This tool returns a session_id in the response metadata. You MUST use this session_id for ALL subsequent tool calls (convoreply, tfgenerate, tfdeploy, etc.). Use when the user mentions keywords like: 'setup my cloud infra', 'provision infrastructure', 'deploy infra', 'start insideout', 'use insideout', or similar intent to begin infra setup. OPTIONAL: project_context (string) - General tech stack summary so Riley can skip discovery questions and jump to recommendations. The agent should confirm this with the user before sending. Include whichever apply: language/framework, databases/services, container usage, existing IaC, CI/CD platform, cloud provider, Kubernetes usage, what the project does. Example: 'Next.js 14 + TypeScript, PostgreSQL, Redis, Docker Compose, deployed to AWS ECS, GitHub Actions CI/CD, ~50k MAU'. NEVER include credentials, secrets, API keys, PII, source code, or internal URLs/IPs -- only general metadata summaries useful to a cloud architect agent. IMPORTANT: source (string) - You MUST set this to identify which IDE/tool you are. Auto-detect from your environment: 'claude-code', 'codex', 'antigravity', 'kiro', 'vscode', 'web', 'mcp'. If unsure, use the name of your IDE/tool in lowercase. Do NOT omit this — it controls the 'Open {IDE}' button on the credential connect screen. OPTIONAL: github_username (string) - GitHub username for deploy commit attribution. Pre-populates the GitHub username field on the connect page. 💡 TIP: Examine workflow.usage prompt for more context on how to properly use these tools.
    Connector
  • View account info, pricing, entitlements, or list keys. Actions: "status" (default) → tier, quota, usage from /me/entitlements "pricing" → public pricing tiers (no auth required) "keys" → list user's API keys with per-key usage "usage" → alias for "keys" (per-key usage is shown there)
    Connector
  • Search for eSIM data packages by country. Returns up to 10 packages per page sorted by price. Use the page parameter to paginate. No auth required. Call get_business_context first to understand IP routing and package types. Package types: - "regular": Fixed data pool (e.g. 3GB for 30 days). Best for most travelers. - "daily": Data resets each day (e.g. 2GB/day for 5 days). Good for short trips with predictable daily usage. Top-up days are available. IP routing (important for Asia): - "breakout": Local IP in destination country. Best for streaming, banking, social media. ALWAYS recommend by default. - "hk": Hong Kong IP. Cheapest but TikTok app and Facebook app are BLOCKED. - "nonhk": Third-country IP (UK, Singapore). No HK restrictions but IP won't match destination.
    Connector
  • WORKFLOW: Step 1 of 4 - Start infrastructure design conversation Open an InsideOut V2 session and receive the assistant's intro message. The response contains a clean message from Riley (the infrastructure advisor) - display it to the user. ⚠️ Riley will ask questions - forward these to the user, DO NOT answer on their behalf. CRITICAL: This tool returns a session_id in the response metadata. You MUST use this session_id for ALL subsequent tool calls (convoreply, tfgenerate, tfdeploy, etc.). Use when the user mentions keywords like: 'setup my cloud infra', 'provision infrastructure', 'deploy infra', 'start insideout', 'use insideout', or similar intent to begin infra setup. OPTIONAL: project_context (string) - General tech stack summary so Riley can skip discovery questions and jump to recommendations. The agent should confirm this with the user before sending. Include whichever apply: language/framework, databases/services, container usage, existing IaC, CI/CD platform, cloud provider, Kubernetes usage, what the project does. Example: 'Next.js 14 + TypeScript, PostgreSQL, Redis, Docker Compose, deployed to AWS ECS, GitHub Actions CI/CD, ~50k MAU'. NEVER include credentials, secrets, API keys, PII, source code, or internal URLs/IPs -- only general metadata summaries useful to a cloud architect agent. IMPORTANT: source (string) - You MUST set this to identify which IDE/tool you are. Auto-detect from your environment: 'claude-code', 'codex', 'antigravity', 'kiro', 'vscode', 'web', 'mcp'. If unsure, use the name of your IDE/tool in lowercase. Do NOT omit this — it controls the 'Open {IDE}' button on the credential connect screen. OPTIONAL: github_username (string) - GitHub username for deploy commit attribution. Pre-populates the GitHub username field on the connect page. 💡 TIP: Examine workflow.usage prompt for more context on how to properly use these tools.
    Connector
  • Scaffold an api/<route>.js that takes a prompt and returns the model's response, using lib/llm.js so every call lands in the llm_calls table for free observability. Produces: - POST handler with requireAuth - body: { prompt, model? } - calls llm.run({ user_id, purpose, prompt, model, system }) - returns { text, usage, llm_call_id } Example: add_llm_endpoint({ route: 'summarize', purpose: 'summarize', system_prompt: 'Summarize the user\'s text in 2-3 sentences.' }) Note: The project must have lib/llm.js (it ships in tpl-prompt-playground; copy if needed).
    Connector
  • Export a generated image asset by session and asset ID. Returns the image inline as base64 along with metadata (format, dimensions, size). When running locally (stdio transport), you can optionally provide a destinationPath to save the image to disk. USAGE: After generating an image with generateImage, use the sessionId and assetId to export: exportImageAsset(sessionId="...", assetId="...") To save to disk (local/stdio only): exportImageAsset(sessionId="...", assetId="...", destinationPath="/Users/me/project/images/logo.png")
    Connector
  • Discover all knowledge bases you have access to. Returns collection names, descriptions, content types, stats, available operations, and usage examples for each collection. Call this first to understand what data is available before searching.
    Connector
  • Extraction du texte intégral d'une décision relevant de l'ordre administratif (Conseil d'État, TA, CAA). Usage strictement réservé aux identifiants normés issus des recherches administratives : `DCE_XXX_YYYYMMDD` (Conseil d'État), `DTA_XXX_YYYYMMDD` (TA), `DCAA_XXX_YYYYMMDD` (CAA). INCOMPATIBILITÉS MAJEURES : - Identifiants ArianeWeb `/Ariane_Web/AW_DCE/|XXXXXX` — procéder à une ré-indexation via `search_admin` pour obtenir un identifiant compatible. - Identifiants JURITEXT — rediriger vers `get_decision_judiciaire_libre` ou `get_decision_judiciaire`. - Identifiants CELEX `6XXXXCJXXXX` — rediriger vers `get_decision_cjue`. - Identifiants HUDOC `001-XXXXXX` — rediriger vers `get_decision_cedh`. Args: decision_id: identifiant de la décision (avec ou sans suffixe .xml) Returns: Dict comportant les métadonnées complètes, `text_segments` (liste des paragraphes) et `full_text` (texte intégral joint), ou None si la décision est introuvable.
    Connector
  • Collects user feedback on the provided response. **When to use this tool:** - After providing an analysis, a SQL query, or an important response - When you want to know if the response was helpful - Naturally suggest: "Was this response helpful? 👍 👎" **Ratings:** - 'positive': The response was helpful and accurate - 'negative': The response was not satisfactory - 'neutral': Neither satisfied nor dissatisfied **Categories (optional):** - 'accuracy': Was the response accurate? - 'relevance': Did the response address the question? - 'completeness': Was the response complete? - 'speed': Was the response time acceptable? - 'other': Other feedback **Feedback usage:** Feedback is used to improve future responses (RAG, analytics).
    Connector