Skip to main content
Glama

dolphin-mcp

NPM Version License: MIT

MCP server for Dolphin semantic code search. Conforms to MCP spec.

Quick Start

No installation needed - use bunx:

bunx dolphin-mcp

Configuration

Continue.dev

Add to config.yaml:

mcpServers:
  - name: Dolphin-KB
    command: bunx
    args:
      - dolphin-mcp
    env:
      DOLPHIN_API_URL: "http://127.0.0.1:7777"
      # Optional: Performance optimization for parallel snippet fetching
      MAX_CONCURRENT_SNIPPET_FETCH: "8"
      SNIPPET_FETCH_TIMEOUT_MS: "2000"
      SNIPPET_FETCH_RETRY_ATTEMPTS: "1"

Claude Desktop

Add to claude_desktop_config.json:

{
  "mcpServers": {
    "dolphin-kb": {
      "command": "bunx",
      "args": ["dolphin-mcp"],
      "env": {
        "DOLPHIN_API_URL": "http://127.0.0.1:7777",
        "MAX_CONCURRENT_SNIPPET_FETCH": "8",
        "SNIPPET_FETCH_TIMEOUT_MS": "2000",
        "SNIPPET_FETCH_RETRY_ATTEMPTS": "1"
      }
    }
  }
}

Environment Variables

Variable

Default

Description

DOLPHIN_API_URL

http://127.0.0.1:7777

Dolphin API endpoint

LOG_LEVEL

info

Logging level (debug, info, warn, error)

Parallel Snippet Fetching Configuration

These variables control the performance optimization for parallel snippet fetching in search_knowledge:

Variable

Default

Description

Recommended Range

MAX_CONCURRENT_SNIPPET_FETCH

8

Maximum parallel snippet requests

4-12

SNIPPET_FETCH_TIMEOUT_MS

2000

Timeout per snippet request (ms)

1500-3000

SNIPPET_FETCH_RETRY_ATTEMPTS

1

Retry attempts for failed requests

0-3

Configuration Presets

Conservative (recommended for limited resources):

MAX_CONCURRENT_SNIPPET_FETCH=4
SNIPPET_FETCH_TIMEOUT_MS=1500
SNIPPET_FETCH_RETRY_ATTEMPTS=1

Recommended (balanced performance):

MAX_CONCURRENT_SNIPPET_FETCH=8
SNIPPET_FETCH_TIMEOUT_MS=2000
SNIPPET_FETCH_RETRY_ATTEMPTS=1

Performance (maximum throughput):

MAX_CONCURRENT_SNIPPET_FETCH=10
SNIPPET_FETCH_TIMEOUT_MS=3000
SNIPPET_FETCH_RETRY_ATTEMPTS=2

Available Tools

search_knowledge

Semantically query code and docs across indexed repositories and return ranked snippets with citations.

{
  "query": "string (required)",
  "repos": ["string"],
  "path_prefix": ["string"],
  "exclude_paths": ["string"],
  "exclude_patterns": ["string"],
  "top_k": "number (1-100)",
  "max_snippets": "number",
  "embed_model": "small | large",
  "score_cutoff": "number"
}

Filtering Options:

  • repos: Include only specific repositories

  • path_prefix: Include only paths matching these prefixes (e.g., ["src/", "lib/"])

  • exclude_paths: Exclude paths matching these prefixes (e.g., ["tests/", "node_modules/", "dist/"])

  • exclude_patterns: Exclude files matching glob patterns (e.g., ["*.test.ts", "*.config.json"])

Example:

{
  "query": "authentication logic",
  "repos": ["myapp"],
  "path_prefix": ["src/"],
  "exclude_paths": ["tests/"],
  "exclude_patterns": ["*.spec.ts", "*.mock.ts"]
}

fetch_chunk

Fetch a chunk by chunk_id and return fenced code with citation.

{
  "chunk_id": "string (required)"
}

fetch_lines

Fetch a file slice [start, end] inclusive from disk and return fenced code with citation.

{
  "repo": "string (required)",
  "path": "string (required)",
  "start": "number (required, 1-indexed)",
  "end": "number (required, inclusive)"
}

get_vector_store_info

Report namespaces, dims, limits, and approximate counts.

{}

open_in_editor

Compute a vscode://file URI for a repo path and optional position.

{
  "repo": "string (required)",
  "path": "string (required)",
  "line": "number (1-indexed)",
  "column": "number (1-indexed)"
}

Installation (Optional)

If you prefer installing globally:

bun install -g dolphin-mcp

Then use dolphin-mcp instead of bunx dolphin-mcp.

Requirements

License

MIT - see LICENSE file for details.

-
security - not tested
A
license - permissive license
-
quality - not tested

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/plasticbeachllc/dolphin-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server