Server Configuration
Describes the environment variables required to run the server.
| Name | Required | Description | Default |
|---|---|---|---|
| DATAPROC_MCP_TRANSPORT | No | Custom transport method for the MCP server | |
| GOOGLE_APPLICATION_CREDENTIALS | No | Path to service account JSON file for Google Cloud authentication |
Schema
Prompts
Interactive templates invoked by user choice
| Name | Description |
|---|---|
No prompts | |
Resources
Contextual data attached and managed by the client
| Name | Description |
|---|---|
| get_clusters_resource | Get list of all Dataproc clusters. |
| get_jobs_resource | Get list of all Dataproc jobs. |
Tools
Functions exposed to the LLM to take actions
| Name | Description |
|---|---|
| list_clusters | List Dataproc clusters in a project and region. Args:
project_id: Google Cloud project ID
region: Dataproc region (e.g., us-central1) |
| create_cluster | Create a new Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name for the new cluster
num_instances: Number of worker instances
machine_type: Machine type for cluster nodes
disk_size_gb: Boot disk size in GB
image_version: Dataproc image version |
| delete_cluster | Delete a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name of the cluster to delete |
| get_cluster | Get details of a specific Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name of the cluster |
| submit_job | Submit a job to a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Target cluster name
job_type: Type of job (spark, pyspark, spark_sql, hive, pig, hadoop)
main_file: Main file/class for the job
args: Job arguments
jar_files: JAR files to include
properties: Job properties |
| list_jobs | List jobs in a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Cluster name (optional)
job_states: Filter by job states |
| get_job | Get details of a specific job. Args:
project_id: Google Cloud project ID
region: Dataproc region
job_id: Job ID |
| cancel_job | Cancel a running job. Args:
project_id: Google Cloud project ID
region: Dataproc region
job_id: Job ID to cancel |
| create_batch_job | Create a Dataproc batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Unique identifier for the batch job
job_type: Type of batch job (spark, pyspark, spark_sql)
main_file: Main file/class for the job
args: Job arguments
jar_files: JAR files to include
properties: Job properties
service_account: Service account email
network_uri: Network URI
subnetwork_uri: Subnetwork URI |
| list_batch_jobs | List Dataproc batch jobs. Args:
project_id: Google Cloud project ID
region: Dataproc region
page_size: Number of results per page |
| get_batch_job | Get details of a specific batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Batch job ID |
| delete_batch_job | Delete a batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Batch job ID to delete |