Server Configuration
Describes the environment variables required to run the server.
Name | Required | Description | Default |
---|---|---|---|
DATAPROC_MCP_TRANSPORT | No | Custom transport method for the MCP server | |
GOOGLE_APPLICATION_CREDENTIALS | No | Path to service account JSON file for Google Cloud authentication |
Schema
Prompts
Interactive templates invoked by user choice
Name | Description |
---|---|
No prompts |
Resources
Contextual data attached and managed by the client
Name | Description |
---|---|
get_clusters_resource | Get list of all Dataproc clusters. |
get_jobs_resource | Get list of all Dataproc jobs. |
Tools
Functions exposed to the LLM to take actions
Name | Description |
---|---|
list_clusters | List Dataproc clusters in a project and region. Args:
project_id: Google Cloud project ID
region: Dataproc region (e.g., us-central1) |
create_cluster | Create a new Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name for the new cluster
num_instances: Number of worker instances
machine_type: Machine type for cluster nodes
disk_size_gb: Boot disk size in GB
image_version: Dataproc image version |
delete_cluster | Delete a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name of the cluster to delete |
get_cluster | Get details of a specific Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Name of the cluster |
submit_job | Submit a job to a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Target cluster name
job_type: Type of job (spark, pyspark, spark_sql, hive, pig, hadoop)
main_file: Main file/class for the job
args: Job arguments
jar_files: JAR files to include
properties: Job properties |
list_jobs | List jobs in a Dataproc cluster. Args:
project_id: Google Cloud project ID
region: Dataproc region
cluster_name: Cluster name (optional)
job_states: Filter by job states |
get_job | Get details of a specific job. Args:
project_id: Google Cloud project ID
region: Dataproc region
job_id: Job ID |
cancel_job | Cancel a running job. Args:
project_id: Google Cloud project ID
region: Dataproc region
job_id: Job ID to cancel |
create_batch_job | Create a Dataproc batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Unique identifier for the batch job
job_type: Type of batch job (spark, pyspark, spark_sql)
main_file: Main file/class for the job
args: Job arguments
jar_files: JAR files to include
properties: Job properties
service_account: Service account email
network_uri: Network URI
subnetwork_uri: Subnetwork URI |
list_batch_jobs | List Dataproc batch jobs. Args:
project_id: Google Cloud project ID
region: Dataproc region
page_size: Number of results per page |
get_batch_job | Get details of a specific batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Batch job ID |
delete_batch_job | Delete a batch job. Args:
project_id: Google Cloud project ID
region: Dataproc region
batch_id: Batch job ID to delete |