Skip to main content
Glama

Dataproc MCP Server

by warrenzhu25

Server Configuration

Describes the environment variables required to run the server.

NameRequiredDescriptionDefault
DATAPROC_MCP_TRANSPORTNoCustom transport method for the MCP server
GOOGLE_APPLICATION_CREDENTIALSNoPath to service account JSON file for Google Cloud authentication

Schema

Prompts

Interactive templates invoked by user choice

NameDescription

No prompts

Resources

Contextual data attached and managed by the client

NameDescription
get_clusters_resourceGet list of all Dataproc clusters.
get_jobs_resourceGet list of all Dataproc jobs.

Tools

Functions exposed to the LLM to take actions

NameDescription
list_clusters

List Dataproc clusters in a project and region.

Args: project_id: Google Cloud project ID region: Dataproc region (e.g., us-central1)
create_cluster

Create a new Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name for the new cluster num_instances: Number of worker instances machine_type: Machine type for cluster nodes disk_size_gb: Boot disk size in GB image_version: Dataproc image version
delete_cluster

Delete a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name of the cluster to delete
get_cluster

Get details of a specific Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name of the cluster
submit_job

Submit a job to a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Target cluster name job_type: Type of job (spark, pyspark, spark_sql, hive, pig, hadoop) main_file: Main file/class for the job args: Job arguments jar_files: JAR files to include properties: Job properties
list_jobs

List jobs in a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Cluster name (optional) job_states: Filter by job states
get_job

Get details of a specific job.

Args: project_id: Google Cloud project ID region: Dataproc region job_id: Job ID
cancel_job

Cancel a running job.

Args: project_id: Google Cloud project ID region: Dataproc region job_id: Job ID to cancel
create_batch_job

Create a Dataproc batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Unique identifier for the batch job job_type: Type of batch job (spark, pyspark, spark_sql) main_file: Main file/class for the job args: Job arguments jar_files: JAR files to include properties: Job properties service_account: Service account email network_uri: Network URI subnetwork_uri: Subnetwork URI
list_batch_jobs

List Dataproc batch jobs.

Args: project_id: Google Cloud project ID region: Dataproc region page_size: Number of results per page
get_batch_job

Get details of a specific batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Batch job ID
delete_batch_job

Delete a batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Batch job ID to delete

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/warrenzhu25/dataproc-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server