Skip to main content
Glama

Server Configuration

Describes the environment variables required to run the server.

NameRequiredDescriptionDefault
DATAPROC_MCP_TRANSPORTNoCustom transport method for the MCP server
GOOGLE_APPLICATION_CREDENTIALSNoPath to service account JSON file for Google Cloud authentication

Tools

Functions exposed to the LLM to take actions

NameDescription
list_clusters

List Dataproc clusters in a project and region.

Args: project_id: Google Cloud project ID region: Dataproc region (e.g., us-central1)
create_cluster

Create a new Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name for the new cluster num_instances: Number of worker instances machine_type: Machine type for cluster nodes disk_size_gb: Boot disk size in GB image_version: Dataproc image version
delete_cluster

Delete a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name of the cluster to delete
get_cluster

Get details of a specific Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Name of the cluster
submit_job

Submit a job to a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Target cluster name job_type: Type of job (spark, pyspark, spark_sql, hive, pig, hadoop) main_file: Main file/class for the job args: Job arguments jar_files: JAR files to include properties: Job properties
list_jobs

List jobs in a Dataproc cluster.

Args: project_id: Google Cloud project ID region: Dataproc region cluster_name: Cluster name (optional) job_states: Filter by job states
get_job

Get details of a specific job.

Args: project_id: Google Cloud project ID region: Dataproc region job_id: Job ID
cancel_job

Cancel a running job.

Args: project_id: Google Cloud project ID region: Dataproc region job_id: Job ID to cancel
create_batch_job

Create a Dataproc batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Unique identifier for the batch job job_type: Type of batch job (spark, pyspark, spark_sql) main_file: Main file/class for the job args: Job arguments jar_files: JAR files to include properties: Job properties service_account: Service account email network_uri: Network URI subnetwork_uri: Subnetwork URI
list_batch_jobs

List Dataproc batch jobs.

Args: project_id: Google Cloud project ID region: Dataproc region page_size: Number of results per page
get_batch_job

Get details of a specific batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Batch job ID
delete_batch_job

Delete a batch job.

Args: project_id: Google Cloud project ID region: Dataproc region batch_id: Batch job ID to delete

Prompts

Interactive templates invoked by user choice

NameDescription

No prompts

Resources

Contextual data attached and managed by the client

NameDescription
get_clusters_resourceGet list of all Dataproc clusters.
get_jobs_resourceGet list of all Dataproc jobs.

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/warrenzhu25/dataproc-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server