Submit and manage batch jobs on Google Cloud Dataproc. Define job types (Spark, PySpark, Spark SQL), include main files, JARs, arguments, and configure properties for efficient job execution.
Set up a new Databricks cluster with custom parameters like cluster name, Spark version, node type, worker count, and auto-termination settings using Databricks MCP Server.
Submit a job to a Dataproc cluster by specifying project ID, region, cluster name, job type, main file, and optional arguments, JAR files, and properties. Supports Spark, PySpark, Hive, Pig, and Hadoop job types.
Create dynamic, interactive charts using Apache ECharts by providing customizable configurations. Export visualizations as PNG, SVG, or raw option formats for seamless integration into web applications.
Enables comprehensive analysis of Apache Spark event logs from S3, HTTP, or local sources, providing performance metrics, resource monitoring, shuffle analysis, and automated optimization recommendations with interactive HTML reports.