gcloud beta dataproc workflow-templates add-job - add Dataproc jobs to workflow template
gcloud beta dataproc workflow-templates add-job COMMAND [GCLOUD_WIDE_FLAG ...]
To add a Hadoop MapReduce job, run:
$ gcloud beta dataproc workflow-templates add-job hadoop \ --workflow-template my_template --jar my_jar.jar -- arg1 arg2
To add a Spark Scala or Java job, run:
$ gcloud beta dataproc workflow-templates add-job spark \ --workflow-template my_template --jar my_jar.jar -- arg1 arg2
To add a PySpark job, run:
$ gcloud beta dataproc workflow-templates add-job pyspark \ --workflow-template my_template my_script.py -- arg1 arg2
To add a Spark SQL job, run:
$ gcloud beta dataproc workflow-templates add-job spark-sql \ --workflow-template my_template --file my_queries.q
To add a Pig job, run:
$ gcloud beta dataproc workflow-templates add-job pig \ --workflow-template my_template --file my_script.pig
To add a Hive job, run:
$ gcloud beta dataproc workflow-templates add-job hive \ --workflow-template my_template --file my_queries.q
These flags are available to all commands: --help.
Run $ gcloud help for details.
COMMAND is one of the following:
- hadoop
(BETA) Add a hadoop job to the workflow template.
- hive
(BETA) Add a Hive job to the workflow template.
- pig
(BETA) Add a Pig job to the workflow template.
- presto
(BETA) Add a Presto job to the workflow template.
- pyspark
(BETA) Add a PySpark job to the workflow template.
- spark
(BETA) Add a Spark job to the workflow template.
- spark-r
(BETA) Add a SparkR job to the workflow template.
- spark-sql
(BETA) Add a SparkSql job to the workflow template.
- trino
(BETA) Add a Trino job to the workflow template.
This command is currently in beta and might change without notice. These variants are also available:
$ gcloud dataproc workflow-templates add-job
$ gcloud alpha dataproc workflow-templates add-job