gcloud dataproc workflow-templates add-job - add Dataproc jobs to workflow template
gcloud dataproc workflow-templates add-job COMMAND [GCLOUD_WIDE_FLAG ...]
To add a Hadoop MapReduce job, run:
$ gcloud dataproc workflow-templates add-job hadoop \ --workflow-template my_template --jar my_jar.jar -- arg1 arg2
To add a Spark Scala or Java job, run:
$ gcloud dataproc workflow-templates add-job spark \ --workflow-template my_template --jar my_jar.jar -- arg1 arg2
To add a PySpark job, run:
$ gcloud dataproc workflow-templates add-job pyspark \ --workflow-template my_template my_script.py -- arg1 arg2
To add a Spark SQL job, run:
$ gcloud dataproc workflow-templates add-job spark-sql \ --workflow-template my_template --file my_queries.q
To add a Pig job, run:
$ gcloud dataproc workflow-templates add-job pig \ --workflow-template my_template --file my_script.pig
To add a Hive job, run:
$ gcloud dataproc workflow-templates add-job hive \ --workflow-template my_template --file my_queries.q
These flags are available to all commands: --help.
Run $ gcloud help for details.
COMMAND is one of the following:
- hadoop
Add a hadoop job to the workflow template.
- hive
Add a Hive job to the workflow template.
- pig
Add a Pig job to the workflow template.
- presto
Add a Presto job to the workflow template.
- pyspark
Add a PySpark job to the workflow template.
- spark
Add a Spark job to the workflow template.
- spark-r
Add a SparkR job to the workflow template.
- spark-sql
Add a SparkSql job to the workflow template.
- trino
Add a Trino job to the workflow template.
These variants are also available:
$ gcloud alpha dataproc workflow-templates add-job
$ gcloud beta dataproc workflow-templates add-job