gcloud_dataproc_workflow-templates_add-job (1)
NAME
- gcloud dataproc workflow-templates add-job - add Google Cloud Dataproc jobs to workflow template
SYNOPSIS
-
gcloud dataproc workflow-templates add-job COMMAND [--region=REGION] [GCLOUD_WIDE_FLAG ...]
FLAGS
-
- --region=REGION
-
- Cloud Dataproc region to use. Each Cloud Dataproc region constitutes an independent resource namespace constrained to deploying instances into Compute Engine zones inside the region. The default value of global is a special multi-region namespace which is capable of deploying instances into all Compute Engine zones globally, and is disjoint from other Cloud Dataproc regions. Overrides the default dataproc/region property value for this command invocation.
- Cloud Dataproc region to use. Each Cloud Dataproc region constitutes an independent resource namespace constrained to deploying instances into Compute Engine zones inside the region. The default value of global is a special multi-region namespace which is capable of deploying instances into all Compute Engine zones globally, and is disjoint from other Cloud Dataproc regions. Overrides the default dataproc/region property value for this command invocation.
GCLOUD WIDE FLAGS
These flags are available to all commands: --account, --configuration, --flags-file, --flatten, --format, --help, --log-http, --project, --quiet, --trace-token, --user-output-enabled, --verbosity. Run $ gcloud help for details.
COMMANDS
COMMAND is one of the following:
-
- hadoop
-
Add a hadoop job to the workflow template.
- hive
-
Add a Hive job to the workflow template.
- pig
-
Add a Pig job to the workflow template.
- pyspark
-
Add a PySpark job to the workflow template.
- spark
-
Add a Spark job to the workflow template.
- spark-sql
-
Add a SparkSql job to the workflow template.
EXAMPLES
To add a Hadoop MapReduce job, run:
-
$ gcloud dataproc workflow-templates add-job hadoop \
--workflow-template my_template --jar my_jar.jar \
-- arg1 arg2
To add a Spark Scala or Java job, run:
-
$ gcloud dataproc workflow-templates add-job spark \
--workflow-template my_template --jar my_jar.jar \
-- arg1 arg2
To add a PySpark job, run:
-
$ gcloud dataproc workflow-templates add-job pyspark \
--workflow-template my_template my_script.py \
-- arg1 arg2
To add a Spark SQL job, run:
-
$ gcloud dataproc workflow-templates add-job spark-sql \
--workflow-template my_template --file my_queries.q
To add a Pig job, run:
-
$ gcloud dataproc workflow-templates add-job pig \
--workflow-template my_template --file my_script.pig
To add a Hive job, run:
-
$ gcloud dataproc workflow-templates add-job hive \
--workflow-template my_template --file my_queries.q
NOTES
These variants are also available:
- $ gcloud alpha dataproc workflow-templates add-job $ gcloud beta dataproc workflow-templates add-job