wb resource create dataproc-cluster
Categories:
Name
wb-resource-create-dataproc-cluster - Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster
Synopsis
wb resource create dataproc-cluster [--quiet] [--autoscaling-policy=<autoscalingPolicy>] [--bucket=<configBucket>] [--cluster-id=<clusterId>] [--description=<description>] [--format=<format>] [--idle-delete-ttl=<idleDeleteTtl>] [--image-version=<imageVersion>] [--region=<region>] [--software-framework=<softwareFrameworkType _ >] [--temp-bucket=<tempBucket>] [--workspace=<id>] [--components=<components>[, <components>...]]... [--initialization-actions=<initializationAct_ _ ions>[,<initializationActions>...]]... [-M=<String=String>[, <String=String>...]]... [--properties=<String=String>[, <String=String>...]]... (--id=<id>) [[--manager-machine-type=<machineType>] [--manager-image-uri=<imageUri>] [[--manager-accelerator-type=<type>] [--manager-accelerator-count=<count>]] [[--manager-boot-disk-type=<bootDiskType>] [--manager-boot-disk-size=<bootDiskSizeGb>] [--manager-num-local-ssds=<numLocalSsds>] [--manager-local-ssd-interface=<localSsdInte_ _ rface>]]] [[--num-workers=<numNodes>] [--worker-machine-type=<machineType>] [--worker-image-uri=<imageUri>] [[--worker-accelerator-type=<type>] [--worker-accelerator-count=<count>]] [[--worker-boot-disk-type=<bootDiskType>] [--worker-boot-disk-size=<bootDiskSizeGb>] [--worker-num-local-ssds=<numLocalSsds>] [--worker-local-ssd-interface=<localSsdInter_ _ face>]]] [[--num-secondary-workers=<numNodes>] [--secondary-worker-machine-type=<machineTyp_ _ e>] [--secondary-worker-image-uri=<imageUri>] [--secondary-worker-type=<type>] [[--secondary-worker-accelerator-type=<type>_ ] [--secondary-worker-accelerator-count=<count _ >]] [[--secondary-worker-boot-disk-type=<bootDis_ _ kType>] [--secondary-worker-boot-disk-size=<bootDisk_ _ SizeGb>] [--secondary-worker-num-local-ssds=<numLocal_ _ Ssds>] [--secondary-worker-local-ssd-interface=<loc_ _ alSsdInterface>_]]]
Description
Add a controlled GCP Dataproc cluster resource with Jupyter. For a detailed explanation of parameters, see https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters#Cluster
Options
-
--id=<id>
ID of the resource, scoped to the workspace. Only use letters, numbers, dashes, and underscores. -
--id=<id>
ID of the resource, scoped to the workspace. Only use letters, numbers, dashes, and underscores. -
--description=<description>
Description of the resource. -
--workspace=<id>
Workspace id to use for this command only. -
--format=<format>
Set the format for printing command output: JSON, TEXT. Defaults to the config format property.Default: null
-
--quiet
Suppress interactive prompt. -
--cluster-id=<clusterId>
The unique name to give to the dataproc cluster. Cannot be changed later. The instance name must be 1 to 52 characters long and contain only lowercase letters, numeric characters, and dashes. The first character must be a lowercase letter and the last character cannot be a dash. If not specified, a value will be auto-generated for you. -
--region=<region>
The Google Cloud region of the cluster. -
--image-version=<imageVersion>
The dataproc cluster image version containing versions of its software components. See https://cloud.google.com/dataproc/docs/concepts/versioning/dataproc-version-clusters for the full list of image versions and their bundled software components. -
--initialization-actions=<initializationActions>[,<initializationActions>...]
A comma separated list of initialization scripts to run during cluster creation.The path must be a URL or Cloud Storage path, e.g. 'gs://path-to-file/file-name'. -
--components=<components>[,<components>...]
Comma-separated list of components. -
--properties=<String=String>[,<String=String>...]
Properties in the format key=value. -
--software-framework=<softwareFrameworkType>
Software framework for the cluster. Available frameworks are: NONE, HAIL.Default: NONE
-
--bucket=<configBucket>
Resource name of the cluster staging bucket. If not specified, a default staging bucket will be created. -
--temp-bucket=<tempBucket>
Resource name of the cluster temp bucket. If not specified, a default temp bucket will be created. -
--autoscaling-policy=<autoscalingPolicy>
Autoscaling policy id to attach to the cluster. -
-M, --metadata=<String=String>[,<String=String>...]
Custom metadata to apply to this cluster.specify multiple metadata in the format of --metadata="key1=value1" -Mkey2=value2.
It allows multiple metadata entries split by "," like --metadata=key1=value1,key2=value2
By default, set Workbench CLI server terra-cli-server=[CLI_SERVER_ID]
and the Workbench workspace id (terra-workspace-id=[WORKSPACE_ID]).
-
--idle-delete-ttl=<idleDeleteTtl>
Time-to-live after which the resource becomes idle and is deleted.
Manager node configurations
-
--manager-machine-type=<machineType>
The machine type of the manager node.Default: n2-standard-2
-
--manager-image-uri=<imageUri>
The image URI for the manager node. -
--manager-accelerator-type=<type>
The type of accelerator for the manager. -
--manager-accelerator-count=<count>
The count of accelerators for the manager.Default: 0
-
--manager-boot-disk-type=<bootDiskType>
The type of boot disk for the manager node. -
--manager-boot-disk-size=<bootDiskSizeGb>
The size of the boot disk in GB for the manager node.Default: 500
-
--manager-num-local-ssds=<numLocalSsds>
The number of local SSDs for the manager node.Default: 0
-
--manager-local-ssd-interface=<localSsdInterface>
The interface type of local SSDs for the manager node.Default: scsi
Worker node configurations
-
--num-workers=<numNodes>
The number of worker nodes.Default: 2
-
--worker-machine-type=<machineType>
The machine type of the worker node.Default: n2-standard-2
-
--worker-image-uri=<imageUri>
The image URI for the worker node. -
--worker-accelerator-type=<type>
The type of accelerator for the worker. -
--worker-accelerator-count=<count>
The count of accelerators for the worker.Default: 0
-
--worker-boot-disk-type=<bootDiskType>
The type of boot disk for the worker node. -
--worker-boot-disk-size=<bootDiskSizeGb>
The size of the boot disk in GB for the worker node.Default: 500
-
--worker-num-local-ssds=<numLocalSsds>
The number of local SSDs for the worker node.Default: 0
-
--worker-local-ssd-interface=<localSsdInterface>
The interface type of local SSDs for the worker node.Default: scsi
Secondary worker node configurations
-
--num-secondary-workers=<numNodes>
The number of secondary worker nodes.Default: 0
-
--secondary-worker-machine-type=<machineType>
The machine type of the secondary worker node.Default: n2-standard-2
-
--secondary-worker-image-uri=<imageUri>
The image URI for the secondary worker node. -
--secondary-worker-type=<type>
The type of the secondary worker. Valid values are preemptible, non-preemptible, and spot.Default: spot
-
--secondary-worker-accelerator-type=<type>
The type of accelerator for the secondary worker. -
--secondary-worker-accelerator-count=<count>
The count of accelerators for the secondary worker.Default: 0
-
--secondary-worker-boot-disk-type=<bootDiskType>
The type of boot disk for the secondary worker node. -
--secondary-worker-boot-disk-size=<bootDiskSizeGb>
The size of the boot disk in GB for the secondary worker node.Default: 500
-
--secondary-worker-num-local-ssds=<numLocalSsds>
The number of local SSDs for the secondary worker node.Default: 0
-
--secondary-worker-local-ssd-interface=<localSsdInterface>
The interface type of local SSDs for the secondary worker node.Default: scsi
Last Modified: 16 December 2024