NewCluster
Properties
num_workers
Option<i32>
If num_workers, number of worker nodes that this cluster must have. A cluster has one Spark driver and num_workers executors for a total of num_workers + 1 Spark nodes. When reading the properties of a cluster, this field reflects the desired number of workers rather than the actual current number of workers. For example, if a cluster is resized from 5 to 10 workers, this field immediately updates to reflect the target size of 10 workers, whereas the workers listed in spark_info
gradually increase from 5 to 10 as the new nodes are provisioned.
[optional]
autoscale
[optional]
spark_version
String
spark_conf
An arbitrary object where the object key is a configuration propery name and the value is a configuration property value.
[optional]
aws_attributes
[optional]
node_type_id
Option<String>
[optional]
data_security_mode
Option<String>
Data security mode decides what data governance model to use when accessing data from a cluster.
[optional]
single_user_name
Option<String>
Single user name if data_security_mode is SINGLE_USER
[optional]
driver_node_type_id
Option<String>
The node type of the Spark driver. This field is optional; if unset, the driver node type is set as the same value as node_type_id
defined above.
[optional]
ssh_public_keys
Option<Vec>
SSH public key contents that are added to each Spark node in this cluster. The corresponding private keys can be used to login with the user name ubuntu
on port 2200
. Up to 10 keys can be specified.
[optional]
custom_tags
[optional]
cluster_log_conf
[optional]
init_scripts
The configuration for storing init scripts. Any number of scripts can be specified. The scripts are executed sequentially in the order provided. If cluster_log_conf
is specified, init script logs are sent to <destination>/<cluster-id>/init_scripts
.
[optional]
spark_env_vars
An arbitrary object where the object key is an environment variable name and the value is an environment variable value.
[optional]
enable_elastic_disk
Option<bool>
[optional]
driver_instance_pool_id
Option<String>
[optional]
instance_pool_id
Option<String>
[optional]
policy_id
Option<String>
[optional]
enable_local_disk_encryption
Option<bool>
Determines whether encryption of disks locally attached to the cluster is enabled.
[optional]
docker_image
[optional]
runtime_engine
Option<String>
The type of runtime engine to use. If not specified, the runtime engine type is inferred based on the spark_version
value. Allowed values include: * PHOTON
: Use the Photon runtime engine type. * STANDARD
: Use the standard runtime engine type. This field is optional.
[optional]
Last updated