Spark K8S Operator
The Kubernetes Operator for Apache Spark aims to make specifying and running Spark applications as easy and idiomatic as running other workloads on Kubernetes. It uses Kubernetes custom resources for specifying, running, and surfacing status of Spark applications. For a complete reference of the custom resource definitions, please refer to the API Definition. For details on its design, please refer to the design doc. It requires Spark 2.3 and above that supports Kubernetes as a native scheduler backend.
For complete project documentation, please visit the Spark K8S Operator documentation site.
Usage
Spark K8S Operator can be deployed by enabling the add-on via the following.
Basic Example
enable_spark_k8s_operator = true
Advanced Example
enable_spark_k8s_operator = true
# Optional Map value
# NOTE: This block requires passing the helm values.yaml
spark_k8s_operator_helm_config = {
name = "spark-operator"
chart = "spark-operator"
repository = "https://googlecloudplatform.github.io/spark-on-k8s-operator"
version = "1.1.19"
namespace = "spark-k8s-operator"
timeout = "1200"
create_namespace = true
values = [templatefile("${path.module}/values.yaml", {})]
}
GitOps Configuration
The following properties are made available for use when managing the add-on via GitOps
sparkK8sOperator = {
enable = true
}