2.0.11

Spark

Package: flyteplugins.spark

Use this to configure a SparkContext for a your task. Task’s marked with this will automatically execute natively onto K8s as a distributed execution of spark

Parameters

class Spark(
    spark_conf: typing.Optional[typing.Dict[str, str]],
    hadoop_conf: typing.Optional[typing.Dict[str, str]],
    executor_path: typing.Optional[str],
    applications_path: typing.Optional[str],
    driver_pod: typing.Optional[flyte._pod.PodTemplate],
    executor_pod: typing.Optional[flyte._pod.PodTemplate],
)
Parameter Type Description
spark_conf typing.Optional[typing.Dict[str, str]] Spark configuration dictionary.
hadoop_conf typing.Optional[typing.Dict[str, str]] Hadoop configuration dictionary.
executor_path typing.Optional[str] Path to the Python binary for PySpark execution.
applications_path typing.Optional[str] Path to the main application file.
driver_pod typing.Optional[flyte._pod.PodTemplate] Pod template for the driver pod.
executor_pod typing.Optional[flyte._pod.PodTemplate] Pod template for the executor pods.