dagster_pyspark.
pyspark_resource
ResourceDefinition[source]¶This resource provides access to a PySpark SparkSession for executing PySpark code within Dagster.
Example
@solid(required_resource_keys={"pyspark"})
def my_solid(context):
spark_session = context.pyspark.spark_session
dataframe = spark_session.read.json("examples/src/main/resources/people.json")
my_pyspark_resource = pyspark_resource.configured(
{"spark_conf": {"spark.executor.memory": "2g"}}
)
@pipeline(mode_defs=[ModeDefinition(resource_defs={"pyspark"})])
def my_pipeline():
my_solid()