Skip to content Skip to sidebar Skip to footer

Pyspark Outofmemoryerrors When Performing Many Dataframe Joins

There's many posts about this issue, but none have answered my question. I'm running into OutOfMemoryErrors in PySpark while attempting to join many different dataframes together.

Solution 1:

I faced similar issue and it worked with: Spark Submit:

spark-submit --driver-memory3g\
            --executor-memory14g\
            *.py

Code:

sc = SparkContext().getOrCreate()

Post a Comment for "Pyspark Outofmemoryerrors When Performing Many Dataframe Joins"