Spark Failure : Caused by: org.apache.spark.shuffle.FetchFailedException: Too large frame: 5454002341

后端 未结 5 1127
遥遥无期
遥遥无期 2021-01-06 07:51

I am generating a hierarchy for a table determining the parent child.

Below is the configuration used, even after getting the error with regards to the too large fra

5条回答
  •  鱼传尺愫
    2021-01-06 08:38

    This means that size of your dataset partitions is enormous. You need to repartition your dataset to more partitions.

    you can do this using,

    df.repartition(n)
    

    Here, n is dependent on the size of your dataset.

提交回复
热议问题