Apache spark join with dynamic re-partitionion

前端 未结 1 593
长发绾君心
长发绾君心 2021-01-28 02:56

I\'m trying to do a fairly straightforward join on two tables, nothing complicated. Load both tables, do a join and update columns but it keeps throwing an exception.

I

1条回答
  •  梦毁少年i
    2021-01-28 03:48

    Your values for partitionColumn, upperBound, and lowerBound could cause this exact behavior if they aren't set correctly. For instance, if lowerBound == upperBound, then all of the data would be loaded into a single partition, regardless of numPartitions.

    The combination of these attributes determines which (or how many) records get loaded into your DataFrame partitions from your SQL database.

    0 讨论(0)
提交回复
热议问题