overwrite hive partitions using spark

后端 未结 4 1171
北荒
北荒 2021-02-05 21:44

I am working with AWS and I have workflows that use Spark and Hive. My data is partitioned by the date, so everyday I have a new partition in my S3 storage. My problem is when

4条回答
  •  花落未央
    2021-02-05 22:18

    I would suggest to run sql using sparksession. you can run " insert overwrite partition query" by selecting the columns from existing dataset. this solution will surely overwrites partition only.

提交回复
热议问题