How to efficiently write multiple pyarrow tables (>1,000 tables) to a partitioned parquet dataset?

后端 未结 0 734
不知归路
不知归路 2021-02-05 10:59

I have some big files (around 7,000 in total, 4GB each) in other formats that I want to store into a partitioned (hive) directory using the pyarrow.parquet.write_to_dataset() fo

相关标签:
回答
  • 消灭零回复
提交回复
热议问题