How does spark perform filters and aggregations on datasets which don't fit in memory

后端 未结 0 546
醉话见心
醉话见心 2021-01-23 07:04

Say I have a 1TB parquet file stored in S3. It\'s written out in individual files of size 1GB.

I use spark SQL to filter the file down based on one of the columns and the

相关标签:
回答
  • 消灭零回复
提交回复
热议问题