Apache Spark OutOfMemoryError (HeapSpace)

前端 未结 0 611
小蘑菇
小蘑菇 2021-01-13 12:06

I have a dataset with ~5M rows x 20 columns, containing a groupID and a rowID. My goal is to check whether (some) columns contain more than a fixed fraction (say, 50%) of mi

相关标签:
回答
  • 消灭零回复
提交回复
热议问题