Can someone explain this code that uses PySpark RDD, GroupBy and Lambda?

后端 未结 0 380
礼貌的吻别
礼貌的吻别 2021-01-17 05:31
i=0; col=\'INFANT_ALIVE_AT_REPORT\'
agg = categorical_rdd \\
    .groupBy(lambda row: row[i])

The output when using print(col,agg.collect())

相关标签:
回答
  • 消灭零回复
提交回复
热议问题