Store aggregate value of a PySpark dataframe column into a variable

前端 未结 6 836
故里飘歌
故里飘歌 2021-01-13 09:37

I am working with PySpark dataframes here. \"test1\" is my PySpark dataframe and event_date is a TimestampType. So when I try to get a distinct count of event_date, the resu

6条回答
  •  一生所求
    2021-01-13 10:20

    try this

    loop_cnt=test1.select('event_date').distinct().count()
    var = loop_cnt.collect()[0]
    

    Hope this helps

提交回复
热议问题