Multiple aggregations in Spark Structured Streaming

前端 未结 8 1733
难免孤独
难免孤独 2020-12-09 04:09

I would like to do multiple aggregations in Spark Structured Streaming.

Something like this:

  • Read a stream of input files (from a folder)
  • Per
相关标签:
8条回答
  • 2020-12-09 04:39

    This is not supported, but there are other ways also. Like performing single aggregation and saving it to kafka. Read it from kafka and apply aggregation again. This has worked for me.

    0 讨论(0)
  • 2020-12-09 04:39

    For spark 2.2 and above (not sure about earlier version), if you can design the aggregation to use flatMapGroupWithState with append mode, you can do as many aggregations as you want. The restriction is mentioned here Spark structured streaming - Output mode

    0 讨论(0)
提交回复
热议问题