Spark structured streaming - ways to lookup high volume non-static dataset?

时间秒杀一切 提交于 2021-01-29 10:32:34

问题


I wish to build a spark structured streaming job that does something like below(lookup a huge non-static dataset)

  1. Read from kafka(json record)
  2. For each json record

    • Get {user_key}
    • Read from huge Phoenix table(non-static) filter by {user_key}
    • Further DF transformations
    • Write to another phoenix table

How to lookup huge volume non-static dataset per kafka message?

来源:https://stackoverflow.com/questions/62421785/spark-structured-streaming-ways-to-lookup-high-volume-non-static-dataset

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!