Memory profiling on Google Cloud Dataflow

前端 未结 1 1551
面向向阳花
面向向阳花 2020-12-04 00:18

What would be the best way to debug memory issues of a dataflow job?

My job was failing with a GC OOM error, but when I profile it locally I cannot reproduce the exa

相关标签:
1条回答
  • 2020-12-04 00:41

    Please use the option --dumpHeapOnOOM and --saveHeapDumpsToGcsPath (see docs).

    This will only help if one of your workers actually OOMs. Additionally you can try running jmap -dump PID on the harness process on the worker to obtain a heap dump at runtime if it's not OOMing but if you observe high memory usage nevertheless.

    0 讨论(0)
提交回复
热议问题