Why does memory usage of spark worker increases with time?
问题 I have a Spark Streaming application running which uses mapWithState function to track state of RDD. The application runs fine for few minutes but then crashes with org.apache.spark.shuffle.MetadataFetchFailedException: Missing an output location for shuffle 373 I observed that Memory usage of Spark application increases over time linearly even though i have set the timeout for mapWithStateRDD. Please see the code snippet below and memory usage - val completedSess = sessionLines .mapWithState