Memory Leak in pyspark Driver Node in databricks
I’m encountering a memory issue in my PySpark application in databricks. where the memory usage on the driver node keeps increasing over time until it eventually crashes due to an out-of-memory (OOM) error. The problem occurs when I execute the following function in a loop for each image in a folder: