Databricks OutOfMemory error on code that previously worked without issue
I have a notebook in Azure Databricks that does some transformations on a bronze tier table and inserts the transformed data into a silver tier table. This notebook is used to do an initial load of the data from our existing system into our new datalake. Specifically, it converts a list of columns into individual child rows in another table (sort of a key:value pair type of table). I’ve been running this code for weeks without issue: same code, same data, same cluster configuration. Suddenly, I am now consistently receiving an OutOfMemory
error when I execute this particular section of the notebook. The data is not overly large, and the table I’m inserting into is only 6 columns (3 integer type and 3 string type).