What is the best way to do schema validation of a complex nested json in pyspark in databricks. My current input is a dataframe with one of the columns as a json.
I am trying to use Jsonschema library to validate it using udf and it is making the code extremely slow.
Any help here would be appreciated