count of the table : 1125089 for october month data , So I am optimizing the table.
optimize table where batchday >="2022-10-01" and batchday<="2022-10-31"
I am getting error like : GC overhead limit exceeded
at org.apache.spark.unsafe.types.UTF8String.fromBytes(UTF8String.java:136)
I have increased my driver and executor worker nodes from 2 worker nodes to 10 worker nodes with memory size increased from 32 GB .
When i run other optimize on other batchday i am not finding any issue.
Could you tell me why UTFString.fromBytes exception comes during optimizing a partition of data?