Root Cause
When writing data to S3 during a multi-part upload , the maximum number of parts can be 10000
Due to this threshold the jobs fail during execution
Solution
To overcome this limitation we need to increase the size of the ‘part’ with the below advanced configuration.
The part size can be up to 5 GB
Set the below property as an additional configuration in table properties as shown in below image. This will increase the capacity for each 'part'
key: ingestion_spark_configs
value: spark.hadoop.fs.s3a.multipart.size=104857600
Please rerun the job after setting the above advanced configuration.
Affects Version:
IWX 3.2,4.2 and above