-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SUPPORT] Metadata compaction periodically fails/hangs #12261
Comments
@liiang-huang Can you collect more stats from metadata table? I see executors getting lost. |
@ad1happy2go Yes, the reason is
There are 229 objects in |
Could you please share hudi writer configuration and spark configuration? It is possible to provide timeline to check our end. |
@liiang-huang Were you able to get it resolved as i see no update here? Can you please share insights on what was the issue. |
Hi @ad1happy2go @rangareddy This is still happening sometimes, there are not much logs other than heartbeat SaveIntoDataSourceCommand org.apache.hudi.Spark31DefaultSource@4fed7de0, Map(hoodie.payload.ordering.field -> daas_internal_ts, hoodie.datasource.hive_sync.database -> pay2bronze, hoodie.datasource.hive_sync.mode -> GLUE, hoodie.filesystem.view.incr.timeline.sync.enable -> false, hoodie.schema.on.read.enable -> true, path -> s3://pay2-datalake-prod-standard/datasets/bronze/paylite-payment-db1-w-slave-20220523-aurora-prod/paylite_payment/sub_payments-1661338391, hoodie.compact.inline.max.delta.seconds -> 3600, hoodie.datasource.write.precombine.field -> daas_internal_ts, hoodie.datasource.write.payload.class -> jp.ne.paypay.daas.data.util.DaaSOverwritePayload, hoodie.compact.inline.trigger.strategy -> NUM_OR_TIME, hoodie.cleaner.fileversions.retained -> 6, hoodie.datasource.meta.sync.enable -> true, hoodie.write.commit.callback.on -> true, hoodie.metadata.enable -> true, hoodie.datasource.hive_sync.table -> paylite_payment_sub_payments, hoodie.datasource.meta_sync.condition.sync -> false, hoodie.write.commit.callback.class -> jp.ne.paypay.daas.data.metrics.DaasHudiWriteCommitCallback, hoodie.index.type -> BLOOM, hoodie.datasource.write.operation -> upsert, hoodie.rollback.using.markers -> false, hoodie.metrics.reporter.type -> CLOUDWATCH, hoodie.datasource.write.recordkey.field -> id, hoodie.table.name -> paylite_payment_sub_payments, hoodie.datasource.write.table.type -> MERGE_ON_READ, hoodie.datasource.write.hive_style_partitioning -> true, hoodie.datasource.write.table.name -> paylite_payment_sub_payments, hoodie.cleaner.policy -> KEEP_LATEST_FILE_VERSIONS, hoodie.write.markers.type -> DIRECT, hoodie.compact.inline -> true, hoodie.datasource.compaction.async.enable -> false, hoodie.metrics.on -> true, hoodie.upsert.shuffle.parallelism -> 200, hoodie.meta.sync.client.tool.class -> org.apache.hudi.aws.sync.AwsGlueCatalogSyncTool, hoodie.datasource.write.partitionpath.field -> daas_date, hoodie.compact.inline.max.delta.commits -> 1, hoodie.payload.event.time.field -> daas_internal_ts Spark config:
|
Describe the problem you faced
Hi Hudi community, I have a glue job that is ingesting data to a Hudi MOR table. However, this job periodically failed in the below stage



Could you help to investigate this issue? I have went through this issue, but doesn't seem like the same issue. When I deleted the requested/inflight deltacommit, also tried to increase resources, the errors still persisted. Thanks!
Environment Description
Hudi version : 0.13.1
Spark version : 3.1
Storage (HDFS/S3/GCS..) : S3
Additional context
Add any other context about the problem here.
Stacktrace
The text was updated successfully, but these errors were encountered: