Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-03 Thread via GitHub


gsudhanshu closed issue #10432: [SUPPORT] Failed to create marker file 
Exception when trying to write data on Hudi
URL: https://github.com/apache/hudi/issues/10432


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-03 Thread via GitHub


gsudhanshu commented on issue #10432:
URL: https://github.com/apache/hudi/issues/10432#issuecomment-1875003177

   Thanks


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-03 Thread via GitHub


gsudhanshu commented on issue #10432:
URL: https://github.com/apache/hudi/issues/10432#issuecomment-1875003031

   > @gsudhanshu Can you try disabling the timeline server?
   > 
   > hoodie.write.markers.type= 'direct', hoodie.embed.timeline.server= 'false'
   > 
   > We had a silmilar issue (#4230) before which we had fixed. I see you using 
0.14.0, So adding @yihua in case he have more insights.
   
   @ad1happy2go 
   Thanks for reply.. these configurations solved the issue for me. We can 
close this ticker. I had to set these in hudi options


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-02 Thread via GitHub


ad1happy2go commented on issue #10432:
URL: https://github.com/apache/hudi/issues/10432#issuecomment-1874904475

   @gsudhanshu After setting these, it should not use timeline server. Do you 
still see references of TimelineServerBasedWriteMarkers in the stack trace?
   
   can you paste the new stack trace please


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-02 Thread via GitHub


gsudhanshu commented on issue #10432:
URL: https://github.com/apache/hudi/issues/10432#issuecomment-1873985101

   @ad1happy2go thanks for your reply.
   
   I have added as following:
   
![image](https://github.com/apache/hudi/assets/45429552/a3516567-ba3c-47ec-aaff-56489d025cf8)
   but still getting same error.
   
   should I add these in hudi_options?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2024-01-02 Thread via GitHub


ad1happy2go commented on issue #10432:
URL: https://github.com/apache/hudi/issues/10432#issuecomment-1873953371

   @gsudhanshu Can you try disabling the timeline server? 
   
   hoodie.write.markers.type= 'direct',
   hoodie.embed.timeline.server= 'false'
   
   We had a silmilar issue (https://github.com/apache/hudi/issues/4230) before 
which we had fixed. I see you using 0.14.0, So adding @yihua in case he have 
more insights.
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@hudi.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[I] [SUPPORT] Failed to create marker file Exception when trying to write data on Hudi [hudi]

2023-12-29 Thread via GitHub


gsudhanshu opened a new issue, #10432:
URL: https://github.com/apache/hudi/issues/10432

   My Setup:
   My Django application on app server uses pyspark 3.4.2 to connect to 
spark-hudi bundle (3.4.2:2_12:0.14.0) on db server
   
   spark config:
   
   ```
   spark = SparkSession.builder \
   .appName("dataHudi") \
   .master('spark://db_server:7077') \
   .config('spark.driver.bindAddress', '0.0.0.0') \
   .config('spark.driver.host', 'app_server') \
   .config('spark.driver.port', '37077') \
   .config('spark.driver.blockManager.port', '37078') \
   .config('spark.executor.host', 'db_server') \
   .config("spark.executor.port", "37079") \
   .config('spark.fileserver.host', 'db_server') \
   .config("spark.fileserver.port", "37080") \
   .config('spark.replClassServer.host', 'db_server') \
   .config("spark.replClassServer.port", "37081") \
   .config('spark.broadcast.host', 'db_server') \
   .config("spark.broadcast.port", "37082") \
   .config('spark.driver.memory', '7g') \
   .config('spark.executor.memory', '4g') \
   .config('spark.jars.packages', 
'org.apache.hudi:hudi-spark3.4-bundle_2.12:0.14.0') \
   .config('spark.serializer', 
'org.apache.spark.serializer.KryoSerializer') \
   .config('spark.sql.catalog.spark_catalog', 
'org.apache.spark.sql.hudi.catalog.HoodieCatalog') \
   .config('spark.sql.extensions', 
'org.apache.spark.sql.hudi.HoodieSparkSessionExtension') \
   .getOrCreate()
   
   ```
   while writing data from pyspark pandas to hudi, I am getting following errors
   Event Error logs:
   ```
   org.apache.hudi.exception.HoodieRemoteException: Failed to create marker 
file 
/d13b0ab8-8d0d-4743-9f31-b3df2982965f-0_0-2-0_20231229115643749.parquet.marker.CREATE
   Connect to 10.0.0.xxx:46131 [/10.0.0.xxx] failed: Connection timed out 
(Connection timed out)
at 
org.apache.hudi.table.marker.TimelineServerBasedWriteMarkers.executeCreateMarkerRequest(TimelineServerBasedWriteMarkers.java:186)
at 
org.apache.hudi.table.marker.TimelineServerBasedWriteMarkers.create(TimelineServerBasedWriteMarkers.java:141)
at 
org.apache.hudi.table.marker.WriteMarkers.create(WriteMarkers.java:67)
at 
org.apache.hudi.io.storage.row.HoodieRowCreateHandle.createMarkerFile(HoodieRowCreateHandle.java:281)
at 
org.apache.hudi.io.storage.row.HoodieRowCreateHandle.(HoodieRowCreateHandle.java:144)
at 
org.apache.hudi.table.action.commit.BulkInsertDataInternalWriterHelper.createHandle(BulkInsertDataInternalWriterHelper.java:187)
at 
org.apache.hudi.table.action.commit.BulkInsertDataInternalWriterHelper.getRowCreateHandle(BulkInsertDataInternalWriterHelper.java:173)
at 
org.apache.hudi.table.action.commit.BulkInsertDataInternalWriterHelper.write(BulkInsertDataInternalWriterHelper.java:121)
at 
org.apache.hudi.spark3.internal.HoodieBulkInsertDataInternalWriter.write(HoodieBulkInsertDataInternalWriter.java:62)
at 
org.apache.hudi.spark3.internal.HoodieBulkInsertDataInternalWriter.write(HoodieBulkInsertDataInternalWriter.java:38)
at 
org.apache.spark.sql.execution.datasources.v2.DataWritingSparkTask$.write(WriteToDataSourceV2Exec.scala:516)
at 
org.apache.spark.sql.execution.datasources.v2.WritingSparkTask.$anonfun$run$1(WriteToDataSourceV2Exec.scala:471)
at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1563)
at 
org.apache.spark.sql.execution.datasources.v2.WritingSparkTask.run(WriteToDataSourceV2Exec.scala:509)
at 
org.apache.spark.sql.execution.datasources.v2.WritingSparkTask.run$(WriteToDataSourceV2Exec.scala:448)
at 
org.apache.spark.sql.execution.datasources.v2.DataWritingSparkTask$.run(WriteToDataSourceV2Exec.scala:514)
at 
org.apache.spark.sql.execution.datasources.v2.V2TableWriteExec.$anonfun$writeWithV2$2(WriteToDataSourceV2Exec.scala:411)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:92)
at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
at org.apache.spark.scheduler.Task.run(Task.scala:139)
at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
   Caused by: