[
https://issues.apache.org/jira/browse/HCATALOG-538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13538645#comment-13538645
]
Sushanth Sowmyan commented on HCATALOG-538:
-------------------------------------------
Hi, I just created HCATALOG-580 that follows an e2e failure caused due to this
patch.
I want to make sure that any "fixes" we make there does not cause too
significant a performance degradation here, so could you please track that as
we solve it? Also, if you have any solutions, I'm all ears. :)
> HCatalogStorer fails for 100GB of data with dynamic partitioning (number of
> partition is 300)
> ---------------------------------------------------------------------------------------------
>
> Key: HCATALOG-538
> URL: https://issues.apache.org/jira/browse/HCATALOG-538
> Project: HCatalog
> Issue Type: Bug
> Affects Versions: 0.4, 0.5
> Environment: Hadoop 0.23.4
> HCatalog 0.4
> Reporter: Arup Malakar
> Assignee: Arup Malakar
> Fix For: 0.4.1
>
> Attachments: HCATALOG-538-branch0.4-0.patch,
> HCATALOG-538-trunk-0.patch
>
>
> A hadoop job with 100GB of data and 300 partitions fails. All the maps
> succeed fine but the commit job fails after that. This looks like a timeout
> issue as commitJob() takes more than 10 minutes. I am running this on
> hadoop-0.23.4. I am playing with yarn.nm.liveness-monitor.expiry-interval-ms,
> yarn.am.liveness-monitor.expiry-interval-ms etc to make it work.
> This JIRA is for optimizing the commitJob(), as 10 minutes is too long.
> On a side note for storing 100GB of data without partition takes ~12 minutes,
> same amount of data with 300 partitions fails after 45 minutes. These tests
> were run on a 10 node cluster.
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira