[jira] [Commented] (HADOOP-15849) Upgrade netty version to 3.10.6
[ https://issues.apache.org/jira/browse/HADOOP-15849?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16649240#comment-16649240 ] Hudson commented on HADOOP-15849: - SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #15210 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/15210/]) HADOOP-15849. Upgrade netty version to 3.10.6. (xiao: rev 8853fc8a55b07ecdc5ce8d85278b822e5675d97a) * (edit) hadoop-project/pom.xml > Upgrade netty version to 3.10.6 > > > Key: HADOOP-15849 > URL: https://issues.apache.org/jira/browse/HADOOP-15849 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Xiao Chen >Assignee: Xiao Chen >Priority: Major > Fix For: 3.3.0 > > Attachments: HADOOP-15849.01.patch > > > We're currently at 3.10.5. It'd be good to upgrade to the latest 3.10.6 > release. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15849) Upgrade netty version to 3.10.6
[ https://issues.apache.org/jira/browse/HADOOP-15849?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Chen updated HADOOP-15849: --- Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.3.0 Status: Resolved (was: Patch Available) Committed to trunk. Thank you, [~arpitagarwal]. > Upgrade netty version to 3.10.6 > > > Key: HADOOP-15849 > URL: https://issues.apache.org/jira/browse/HADOOP-15849 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Xiao Chen >Assignee: Xiao Chen >Priority: Major > Fix For: 3.3.0 > > Attachments: HADOOP-15849.01.patch > > > We're currently at 3.10.5. It'd be good to upgrade to the latest 3.10.6 > release. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15849) Upgrade netty version to 3.10.6
[ https://issues.apache.org/jira/browse/HADOOP-15849?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Chen updated HADOOP-15849: --- Summary: Upgrade netty version to 3.10.6 (was: Upgrade netty version) > Upgrade netty version to 3.10.6 > > > Key: HADOOP-15849 > URL: https://issues.apache.org/jira/browse/HADOOP-15849 > Project: Hadoop Common > Issue Type: Improvement >Reporter: Xiao Chen >Assignee: Xiao Chen >Priority: Major > Attachments: HADOOP-15849.01.patch > > > We're currently at 3.10.5. It'd be good to upgrade to the latest 3.10.6 > release. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15802) start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} file entry
[ https://issues.apache.org/jira/browse/HADOOP-15802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16649226#comment-16649226 ] Jon Boone commented on HADOOP-15802: It doesn't appear that there are any tests for _start-build-env.sh_, so I didn't modify or add any. > start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} > file entry > --- > > Key: HADOOP-15802 > URL: https://issues.apache.org/jira/browse/HADOOP-15802 > Project: Hadoop Common > Issue Type: Bug > Components: common > Environment: Ubuntu 18.04 x86_64 running in a VM with 4 CPUs / 8 GBs > RAM / 128 GB disk. >Reporter: Jon Boone >Priority: Minor > Attachments: HADOOP-15802.001.patch > > > In my Ubuntu 18.04 dev VM, I cloned the hadoop repo and ran the > start-build-env.sh script. Once the docker build was completed and the > container running, I tried to sudo and it failed. Upon investigation, I > discovered that it was creating an entry in > /etc/sudoers.d/hadoop-build-${USER_ID} that contained the characters '\t' > rather than a tab. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15802) start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} file entry
[ https://issues.apache.org/jira/browse/HADOOP-15802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16649224#comment-16649224 ] Hadoop QA commented on HADOOP-15802: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 26s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 22m 52s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 13m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 53s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 24s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 11m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} shellcheck {color} | {color:green} 0m 0s{color} | {color:green} There were no new shellcheck issues. {color} | | {color:green}+1{color} | {color:green} shelldocs {color} | {color:green} 0m 13s{color} | {color:green} There were no new shelldocs issues. {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 33s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 17m 49s{color} | {color:green} root in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 46s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}110m 57s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:4b8c2b1 | | JIRA Issue | HADOOP-15802 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12943802/HADOOP-15802.001.patch | | Optional Tests | dupname asflicense mvnsite unit shellcheck shelldocs | | uname | Linux 0cfca0b23943 3.13.0-144-generic #193-Ubuntu SMP Thu Mar 15 17:03:53 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 5209c75 | | maven | version: Apache Maven 3.3.9 | | shellcheck | v0.4.6 | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/15363/testReport/ | | Max. process+thread count | 316 (vs. ulimit of 1) | | modules | C: . U: . | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/15363/console | | Powered by | Apache Yetus 0.8.0 http://yetus.apache.org | This message was automatically generated. > start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} > file entry > --- > > Key: HADOOP-15802 > URL: https://issues.apache.org/jira/browse/HADOOP-15802 > Project: Hadoop Common > Issue Type: Bug > Components: common > Environment: Ubuntu 18.04 x86_64 running in a VM with 4 CPUs / 8 GBs > RAM / 128 GB disk. >Reporter: Jon Boone >Priority: Minor > Attachments: HADOOP-15802.001.patch > > > In my Ubuntu 18.04 dev VM, I cloned the hadoop repo and ran the > start-build-env.sh script. Once the docker build was completed and the > container running, I tried to sudo and it failed. Upon investigation, I > discovered that it was creating an entry in > /etc/sudoers.d/hadoop-build-${USER_ID} that contained the characters '\t' > rather than a tab. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (HADOOP-15802) start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} file entry
[ https://issues.apache.org/jira/browse/HADOOP-15802?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jon Boone updated HADOOP-15802: --- Attachment: HADOOP-15802.001.patch Status: Patch Available (was: Open) > start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} > file entry > --- > > Key: HADOOP-15802 > URL: https://issues.apache.org/jira/browse/HADOOP-15802 > Project: Hadoop Common > Issue Type: Bug > Components: common > Environment: Ubuntu 18.04 x86_64 running in a VM with 4 CPUs / 8 GBs > RAM / 128 GB disk. >Reporter: Jon Boone >Priority: Minor > Attachments: HADOOP-15802.001.patch > > > In my Ubuntu 18.04 dev VM, I cloned the hadoop repo and ran the > start-build-env.sh script. Once the docker build was completed and the > container running, I tried to sudo and it failed. Upon investigation, I > discovered that it was creating an entry in > /etc/sudoers.d/hadoop-build-${USER_ID} that contained the characters '\t' > rather than a tab. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
[ https://issues.apache.org/jira/browse/HADOOP-15850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16648975#comment-16648975 ] Ted Yu edited comment on HADOOP-15850 at 10/13/18 2:28 PM: --- [~yzhangal]: When you have chance, can you take a look ? Maybe I missed some existing DistCp functionality. Thanks was (Author: yuzhih...@gmail.com): [~yzhangal]: When you have chance, can you take a look ? Maybe I missed some existing DistCp functionality. > Allow CopyCommitter to skip concatenating source files specified by > DistCpConstants.CONF_LABEL_LISTING_FILE_PATH > > > Key: HADOOP-15850 > URL: https://issues.apache.org/jira/browse/HADOOP-15850 > Project: Hadoop Common > Issue Type: Task >Reporter: Ted Yu >Priority: Major > Attachments: testIncrementalBackupWithBulkLoad-output.txt > > > I was investigating test failure of TestIncrementalBackupWithBulkLoad from > hbase against hadoop 3.1.1 > hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: > {code} > LOG.debug("creating input listing " + listing + " , totalRecords=" + > totalRecords); > cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); > cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, > totalRecords); > {code} > For the test case, two bulk loaded hfiles are in the listing: > {code} > 2018-10-13 14:09:24,123 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for > 2 files of 10242 > {code} > Later on, CopyCommitter#concatFileChunks would throw the following exception: > {code} > 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): > job_local1795473782_0004 > java.io.IOException: Inconsistent sequence file: current chunk file > org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ > > 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry > org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- > > 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > length = 5142 aclEntries = null, xAttrs = null} > at > org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) > at > org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) > {code} > The above warning shouldn't happen - the two bulk loaded hfiles are > independent. > From the contents of the two CopyListingFileStatus instances, we can see that > their isSplit() return false. Otherwise the following from toString should be > logged: > {code} > if (isSplit()) { > sb.append(", chunkOffset = ").append(this.getChunkOffset()); > sb.append(", chunkLength = ").append(this.getChunkLength()); > } > {code} > From hbase side, we can specify one bulk loaded hfile per job but that > defeats the purpose of using DistCp. > There should be a way for DistCp to specify the skipping of source file > concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
[ https://issues.apache.org/jira/browse/HADOOP-15850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16648975#comment-16648975 ] Ted Yu commented on HADOOP-15850: - [~yzhangal]: When you have chance, can you take a look ? Maybe I missed some existing DistCp functionality. > Allow CopyCommitter to skip concatenating source files specified by > DistCpConstants.CONF_LABEL_LISTING_FILE_PATH > > > Key: HADOOP-15850 > URL: https://issues.apache.org/jira/browse/HADOOP-15850 > Project: Hadoop Common > Issue Type: Task >Reporter: Ted Yu >Priority: Major > Attachments: testIncrementalBackupWithBulkLoad-output.txt > > > I was investigating test failure of TestIncrementalBackupWithBulkLoad from > hbase against hadoop 3.1.1 > hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: > {code} > LOG.debug("creating input listing " + listing + " , totalRecords=" + > totalRecords); > cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); > cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, > totalRecords); > {code} > For the test case, two bulk loaded hfiles are in the listing: > {code} > 2018-10-13 14:09:24,123 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for > 2 files of 10242 > {code} > Later on, CopyCommitter#concatFileChunks would throw the following exception: > {code} > 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): > job_local1795473782_0004 > java.io.IOException: Inconsistent sequence file: current chunk file > org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ > > 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry > org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- > > 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > length = 5142 aclEntries = null, xAttrs = null} > at > org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) > at > org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) > {code} > The above warning shouldn't happen - the two bulk loaded hfiles are > independent. > From the contents of the two CopyListingFileStatus instances, we can see that > their isSplit() return false. Otherwise the following from toString should be > logged: > {code} > if (isSplit()) { > sb.append(", chunkOffset = ").append(this.getChunkOffset()); > sb.append(", chunkLength = ").append(this.getChunkLength()); > } > {code} > From hbase side, we can specify one bulk loaded hfile per job but that > defeats the purpose of using DistCp. > There should be a way for DistCp to specify the skipping of source file > concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
[ https://issues.apache.org/jira/browse/HADOOP-15850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16648974#comment-16648974 ] Ted Yu commented on HADOOP-15850: - The quoted test output was from testIncrementalBackupWithBulkLoad-output.txt > Allow CopyCommitter to skip concatenating source files specified by > DistCpConstants.CONF_LABEL_LISTING_FILE_PATH > > > Key: HADOOP-15850 > URL: https://issues.apache.org/jira/browse/HADOOP-15850 > Project: Hadoop Common > Issue Type: Task >Reporter: Ted Yu >Priority: Major > Attachments: testIncrementalBackupWithBulkLoad-output.txt > > > I was investigating test failure of TestIncrementalBackupWithBulkLoad from > hbase against hadoop 3.1.1 > hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: > {code} > LOG.debug("creating input listing " + listing + " , totalRecords=" + > totalRecords); > cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); > cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, > totalRecords); > {code} > For the test case, two bulk loaded hfiles are in the listing: > {code} > 2018-10-13 14:09:24,123 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for > 2 files of 10242 > {code} > Later on, CopyCommitter#concatFileChunks would throw the following exception: > {code} > 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): > job_local1795473782_0004 > java.io.IOException: Inconsistent sequence file: current chunk file > org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ > > 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry > org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- > > 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > length = 5142 aclEntries = null, xAttrs = null} > at > org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) > at > org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) > {code} > The above warning shouldn't happen - the two bulk loaded hfiles are > independent. > From the contents of the two CopyListingFileStatus instances, we can see that > their isSplit() return false. Otherwise the following from toString should be > logged: > {code} > if (isSplit()) { > sb.append(", chunkOffset = ").append(this.getChunkOffset()); > sb.append(", chunkLength = ").append(this.getChunkLength()); > } > {code} > From hbase side, we can specify one bulk loaded hfile per job but that > defeats the purpose of using DistCp. > There should be a way for DistCp to specify the skipping of source file > concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
[ https://issues.apache.org/jira/browse/HADOOP-15850?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16648973#comment-16648973 ] Ted Yu commented on HADOOP-15850: - This is hbase code: https://github.com/apache/hbase/blob/master/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/mapreduce/MapReduceBackupCopyJob.java#L153 > Allow CopyCommitter to skip concatenating source files specified by > DistCpConstants.CONF_LABEL_LISTING_FILE_PATH > > > Key: HADOOP-15850 > URL: https://issues.apache.org/jira/browse/HADOOP-15850 > Project: Hadoop Common > Issue Type: Task >Reporter: Ted Yu >Priority: Major > Attachments: testIncrementalBackupWithBulkLoad-output.txt > > > I was investigating test failure of TestIncrementalBackupWithBulkLoad from > hbase against hadoop 3.1.1 > hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: > {code} > LOG.debug("creating input listing " + listing + " , totalRecords=" + > totalRecords); > cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); > cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, > totalRecords); > {code} > For the test case, two bulk loaded hfiles are in the listing: > {code} > 2018-10-13 14:09:24,123 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for > 2 files of 10242 > {code} > Later on, CopyCommitter#concatFileChunks would throw the following exception: > {code} > 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): > job_local1795473782_0004 > java.io.IOException: Inconsistent sequence file: current chunk file > org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ > > 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry > org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- > > 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > length = 5142 aclEntries = null, xAttrs = null} > at > org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) > at > org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) > {code} > The above warning shouldn't happen - the two bulk loaded hfiles are > independent. > From the contents of the two CopyListingFileStatus instances, we can see that > their isSplit() return false. Otherwise the following from toString should be > logged: > {code} > if (isSplit()) { > sb.append(", chunkOffset = ").append(this.getChunkOffset()); > sb.append(", chunkLength = ").append(this.getChunkLength()); > } > {code} > From hbase side, we can specify one bulk loaded hfile per job but that > defeats the purpose of using DistCp. > There should be a way for DistCp to specify the skipping of source file > concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
[ https://issues.apache.org/jira/browse/HADOOP-15850?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ted Yu updated HADOOP-15850: Attachment: testIncrementalBackupWithBulkLoad-output.txt > Allow CopyCommitter to skip concatenating source files specified by > DistCpConstants.CONF_LABEL_LISTING_FILE_PATH > > > Key: HADOOP-15850 > URL: https://issues.apache.org/jira/browse/HADOOP-15850 > Project: Hadoop Common > Issue Type: Task >Reporter: Ted Yu >Priority: Major > Attachments: testIncrementalBackupWithBulkLoad-output.txt > > > I was investigating test failure of TestIncrementalBackupWithBulkLoad from > hbase against hadoop 3.1.1 > hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: > {code} > LOG.debug("creating input listing " + listing + " , totalRecords=" + > totalRecords); > cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); > cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, > totalRecords); > {code} > For the test case, two bulk loaded hfiles are in the listing: > {code} > 2018-10-13 14:09:24,123 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : > hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > 2018-10-13 14:09:24,125 DEBUG [Time-limited test] > mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for > 2 files of 10242 > {code} > Later on, CopyCommitter#concatFileChunks would throw the following exception: > {code} > 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): > job_local1795473782_0004 > java.io.IOException: Inconsistent sequence file: current chunk file > org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ > > 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ > length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry > org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- > > 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ > length = 5142 aclEntries = null, xAttrs = null} > at > org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) > at > org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) > {code} > The above warning shouldn't happen - the two bulk loaded hfiles are > independent. > From the contents of the two CopyListingFileStatus instances, we can see that > their isSplit() return false. Otherwise the following from toString should be > logged: > {code} > if (isSplit()) { > sb.append(", chunkOffset = ").append(this.getChunkOffset()); > sb.append(", chunkLength = ").append(this.getChunkLength()); > } > {code} > From hbase side, we can specify one bulk loaded hfile per job but that > defeats the purpose of using DistCp. > There should be a way for DistCp to specify the skipping of source file > concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Created] (HADOOP-15850) Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH
Ted Yu created HADOOP-15850: --- Summary: Allow CopyCommitter to skip concatenating source files specified by DistCpConstants.CONF_LABEL_LISTING_FILE_PATH Key: HADOOP-15850 URL: https://issues.apache.org/jira/browse/HADOOP-15850 Project: Hadoop Common Issue Type: Task Reporter: Ted Yu I was investigating test failure of TestIncrementalBackupWithBulkLoad from hbase against hadoop 3.1.1 hbase MapReduceBackupCopyJob$BackupDistCp would create listing file: {code} LOG.debug("creating input listing " + listing + " , totalRecords=" + totalRecords); cfg.set(DistCpConstants.CONF_LABEL_LISTING_FILE_PATH, listing); cfg.setLong(DistCpConstants.CONF_LABEL_TOTAL_NUMBER_OF_RECORDS, totalRecords); {code} For the test case, two bulk loaded hfiles are in the listing: {code} 2018-10-13 14:09:24,123 DEBUG [Time-limited test] mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ 2018-10-13 14:09:24,125 DEBUG [Time-limited test] mapreduce.MapReduceBackupCopyJob$BackupDistCp(195): BackupDistCp : hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ 2018-10-13 14:09:24,125 DEBUG [Time-limited test] mapreduce.MapReduceBackupCopyJob$BackupDistCp(197): BackupDistCp execute for 2 files of 10242 {code} Later on, CopyCommitter#concatFileChunks would throw the following exception: {code} 2018-10-13 14:09:25,351 WARN [Thread-936] mapred.LocalJobRunner$Job(590): job_local1795473782_0004 java.io.IOException: Inconsistent sequence file: current chunk file org.apache.hadoop.tools.CopyListingFileStatus@bb8826ee{hdfs://localhost:42796/user/hbase/test-data/ 160aeab5-6bca-9f87-465e-2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/a7599081e835440eb7bf0dd3ef4fd7a5_SeqId_205_ length = 5100 aclEntries = null, xAttrs = null} doesnt match prior entry org.apache.hadoop.tools.CopyListingFileStatus@243d544d{hdfs://localhost:42796/user/hbase/test-data/160aeab5-6bca-9f87-465e- 2517a0c43119/data/default/test-1539439707496/96b5a3613d52f4df1ba87a1cef20684c/f/394e6d39a9b94b148b9089c4fb967aad_SeqId_205_ length = 5142 aclEntries = null, xAttrs = null} at org.apache.hadoop.tools.mapred.CopyCommitter.concatFileChunks(CopyCommitter.java:276) at org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:100) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567) {code} The above warning shouldn't happen - the two bulk loaded hfiles are independent. >From the contents of the two CopyListingFileStatus instances, we can see that >their isSplit() return false. Otherwise the following from toString should be >logged: {code} if (isSplit()) { sb.append(", chunkOffset = ").append(this.getChunkOffset()); sb.append(", chunkLength = ").append(this.getChunkLength()); } {code} >From hbase side, we can specify one bulk loaded hfile per job but that defeats >the purpose of using DistCp. There should be a way for DistCp to specify the skipping of source file concatenation. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15802) start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} file entry
[ https://issues.apache.org/jira/browse/HADOOP-15802?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16648900#comment-16648900 ] Jon Boone commented on HADOOP-15802: My first thought was to add the '-e' option to the echo command being called in line 64: ??RUN echo "${USER_NAME}\tALL=NOPASSWD: ALL" > "/etc/sudoers.d/hadoop-build-${USER_ID}"?? as this command is running inside a docker container which is running Linux. > start-build-env.sh creates an invalid /etc/sudoers.d/hadoop-build-${USER_ID} > file entry > --- > > Key: HADOOP-15802 > URL: https://issues.apache.org/jira/browse/HADOOP-15802 > Project: Hadoop Common > Issue Type: Bug > Components: common > Environment: Ubuntu 18.04 x86_64 running in a VM with 4 CPUs / 8 GBs > RAM / 128 GB disk. >Reporter: Jon Boone >Priority: Minor > > In my Ubuntu 18.04 dev VM, I cloned the hadoop repo and ran the > start-build-env.sh script. Once the docker build was completed and the > container running, I tried to sudo and it failed. Upon investigation, I > discovered that it was creating an entry in > /etc/sudoers.d/hadoop-build-${USER_ID} that contained the characters '\t' > rather than a tab. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org