[jira] [Created] (HBASE-4435) Add Group By functionality using Coprocessors
Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Adds in a Group By -like fucntionality to HBase using coprocessors -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4435: Attachment: HBase-4435.patch Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Attachments: HBase-4435.patch Adds in a Group By -like fucntionality to HBase using coprocessors -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4435: Attachment: HBase-4435.patch My first patch included some additional unrelated changes to other parts of the code base that I did not want to include in this patch...sorry about that. Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Attachments: HBase-4435.patch Adds in a Group By -like fucntionality to HBase using coprocessors -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4435: Attachment: (was: HBase-4435.patch) Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Attachments: HBase-4435.patch Adds in a Group By -like fucntionality to HBase using coprocessors -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4435: Description: Adds in a Group By -like functionality to HBase, using the Coprocessor framework. It provides the ability to group the result set on one or more columns (groupBy families). It computes statistics (max, min, sum, count, sum of squares, number missing) for a second column, called the stats column. To use, I've provided two implementations. 1. In the first, you specify a single group-by column and a stats field: statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter); The result is a map with the Group By column value (as a String) to a GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. of the stats column for that group. 2. The second implementation allows you to specify a list of group-by columns and a stats field. The List of group-by columns is expected to contain lists of {column family, qualifier} pairs. statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, statsFamily, statsQualifier, statsFieldColumnInterpreter); The GroupByStatsValues code is adapted from the Solr Stats component. was: Adds in a Group By -like functionality to HBase, using the Coprocessor framework. It provides the ability to group the result set on one or more columns (groupBy families). It computes statistics (max, min, sum, count, sum of squares, number missing) for a second column, called the stats column. To use, I've provided two implementations. 1. In the first, you specify a single group-by column and a stats field: statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter); The result is a map with the Group By column value (as a String) to a GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. of the stats column for that group. 2. The second implementation allows you to specify a list of group-by columns and a stats field. The List of group-by columns is expected to contain lists of {column family, qualifier} pairs. statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, statsFamily, statsQualifier, statsFieldColumnInterpreter); Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Attachments: HBase-4435.patch Adds in a Group By -like functionality to HBase, using the Coprocessor framework. It provides the ability to group the result set on one or more columns (groupBy families). It computes statistics (max, min, sum, count, sum of squares, number missing) for a second column, called the stats column. To use, I've provided two implementations. 1. In the first, you specify a single group-by column and a stats field: statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter); The result is a map with the Group By column value (as a String) to a GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. of the stats column for that group. 2. The second implementation allows you to specify a list of group-by columns and a stats field. The List of group-by columns is expected to contain lists of {column family, qualifier} pairs. statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, statsFamily, statsQualifier, statsFieldColumnInterpreter); The GroupByStatsValues code is adapted from the Solr Stats component. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4435: Description: Adds in a Group By -like functionality to HBase, using the Coprocessor framework. It provides the ability to group the result set on one or more columns (groupBy families). It computes statistics (max, min, sum, count, sum of squares, number missing) for a second column, called the stats column. To use, I've provided two implementations. 1. In the first, you specify a single group-by column and a stats field: statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter); The result is a map with the Group By column value (as a String) to a GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. of the stats column for that group. 2. The second implementation allows you to specify a list of group-by columns and a stats field. The List of group-by columns is expected to contain lists of {column family, qualifier} pairs. statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, statsFamily, statsQualifier, statsFieldColumnInterpreter); was:Adds in a Group By -like fucntionality to HBase using coprocessors Add Group By functionality using Coprocessors - Key: HBASE-4435 URL: https://issues.apache.org/jira/browse/HBASE-4435 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Nichole Treadway Priority: Minor Attachments: HBase-4435.patch Adds in a Group By -like functionality to HBase, using the Coprocessor framework. It provides the ability to group the result set on one or more columns (groupBy families). It computes statistics (max, min, sum, count, sum of squares, number missing) for a second column, called the stats column. To use, I've provided two implementations. 1. In the first, you specify a single group-by column and a stats field: statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter); The result is a map with the Group By column value (as a String) to a GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. of the stats column for that group. 2. The second implementation allows you to specify a list of group-by columns and a stats field. The List of group-by columns is expected to contain lists of {column family, qualifier} pairs. statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, statsFamily, statsQualifier, statsFieldColumnInterpreter); -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4223) Support the ability to return a set of rows using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4223: Affects Version/s: 0.92.0 Status: Patch Available (was: Open) Support the ability to return a set of rows using Coprocessors -- Key: HBASE-4223 URL: https://issues.apache.org/jira/browse/HBASE-4223 Project: HBase Issue Type: Improvement Components: coprocessors Affects Versions: 0.92.0 Reporter: Nichole Treadway Priority: Minor Currently HBase supports returning the results of aggregation operations using coprocessors with the AggregationClient. It would be useful to include a client and implementation which would return a set of rows which match a certain criteria using coprocessors as well. We have a use case in our business process for this. We have an initial implementation of this, which I've attached. The only limitation that we've found is that it cannot be used to return very large sets of rows. If the result set is very large, it would probably require some sort of pagination. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4223) Support the ability to return a set of rows using Coprocessors
[ https://issues.apache.org/jira/browse/HBASE-4223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-4223: Attachment: HBASE-4223.patch Support the ability to return a set of rows using Coprocessors -- Key: HBASE-4223 URL: https://issues.apache.org/jira/browse/HBASE-4223 Project: HBase Issue Type: Improvement Components: coprocessors Affects Versions: 0.92.0 Reporter: Nichole Treadway Priority: Minor Attachments: HBASE-4223.patch Currently HBase supports returning the results of aggregation operations using coprocessors with the AggregationClient. It would be useful to include a client and implementation which would return a set of rows which match a certain criteria using coprocessors as well. We have a use case in our business process for this. We have an initial implementation of this, which I've attached. The only limitation that we've found is that it cannot be used to return very large sets of rows. If the result set is very large, it would probably require some sort of pagination. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception
[ https://issues.apache.org/jira/browse/HBASE-3782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13031744#comment-13031744 ] Nichole Treadway commented on HBASE-3782: - Yes that's right...my patch is modifying the HBASE1861-incomplete.patch. Multi-Family support for bulk upload tools causes File Not Found Exception -- Key: HBASE-3782 URL: https://issues.apache.org/jira/browse/HBASE-3782 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.3 Reporter: Nichole Treadway Attachments: HBASE-3782.patch I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for bulk upload tools. I found that when running the importtsv program, some reduce tasks fail with a File Not Found exception if there are no keys in the input data which fall into the region assigned to that reduce task. From what I can determine, it seems that an output directory is created in the write() method and expected to exist in the writeMetaData() method...if there are no keys to be written for that reduce task, the write method is never called and the output directory is never created, but writeMetaData is expecting the output directory to exist...thus the FnF exception: 2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error running child java.io.FileNotFoundException: File does not exist: hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0 at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468) at org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247) at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408) at org.apache.hadoop.mapred.Child.main(Child.java:170) Simply checking if the file exists should fix the issue. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor
[ https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3691: Attachment: hbase-snappy-3691-trunk-003.patch Thanks for the patch...I made a few additional changes in HColumnDescriptor, and I updated the test files to include snappy. I noticed there are places in the hbase.avro classes where snappy support would need to be added in. Is it ok to add these changes in the patch, or do the avro classes need to be auto-generated somehow? Add compressor support for 'snappy', google's compressor Key: HBASE-3691 URL: https://issues.apache.org/jira/browse/HBASE-3691 Project: HBase Issue Type: Task Reporter: stack Priority: Critical Fix For: 0.92.0 Attachments: hbase-snappy-3691-trunk-002.patch, hbase-snappy-3691-trunk.patch http://code.google.com/p/snappy/ is apache licensed. bq. Snappy is a compression/decompression library. It does not aim for maximum compression, or compatibility with any other compression library; instead, it aims for very high speeds and reasonable compression. For instance, compared to the fastest mode of zlib, Snappy is an order of magnitude faster for most inputs, but the resulting compressed files are anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses at about 500 MB/sec or more. bq. Snappy is widely used inside Google, in everything from BigTable and MapReduce to our internal RPC systems. (Snappy has previously been referred to as Zippy in some presentations and the likes.) Lets get it in. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor
[ https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3691: Attachment: (was: hbase-snappy-3691-trunk-003.patch) Add compressor support for 'snappy', google's compressor Key: HBASE-3691 URL: https://issues.apache.org/jira/browse/HBASE-3691 Project: HBase Issue Type: Task Reporter: stack Priority: Critical Fix For: 0.92.0 Attachments: hbase-snappy-3691-trunk-002.patch, hbase-snappy-3691-trunk.patch http://code.google.com/p/snappy/ is apache licensed. bq. Snappy is a compression/decompression library. It does not aim for maximum compression, or compatibility with any other compression library; instead, it aims for very high speeds and reasonable compression. For instance, compared to the fastest mode of zlib, Snappy is an order of magnitude faster for most inputs, but the resulting compressed files are anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses at about 500 MB/sec or more. bq. Snappy is widely used inside Google, in everything from BigTable and MapReduce to our internal RPC systems. (Snappy has previously been referred to as Zippy in some presentations and the likes.) Lets get it in. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor
[ https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3691: Attachment: hbase-snappy-3691-trunk-003.patch Accidentally selected wrong license option. Add compressor support for 'snappy', google's compressor Key: HBASE-3691 URL: https://issues.apache.org/jira/browse/HBASE-3691 Project: HBase Issue Type: Task Reporter: stack Priority: Critical Fix For: 0.92.0 Attachments: hbase-snappy-3691-trunk-002.patch, hbase-snappy-3691-trunk-003.patch, hbase-snappy-3691-trunk.patch http://code.google.com/p/snappy/ is apache licensed. bq. Snappy is a compression/decompression library. It does not aim for maximum compression, or compatibility with any other compression library; instead, it aims for very high speeds and reasonable compression. For instance, compared to the fastest mode of zlib, Snappy is an order of magnitude faster for most inputs, but the resulting compressed files are anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses at about 500 MB/sec or more. bq. Snappy is widely used inside Google, in everything from BigTable and MapReduce to our internal RPC systems. (Snappy has previously been referred to as Zippy in some presentations and the likes.) Lets get it in. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception
Multi-Family support for bulk upload tools causes File Not Found Exception -- Key: HBASE-3782 URL: https://issues.apache.org/jira/browse/HBASE-3782 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.3 Reporter: Nichole Treadway I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for bulk upload tools. I found that when running the importtsv program, some reduce tasks fail with a File Not Found exception if there are no keys in the input data which fall into the region assigned to that reduce task. From what I can determine, it seems that an output directory is created in the write() method and expected to exist in the writeMetaData() method...if there are no keys to be written for that reduce task, the write method is never called and the output directory is never created, but writeMetaData is expecting the output directory to exist...thus the FnF exception: 2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error running child java.io.FileNotFoundException: File does not exist: hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0 at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468) at org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247) at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408) at org.apache.hadoop.mapred.Child.main(Child.java:170) Simply checking if the file exists should fix the issue. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception
[ https://issues.apache.org/jira/browse/HBASE-3782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3782: Attachment: HBASE-3782.patch Just modifying the patch in HBASE-1861 Multi-Family support for bulk upload tools causes File Not Found Exception -- Key: HBASE-3782 URL: https://issues.apache.org/jira/browse/HBASE-3782 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.3 Reporter: Nichole Treadway Attachments: HBASE-3782.patch I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for bulk upload tools. I found that when running the importtsv program, some reduce tasks fail with a File Not Found exception if there are no keys in the input data which fall into the region assigned to that reduce task. From what I can determine, it seems that an output directory is created in the write() method and expected to exist in the writeMetaData() method...if there are no keys to be written for that reduce task, the write method is never called and the output directory is never created, but writeMetaData is expecting the output directory to exist...thus the FnF exception: 2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error running child java.io.FileNotFoundException: File does not exist: hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0 at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468) at org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186) at org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247) at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408) at org.apache.hadoop.mapred.Child.main(Child.java:170) Simply checking if the file exists should fix the issue. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-3714) completebulkload does not use HBase configuration
completebulkload does not use HBase configuration - Key: HBASE-3714 URL: https://issues.apache.org/jira/browse/HBASE-3714 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.1, 0.90.0, 0.90.2, 0.90.3 Reporter: Nichole Treadway Attachments: HBASE-3714.txt The completebulkupload tool should be using the HBaseConfiguration.create() method to get the HBase configuration in 0.90.*. In it's present state, you receive a connection error when running this tool. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3714) completebulkload does not use HBase configuration
[ https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3714: Attachment: HBASE-3714.txt completebulkload does not use HBase configuration - Key: HBASE-3714 URL: https://issues.apache.org/jira/browse/HBASE-3714 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3 Reporter: Nichole Treadway Attachments: HBASE-3714.txt The completebulkupload tool should be using the HBaseConfiguration.create() method to get the HBase configuration in 0.90.*. In it's present state, you receive a connection error when running this tool. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3714) completebulkload does not use HBase configuration
[ https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nichole Treadway updated HBASE-3714: Priority: Minor (was: Major) completebulkload does not use HBase configuration - Key: HBASE-3714 URL: https://issues.apache.org/jira/browse/HBASE-3714 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3 Reporter: Nichole Treadway Priority: Minor Attachments: HBASE-3714.txt The completebulkupload tool should be using the HBaseConfiguration.create() method to get the HBase configuration in 0.90.*. In it's present state, you receive a connection error when running this tool. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-3714) completebulkload does not use HBase configuration
[ https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13013007#comment-13013007 ] Nichole Treadway commented on HBASE-3714: - Ted, LoadIncrementalHFiles and which other class did you mean? completebulkload does not use HBase configuration - Key: HBASE-3714 URL: https://issues.apache.org/jira/browse/HBASE-3714 Project: HBase Issue Type: Bug Components: mapreduce Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3 Reporter: Nichole Treadway Priority: Minor Attachments: HBASE-3714.txt The completebulkupload tool should be using the HBaseConfiguration.create() method to get the HBase configuration in 0.90.*. In it's present state, you receive a connection error when running this tool. -- This message is automatically generated by JIRA. For more information on JIRA, see: http://www.atlassian.com/software/jira