[jira] [Created] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)
Add Group By functionality using Coprocessors
-

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor


Adds in a Group By -like fucntionality to HBase using coprocessors

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4435:


Attachment: HBase-4435.patch

 Add Group By functionality using Coprocessors
 -

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBase-4435.patch


 Adds in a Group By -like fucntionality to HBase using coprocessors

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4435:


Attachment: HBase-4435.patch

My first patch included some additional unrelated changes to other parts of the 
code base that I did not want to include in this patch...sorry about that.

 Add Group By functionality using Coprocessors
 -

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBase-4435.patch


 Adds in a Group By -like fucntionality to HBase using coprocessors

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4435:


Attachment: (was: HBase-4435.patch)

 Add Group By functionality using Coprocessors
 -

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBase-4435.patch


 Adds in a Group By -like fucntionality to HBase using coprocessors

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4435:


Description: 
Adds in a Group By -like functionality to HBase, using the Coprocessor 
framework. 

It provides the ability to group the result set on one or more columns (groupBy 
families). It computes statistics (max, min, sum, count, sum of squares, number 
missing) for a second column, called the stats column. 

To use, I've provided two implementations.

1. In the first, you specify a single group-by column and a stats field:

  statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);

The result is a map with the Group By column value (as a String) to a 
GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. 
of the stats column for that group.

2. The second implementation allows you to specify a list of group-by columns 
and a stats field. The List of group-by columns is expected to contain lists of 
{column family, qualifier} pairs. 

  statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);


The GroupByStatsValues code is adapted from the Solr Stats component.



  was:
Adds in a Group By -like functionality to HBase, using the Coprocessor 
framework. 

It provides the ability to group the result set on one or more columns (groupBy 
families). It computes statistics (max, min, sum, count, sum of squares, number 
missing) for a second column, called the stats column. 

To use, I've provided two implementations.

1. In the first, you specify a single group-by column and a stats field:

  statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);

The result is a map with the Group By column value (as a String) to a 
GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. 
of the stats column for that group.

2. The second implementation allows you to specify a list of group-by columns 
and a stats field. The List of group-by columns is expected to contain lists of 
{column family, qualifier} pairs. 

  statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);




 Add Group By functionality using Coprocessors
 -

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBase-4435.patch


 Adds in a Group By -like functionality to HBase, using the Coprocessor 
 framework. 
 It provides the ability to group the result set on one or more columns 
 (groupBy families). It computes statistics (max, min, sum, count, sum of 
 squares, number missing) for a second column, called the stats column. 
 To use, I've provided two implementations.
 1. In the first, you specify a single group-by column and a stats field:
   statsMap = gbc.getStats(tableName, scan, groupByFamily, 
 groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter);
 The result is a map with the Group By column value (as a String) to a 
 GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. 
 of the stats column for that group.
 2. The second implementation allows you to specify a list of group-by columns 
 and a stats field. The List of group-by columns is expected to contain lists 
 of {column family, qualifier} pairs. 
   statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, 
 statsFamily, statsQualifier, statsFieldColumnInterpreter);
 The GroupByStatsValues code is adapted from the Solr Stats component.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4435) Add Group By functionality using Coprocessors

2011-09-19 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4435:


Description: 
Adds in a Group By -like functionality to HBase, using the Coprocessor 
framework. 

It provides the ability to group the result set on one or more columns (groupBy 
families). It computes statistics (max, min, sum, count, sum of squares, number 
missing) for a second column, called the stats column. 

To use, I've provided two implementations.

1. In the first, you specify a single group-by column and a stats field:

  statsMap = gbc.getStats(tableName, scan, groupByFamily, groupByQualifier, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);

The result is a map with the Group By column value (as a String) to a 
GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. 
of the stats column for that group.

2. The second implementation allows you to specify a list of group-by columns 
and a stats field. The List of group-by columns is expected to contain lists of 
{column family, qualifier} pairs. 

  statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, 
statsFamily, statsQualifier, statsFieldColumnInterpreter);



  was:Adds in a Group By -like fucntionality to HBase using coprocessors


 Add Group By functionality using Coprocessors
 -

 Key: HBASE-4435
 URL: https://issues.apache.org/jira/browse/HBASE-4435
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBase-4435.patch


 Adds in a Group By -like functionality to HBase, using the Coprocessor 
 framework. 
 It provides the ability to group the result set on one or more columns 
 (groupBy families). It computes statistics (max, min, sum, count, sum of 
 squares, number missing) for a second column, called the stats column. 
 To use, I've provided two implementations.
 1. In the first, you specify a single group-by column and a stats field:
   statsMap = gbc.getStats(tableName, scan, groupByFamily, 
 groupByQualifier, statsFamily, statsQualifier, statsFieldColumnInterpreter);
 The result is a map with the Group By column value (as a String) to a 
 GroupByStatsValues object. The GroupByStatsValues object has max,min,sum etc. 
 of the stats column for that group.
 2. The second implementation allows you to specify a list of group-by columns 
 and a stats field. The List of group-by columns is expected to contain lists 
 of {column family, qualifier} pairs. 
   statsMap = gbc.getStats(tableName, scan, listOfGroupByColumns, 
 statsFamily, statsQualifier, statsFieldColumnInterpreter);

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4223) Support the ability to return a set of rows using Coprocessors

2011-08-17 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4223:


Affects Version/s: 0.92.0
   Status: Patch Available  (was: Open)

 Support the ability to return a set of rows using Coprocessors
 --

 Key: HBASE-4223
 URL: https://issues.apache.org/jira/browse/HBASE-4223
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Affects Versions: 0.92.0
Reporter: Nichole Treadway
Priority: Minor

 Currently HBase supports returning the results of aggregation operations 
 using coprocessors with the AggregationClient. It would be useful to include 
 a client and implementation which would return a set of rows which match a 
 certain criteria using coprocessors as well. We have a use case in our 
 business process for this. 
 We have an initial implementation of this, which I've attached. The only 
 limitation that we've found is that it cannot be used to return very large 
 sets of rows. If the result set is very large, it would probably require some 
 sort of pagination.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Updated] (HBASE-4223) Support the ability to return a set of rows using Coprocessors

2011-08-17 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-4223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-4223:


Attachment: HBASE-4223.patch

 Support the ability to return a set of rows using Coprocessors
 --

 Key: HBASE-4223
 URL: https://issues.apache.org/jira/browse/HBASE-4223
 Project: HBase
  Issue Type: Improvement
  Components: coprocessors
Affects Versions: 0.92.0
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBASE-4223.patch


 Currently HBase supports returning the results of aggregation operations 
 using coprocessors with the AggregationClient. It would be useful to include 
 a client and implementation which would return a set of rows which match a 
 certain criteria using coprocessors as well. We have a use case in our 
 business process for this. 
 We have an initial implementation of this, which I've attached. The only 
 limitation that we've found is that it cannot be used to return very large 
 sets of rows. If the result set is very large, it would probably require some 
 sort of pagination.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Commented] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception

2011-05-11 Thread Nichole Treadway (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-3782?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13031744#comment-13031744
 ] 

Nichole Treadway commented on HBASE-3782:
-

Yes that's right...my patch is modifying the HBASE1861-incomplete.patch. 

 Multi-Family support for bulk upload tools causes File Not Found Exception
 --

 Key: HBASE-3782
 URL: https://issues.apache.org/jira/browse/HBASE-3782
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.3
Reporter: Nichole Treadway
 Attachments: HBASE-3782.patch


 I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for 
 bulk upload tools.
 I found that when running the importtsv program, some reduce tasks fail with 
 a File Not Found exception if there are no keys in the input data which fall 
 into the region assigned to that reduce task.  From what I can determine, it 
 seems that an output directory is created in the write() method and expected 
 to exist in the writeMetaData() method...if there are no keys to be written 
 for that reduce task, the write method is never called and the output 
 directory is never created, but writeMetaData is expecting the output 
 directory to exist...thus the FnF exception:
 2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error 
 running child
 java.io.FileNotFoundException: File does not exist: 
 hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0
   at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468)
   at 
 org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580)
   at 
 org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186)
   at 
 org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247)
   at 
 org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
   at org.apache.hadoop.mapred.Child.main(Child.java:170)
 Simply checking if the file exists should fix the issue. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor

2011-05-06 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3691:


Attachment: hbase-snappy-3691-trunk-003.patch

Thanks for the patch...I made a few additional changes in HColumnDescriptor, 
and I updated the test files to include snappy.

I noticed there are places in the hbase.avro classes where snappy support would 
need to be added in. Is it ok to add these changes in the patch, or do the avro 
classes need to be auto-generated somehow?

 Add compressor support for 'snappy', google's compressor
 

 Key: HBASE-3691
 URL: https://issues.apache.org/jira/browse/HBASE-3691
 Project: HBase
  Issue Type: Task
Reporter: stack
Priority: Critical
 Fix For: 0.92.0

 Attachments: hbase-snappy-3691-trunk-002.patch, 
 hbase-snappy-3691-trunk.patch


 http://code.google.com/p/snappy/ is apache licensed.
 bq. Snappy is a compression/decompression library. It does not aim for 
 maximum compression, or compatibility with any other compression library; 
 instead, it aims for very high speeds and reasonable compression. For 
 instance, compared to the fastest mode of zlib, Snappy is an order of 
 magnitude faster for most inputs, but the resulting compressed files are 
 anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 
 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses 
 at about 500 MB/sec or more.
 bq. Snappy is widely used inside Google, in everything from BigTable and 
 MapReduce to our internal RPC systems. (Snappy has previously been referred 
 to as Zippy in some presentations and the likes.)
 Lets get it in.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor

2011-05-06 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3691:


Attachment: (was: hbase-snappy-3691-trunk-003.patch)

 Add compressor support for 'snappy', google's compressor
 

 Key: HBASE-3691
 URL: https://issues.apache.org/jira/browse/HBASE-3691
 Project: HBase
  Issue Type: Task
Reporter: stack
Priority: Critical
 Fix For: 0.92.0

 Attachments: hbase-snappy-3691-trunk-002.patch, 
 hbase-snappy-3691-trunk.patch


 http://code.google.com/p/snappy/ is apache licensed.
 bq. Snappy is a compression/decompression library. It does not aim for 
 maximum compression, or compatibility with any other compression library; 
 instead, it aims for very high speeds and reasonable compression. For 
 instance, compared to the fastest mode of zlib, Snappy is an order of 
 magnitude faster for most inputs, but the resulting compressed files are 
 anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 
 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses 
 at about 500 MB/sec or more.
 bq. Snappy is widely used inside Google, in everything from BigTable and 
 MapReduce to our internal RPC systems. (Snappy has previously been referred 
 to as Zippy in some presentations and the likes.)
 Lets get it in.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3691) Add compressor support for 'snappy', google's compressor

2011-05-06 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3691:


Attachment: hbase-snappy-3691-trunk-003.patch

Accidentally selected wrong license option.

 Add compressor support for 'snappy', google's compressor
 

 Key: HBASE-3691
 URL: https://issues.apache.org/jira/browse/HBASE-3691
 Project: HBase
  Issue Type: Task
Reporter: stack
Priority: Critical
 Fix For: 0.92.0

 Attachments: hbase-snappy-3691-trunk-002.patch, 
 hbase-snappy-3691-trunk-003.patch, hbase-snappy-3691-trunk.patch


 http://code.google.com/p/snappy/ is apache licensed.
 bq. Snappy is a compression/decompression library. It does not aim for 
 maximum compression, or compatibility with any other compression library; 
 instead, it aims for very high speeds and reasonable compression. For 
 instance, compared to the fastest mode of zlib, Snappy is an order of 
 magnitude faster for most inputs, but the resulting compressed files are 
 anywhere from 20% to 100% bigger. On a single core of a Core i7 processor in 
 64-bit mode, Snappy compresses at about 250 MB/sec or more and decompresses 
 at about 500 MB/sec or more.
 bq. Snappy is widely used inside Google, in everything from BigTable and 
 MapReduce to our internal RPC systems. (Snappy has previously been referred 
 to as Zippy in some presentations and the likes.)
 Lets get it in.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception

2011-04-14 Thread Nichole Treadway (JIRA)
Multi-Family support for bulk upload tools causes File Not Found Exception
--

 Key: HBASE-3782
 URL: https://issues.apache.org/jira/browse/HBASE-3782
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.3
Reporter: Nichole Treadway


I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for 
bulk upload tools.

I found that when running the importtsv program, some reduce tasks fail with a 
File Not Found exception if there are no keys in the input data which fall into 
the region assigned to that reduce task.  From what I can determine, it seems 
that an output directory is created in the write() method and expected to exist 
in the writeMetaData() method...if there are no keys to be written for that 
reduce task, the write method is never called and the output directory is never 
created, but writeMetaData is expecting the output directory to exist...thus 
the FnF exception:

2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error 
running child
java.io.FileNotFoundException: File does not exist: 
hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0
at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468)
at 
org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580)
at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186)
at 
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247)
at 
org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
at org.apache.hadoop.mapred.Child.main(Child.java:170)

Simply checking if the file exists should fix the issue. 



--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3782) Multi-Family support for bulk upload tools causes File Not Found Exception

2011-04-14 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3782?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3782:


Attachment: HBASE-3782.patch

Just modifying the patch in HBASE-1861

 Multi-Family support for bulk upload tools causes File Not Found Exception
 --

 Key: HBASE-3782
 URL: https://issues.apache.org/jira/browse/HBASE-3782
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.3
Reporter: Nichole Treadway
 Attachments: HBASE-3782.patch


 I've been testing HBASE-1861 in 0.90.2, which adds multi-family support for 
 bulk upload tools.
 I found that when running the importtsv program, some reduce tasks fail with 
 a File Not Found exception if there are no keys in the input data which fall 
 into the region assigned to that reduce task.  From what I can determine, it 
 seems that an output directory is created in the write() method and expected 
 to exist in the writeMetaData() method...if there are no keys to be written 
 for that reduce task, the write method is never called and the output 
 directory is never created, but writeMetaData is expecting the output 
 directory to exist...thus the FnF exception:
 2011-03-17 11:52:48,095 WARN org.apache.hadoop.mapred.TaskTracker: Error 
 running child
 java.io.FileNotFoundException: File does not exist: 
 hdfs://master:9000/awardsData/_temporary/_attempt_201103151859_0066_r_00_0
   at 
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:468)
   at 
 org.apache.hadoop.hbase.regionserver.StoreFile.getUniqueFile(StoreFile.java:580)
   at 
 org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.writeMetaData(HFileOutputFormat.java:186)
   at 
 org.apache.hadoop.hbase.mapreduce.HFileOutputFormat$1.close(HFileOutputFormat.java:247)
   at 
 org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:567)
   at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:408)
   at org.apache.hadoop.mapred.Child.main(Child.java:170)
 Simply checking if the file exists should fix the issue. 

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HBASE-3714) completebulkload does not use HBase configuration

2011-03-30 Thread Nichole Treadway (JIRA)
completebulkload does not use HBase configuration
-

 Key: HBASE-3714
 URL: https://issues.apache.org/jira/browse/HBASE-3714
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.1, 0.90.0, 0.90.2, 0.90.3
Reporter: Nichole Treadway
 Attachments: HBASE-3714.txt

The completebulkupload tool should be using the HBaseConfiguration.create() 
method to get the HBase configuration in 0.90.*. In it's present state, you 
receive a connection error when running this tool.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3714) completebulkload does not use HBase configuration

2011-03-30 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3714:


Attachment: HBASE-3714.txt

 completebulkload does not use HBase configuration
 -

 Key: HBASE-3714
 URL: https://issues.apache.org/jira/browse/HBASE-3714
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3
Reporter: Nichole Treadway
 Attachments: HBASE-3714.txt


 The completebulkupload tool should be using the HBaseConfiguration.create() 
 method to get the HBase configuration in 0.90.*. In it's present state, you 
 receive a connection error when running this tool.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HBASE-3714) completebulkload does not use HBase configuration

2011-03-30 Thread Nichole Treadway (JIRA)

 [ 
https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Nichole Treadway updated HBASE-3714:


Priority: Minor  (was: Major)

 completebulkload does not use HBase configuration
 -

 Key: HBASE-3714
 URL: https://issues.apache.org/jira/browse/HBASE-3714
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBASE-3714.txt


 The completebulkupload tool should be using the HBaseConfiguration.create() 
 method to get the HBase configuration in 0.90.*. In it's present state, you 
 receive a connection error when running this tool.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HBASE-3714) completebulkload does not use HBase configuration

2011-03-30 Thread Nichole Treadway (JIRA)

[ 
https://issues.apache.org/jira/browse/HBASE-3714?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13013007#comment-13013007
 ] 

Nichole Treadway commented on HBASE-3714:
-

Ted, LoadIncrementalHFiles and which other class did you mean?

 completebulkload does not use HBase configuration
 -

 Key: HBASE-3714
 URL: https://issues.apache.org/jira/browse/HBASE-3714
 Project: HBase
  Issue Type: Bug
  Components: mapreduce
Affects Versions: 0.90.0, 0.90.1, 0.90.2, 0.90.3
Reporter: Nichole Treadway
Priority: Minor
 Attachments: HBASE-3714.txt


 The completebulkupload tool should be using the HBaseConfiguration.create() 
 method to get the HBase configuration in 0.90.*. In it's present state, you 
 receive a connection error when running this tool.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira