[jira] Commented: (HADOOP-2249) [hbase] Add means of getting the timestamps for all cell versions: e.g. long [] getVersions(row, column)

2007-11-30 Thread Bryan Duxbury (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547389
 ] 

Bryan Duxbury commented on HADOOP-2249:
---

I was hoping for a method that would give you all of the timestamps for a row, 
not per row and column. You'd want to able to do this so that you could know 
what different versions of a particular entity you have in HBase. As such, 
basically all I'm looking for is:

long[] getRowTimestamps(Text row)

> [hbase] Add means of getting the timestamps for all cell versions: e.g. long 
> [] getVersions(row, column)
> 
>
> Key: HADOOP-2249
> URL: https://issues.apache.org/jira/browse/HADOOP-2249
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Priority: Minor
> Fix For: 0.16.0
>
>
> Should be means of asking hbase for list of all the timestamps associated 
> with a particular cell.  The brute force way would be adding a getVersions 
> method but perhaps we can come up w/ something more elegant than this?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547383
 ] 

Hadoop QA commented on HADOOP-2314:
---

+1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370690/block.patch
against trunk revision r600019.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests +1.  The patch passed core unit tests.

contrib tests +1.  The patch passed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1226/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1226/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1226/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1226/console

This message is automatically generated.

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>  Components: dfs
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2326) Add a Random backoff for the initial block report sent to the Name node

2007-11-30 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HADOOP-2326:
-

Attachment: (was: asyncRPC-4.patch)

> Add a Random backoff for the initial block report sent to the Name node
> ---
>
> Key: HADOOP-2326
> URL: https://issues.apache.org/jira/browse/HADOOP-2326
> Project: Hadoop
>  Issue Type: Improvement
>  Components: dfs
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
> Fix For: 0.16.0
>
>
> Startup time can be improved if the initial block reports are spread randomly 
> over small period of time.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2326) Add a Random backoff for the initial block report sent to the Name node

2007-11-30 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HADOOP-2326:
-

Comment: was deleted

> Add a Random backoff for the initial block report sent to the Name node
> ---
>
> Key: HADOOP-2326
> URL: https://issues.apache.org/jira/browse/HADOOP-2326
> Project: Hadoop
>  Issue Type: Improvement
>  Components: dfs
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
> Fix For: 0.16.0
>
>
> Startup time can be improved if the initial block reports are spread randomly 
> over small period of time.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-1841) IPC server should write repsonses asynchronously

2007-11-30 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HADOOP-1841:
-

Attachment: asyncRPC-4.patch

This patch is in response to Doug's comments that we need to show improvement 
in performance in the face of slow clients.

This patch has a unit test that creates a server with one handler thread. One 
thread makes an RPC and stops processing the response. Another thread then 
issues another RPC and it completes successfully even though the first thread 
has not yet consumed the RPC response. This test passes successfully with this 
patch whereas it fails with trunk.

Please let me know is it addresses your concerns. If so, then the only 
remaining thing to make this patch committable is to demonstrate that it does 
not degrade performance for sort runs.

> IPC server should write repsonses asynchronously
> 
>
> Key: HADOOP-1841
> URL: https://issues.apache.org/jira/browse/HADOOP-1841
> Project: Hadoop
>  Issue Type: Improvement
>  Components: ipc
>Reporter: Doug Cutting
>Assignee: dhruba borthakur
> Attachments: asyncRPC-2.patch, asyncRPC-4.patch, asyncRPC.patch, 
> asyncRPC.patch
>
>
> Hadoop's IPC Server currently writes responses from request handler threads 
> using blocking writes.  Performance and scalability might be improved if 
> responses were written asynchronously.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2326) Add a Random backoff for the initial block report sent to the Name node

2007-11-30 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2326?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HADOOP-2326:
-

Attachment: asyncRPC-4.patch

This patch is in response to Doug's comments that we need to show improvement 
in performance in the face of slow clients.

This patch has a unit test that creates a server with one handler thread. One 
thread makes an RPC and stops processing the response. Another thread then 
issues another RPC and it completes successfully even though the first thread 
has not yet consumed the RPC response. This test passes successfully with this 
patch whereas it fails with trunk.

Please let me know is it addresses your concerns. If so, then the only 
remaining thing to make this patch committable is to demonstrate that it does 
not degrade performance for sort runs.

> Add a Random backoff for the initial block report sent to the Name node
> ---
>
> Key: HADOOP-2326
> URL: https://issues.apache.org/jira/browse/HADOOP-2326
> Project: Hadoop
>  Issue Type: Improvement
>  Components: dfs
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
> Fix For: 0.16.0
>
> Attachments: asyncRPC-4.patch
>
>
> Startup time can be improved if the initial block reports are spread randomly 
> over small period of time.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Deleted: (HADOOP-1828) [Hbase Shell] Switch Command for sub-shell

2007-11-30 Thread Edward Yoon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-1828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Yoon deleted HADOOP-1828:



> [Hbase Shell] Switch Command for sub-shell
> --
>
> Key: HADOOP-1828
> URL: https://issues.apache.org/jira/browse/HADOOP-1828
> Project: Hadoop
>  Issue Type: Improvement
> Environment: All environments
>Reporter: Edward Yoon
>Priority: Minor
>
> This is a swtich command patch for the future implementation of sub-shell.
> {code}
> HBase > altools;
> Hbase altools, 0.0.1 version
> Type 'help;' for Hbase altools usage.
> HBase.Altools > help;
> Type 'help ;' to see command-specific usage.
> * Global commands.
> FSHadoop FsShell operations.
> EXIT  Exit shell
> SHOW  List all tables.
> CLEAR Clear the screen.
> DESCRIBE  Describe a table's columnfamilies.
> * Altools Commands.
> Projection ...
> HBase.Altools > exit;
> HBase > exit;
> {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



A tracing framework for Hadoop

2007-11-30 Thread Matei Zaharia

Hi,

We're grad students at UC Berkeley working on a project to instrument  
Hadoop using an open-source path-based tracing framework called X- 
Trace (www.x-trace.net/wiki). X-Trace captures causal dependencies  
between events in addition to timings, letting developers analyze not  
just performance but also context and dependencies for various events.  
We have created a web-based trace analysis UI that shows performance  
of different IPC calls, DFS operations, and phases of a MapReduce job.  
The goal is to let users easily spot the origin of unusual behavior in  
a running system at a centralized location. We believe that this kind  
of tracing can be used for performance tuning and debugging in both  
development and production environments.


We'd like to get feedback on our work and suggestions on what trace  
analyses would be useful to Hadoop developers and users. Some of the  
reports we currently generate include machine utilization over time,  
relative performance of different tasks, and performance of DFS  
operations. You can see an example set of reports at http://www.cs.berkeley.edu/~matei/xtrace_sample_task.html 
 (this is a trace of a Nutch indexing job). You can also read our  
project journal at http://radlab.cs.berkeley.edu/wiki/Projects/Monitoring_Hadoop_through_Tracing 
. We've already spotted some interesting issues, like map tasks and  
DFS reads/writes that are an order of magnitude slower than the  
average, and we are investigating possible causes for them. Most  
importantly, the UI lets a user easily see where the system is  
spending time and reason about how to tune it, and provides much more  
information than the progress data in the JobTracker UI. As a Hadoop  
developer, what kinds of questions do you want answered about running  
jobs that are hard to obtain just from process logs?


Once we've had a discussion on features for a trace analysis UI, we  
would like to contribute our work into the Hadoop codebase. We will  
create a JIRA issue and patch adding this functionality. We're also  
interested in seeing if we can integrate X-Trace logging more tightly  
with the current Apache logging in Hadoop.


Finally, we are currently experimenting on relatively small (<50  
nodes) clusters here at Berkeley, but we would really like to try  
tracing some large (>1000 node) clusters. If there is someone  
interested in evaluating performance on such a cluster, we would be  
very happy to talk about how to set up X-Trace and provide you with a  
patch.


Thanks,

Andy Konwinski and Matei Zaharia



[jira] Assigned: (HADOOP-2225) Joining Three or More Tables

2007-11-30 Thread Edward Yoon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Yoon reassigned HADOOP-2225:
---

Assignee: Edward Yoon

> Joining Three or More Tables
> 
>
> Key: HADOOP-2225
> URL: https://issues.apache.org/jira/browse/HADOOP-2225
> Project: Hadoop
>  Issue Type: Sub-task
>Reporter: Edward Yoon
>Assignee: Edward Yoon
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2240) Truncate for hbase

2007-11-30 Thread Edward Yoon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Yoon updated HADOOP-2240:


Fix Version/s: 0.16.0
Affects Version/s: 0.15.0
   Status: Patch Available  (was: In Progress)

submitting.

> Truncate for hbase
> --
>
> Key: HADOOP-2240
> URL: https://issues.apache.org/jira/browse/HADOOP-2240
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: Billy Pearson
>Assignee: Edward Yoon
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: 2240.patch
>
>
> Would be nice to have a way to truncate the tables from the shell. With out 
> doing a drop and create your self. Maybe the truncate could issue the drop 
> and create command for you based off the layout in the the table before the 
> truncate.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547376
 ] 

Hadoop QA commented on HADOOP-2229:
---

-1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370727/ugi5.patch
against trunk revision r600019.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs -1.  The patch appears to introduce 3 new Findbugs warnings.

core tests -1.  The patch failed core unit tests.

contrib tests -1.  The patch failed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1225/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1225/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1225/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1225/console

This message is automatically generated.

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, 
> ugi4.patch, ugi5.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2240) Truncate for hbase

2007-11-30 Thread Edward Yoon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Edward Yoon updated HADOOP-2240:


Attachment: 2240.patch

- Truncate table is used to clean all data from a table.

SYNTAX: TRUNCATE TABLE table_name;

{code}
hql > insert into test (a,b) values ('aa','bb') where row='row1';
1 row inserted successfully. (0.41 sec)
hql > truncate table test;
'test' is successfully truncated.
hql > select * from test;
0 row(s) in set. (1.16 sec)
hql > exit;
{code}

- We need to get away from functionalities and focus on the internal algorithm 
issues. I'll make these issues.
- i didn't make the test case becuase it already exists. (It just steps of 
creating drop repeats.)


> Truncate for hbase
> --
>
> Key: HADOOP-2240
> URL: https://issues.apache.org/jira/browse/HADOOP-2240
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: Billy Pearson
>Assignee: Edward Yoon
>Priority: Minor
> Attachments: 2240.patch
>
>
> Would be nice to have a way to truncate the tables from the shell. With out 
> doing a drop and create your self. Maybe the truncate could issue the drop 
> and create command for you based off the layout in the the table before the 
> truncate.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2326) Add a Random backoff for the initial block report sent to the Name node

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2326?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547374
 ] 

dhruba borthakur commented on HADOOP-2326:
--

And it would be nice if the backoff period is somehow dependent on either 
namenode load and/or cluster size.

> Add a Random backoff for the initial block report sent to the Name node
> ---
>
> Key: HADOOP-2326
> URL: https://issues.apache.org/jira/browse/HADOOP-2326
> Project: Hadoop
>  Issue Type: Improvement
>  Components: dfs
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
> Fix For: 0.16.0
>
>
> Startup time can be improved if the initial block reports are spread randomly 
> over small period of time.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



RE: hard link count for a file

2007-11-30 Thread dhruba Borthakur
Hi Raghu,

Can you pl explain what curFile and backupFile refer to in this context?
Are you referring to the block data file and the metafile?

Thanks,
dhruba

-Original Message-
From: Raghu Angadi [mailto:[EMAIL PROTECTED] 
Sent: Friday, November 30, 2007 4:05 PM
To: hadoop-dev@lucene.apache.org
Subject: Re: hard link count for a file


I am back to writing this code as part of periodic block verification.

Another approach I am thinking of is to check curFile.lastModified() and

backupFile.lastModified(). As long as these two are different, we could 
assume they are different files. This is simpler than trying to lock the

files.

Raghu.

Raghu Angadi wrote:
> 
> Thanks Nigel.
> 
> Another hack I can think of is, if DataNode is not finalized we can
try 
> to lock current/subdir.../blk_id and previous/subdir../blk_id. If
second 
> lock fails, we can assume they are the same.
> 
> Raghu.
> 
> Nigel Daley wrote:
>> Raghu,
>>
>> LGPL not allowed in distro:
>> http://people.apache.org/~cliffs/3party.html
>>
>> Nige
>>
>> On Nov 6, 2007, at 11:32 PM, Raghu Angadi wrote:
>>
> 



[jira] Work started: (HADOOP-2240) Truncate for hbase

2007-11-30 Thread Edward Yoon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2240?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HADOOP-2240 started by Edward Yoon.

> Truncate for hbase
> --
>
> Key: HADOOP-2240
> URL: https://issues.apache.org/jira/browse/HADOOP-2240
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: Billy Pearson
>Assignee: Edward Yoon
>Priority: Minor
>
> Would be nice to have a way to truncate the tables from the shell. With out 
> doing a drop and create your self. Maybe the truncate could issue the drop 
> and create command for you based off the layout in the the table before the 
> truncate.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547370
 ] 

Hadoop QA commented on HADOOP-2288:
---

-1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370725/2288_20071130c.patch
against trunk revision r600019.

@author +1.  The patch does not contain any @author tags.

javadoc -1.  The javadoc tool appears to have generated  messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs -1.  The patch appears to introduce 1 new Findbugs warnings.

core tests -1.  The patch failed core unit tests.

contrib tests -1.  The patch failed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1224/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1224/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1224/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1224/console

This message is automatically generated.

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130c.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2297) [Hbase Shell] System.exit() Handling in Jar command

2007-11-30 Thread Edward Yoon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2297?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547369
 ] 

Edward Yoon commented on HADOOP-2297:
-

Oh, Thanks michael!
I'll try to do it.

> [Hbase Shell] System.exit() Handling in Jar command
> ---
>
> Key: HADOOP-2297
> URL: https://issues.apache.org/jira/browse/HADOOP-2297
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: Edward Yoon
>Assignee: Edward Yoon
> Fix For: 0.16.0
>
> Attachments: 2297_v02.patch, 2297_v03.patch, Capture.java
>
>
> I'd like to block the exitVM by System.exit().
> Shell should terminate by "quit" command.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2323) JobTracker.close() prints stack traces for exceptions that are not errors

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547362
 ] 

Hadoop QA commented on HADOOP-2323:
---

+1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370714/patch.txt
against trunk revision r599951.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests +1.  The patch passed core unit tests.

contrib tests +1.  The patch passed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1223/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1223/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1223/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1223/console

This message is automatically generated.

> JobTracker.close() prints stack traces for exceptions that are not errors
> -
>
> Key: HADOOP-2323
> URL: https://issues.apache.org/jira/browse/HADOOP-2323
> Project: Hadoop
>  Issue Type: Bug
>  Components: mapred
>Affects Versions: 0.16.0
>Reporter: Jim Kellerman
>Assignee: Jim Kellerman
> Fix For: 0.16.0
>
> Attachments: patch.txt
>
>
> JobTracker.close() prints a stack trace for an interrupted exception even 
> though it was the method that interrupted the thread that threw the 
> exception. For example:
> {code}
>   this.expireTrackers.stopTracker();
>   try {
> this.expireTrackersThread.interrupt();
> this.expireTrackersThread.join();
>   } catch (InterruptedException ex) {
> ex.printStackTrace();
>   }
> {code}
> Well of course it is going to catch an InterruptedException after it just 
> interrupted the thread!
> This is *not* an error and should  *not* be dumped to the logs!
> In other circumstances, catching InterruptedException is entirely 
> appropriate. Just not in close where you've told the thread to shutdown and 
> then interrupted it to ensure it does!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2229:
--

Status: Patch Available  (was: Open)

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, 
> ugi4.patch, ugi5.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2229:
--

Attachment: ugi5.patch

This patch incorporates Sanjay's comments.

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, 
> ugi4.patch, ugi5.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: 2288_20071130c.patch

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130c.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: (was: 2288_20071130b.patch)

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130c.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Status: Patch Available  (was: Open)

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130c.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2326) Add a Random backoff for the initial block report sent to the Name node

2007-11-30 Thread Sanjay Radia (JIRA)
Add a Random backoff for the initial block report sent to the Name node
---

 Key: HADOOP-2326
 URL: https://issues.apache.org/jira/browse/HADOOP-2326
 Project: Hadoop
  Issue Type: Improvement
  Components: dfs
Reporter: Sanjay Radia
Assignee: Sanjay Radia
 Fix For: 0.16.0


Startup time can be improved if the initial block reports are spread randomly 
over small period of time.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547350
 ] 

Sanjay Radia commented on HADOOP-2229:
--

In writeFields, please write  the kind of authentication technology (something 
like "STRING_UGI") to distinguish it from other authentication info such as 
tickets.
 readFields, it should raise an exception if the the string does not match.
(This may be moot if we go to JAAS and GSS, as it probably has its own way of 
taking care of this, but at least it would be consistent for now).


The javadocs comments for some of the public methods are not in the 
conventional javadoc style.

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, ugi4.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2315) [hbase] REST servlet doesn't treat / characters in row key correctly

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2315?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2315:
--

Attachment: fix-urlencode-keys.patch

This patch resolves the problem.

> [hbase] REST servlet doesn't treat / characters in row key correctly
> 
>
> Key: HADOOP-2315
> URL: https://issues.apache.org/jira/browse/HADOOP-2315
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Trivial
> Attachments: fix-urlencode-keys.patch
>
>
> Using row keys like "com.site.www/:http" currently doesn't work. We've 
> tracked it down to the use of request.getPathInfo() instead of 
> request.getRequestURI() in Dispatcher.getPathSegments.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-496) Expose HDFS as a WebDAV store

2007-11-30 Thread Anurag Sharma (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anurag Sharma updated HADOOP-496:
-

Attachment: fuse-j-patch.zip
fuse-j-hadoopfs-0.zip

hi,

Attachments include the following:
- fuse-j-hadoop package
- fuse-j patch.

-thanks

> Expose HDFS as a WebDAV store
> -
>
> Key: HADOOP-496
> URL: https://issues.apache.org/jira/browse/HADOOP-496
> Project: Hadoop
>  Issue Type: New Feature
>  Components: dfs
>Reporter: Michel Tourn
>Assignee: Enis Soztutar
> Attachments: fuse-j-hadoopfs-0.zip, fuse-j-patch.zip, 
> hadoop-496-3.patch, hadoop-496-4.patch, hadoop-496-spool-cleanup.patch, 
> hadoop-webdav.zip, jetty-slide.xml, lib.webdav.tar.gz, screenshot-1.jpg, 
> slideusers.properties, webdav_wip1.patch, webdav_wip2.patch
>
>
> WebDAV stands for Distributed Authoring and Versioning. It is a set of 
> extensions to the HTTP protocol that lets users collaboratively edit and 
> manage files on a remote web server. It is often considered as a replacement 
> for NFS or SAMBA
> HDFS (Hadoop Distributed File System) needs a friendly file system interface. 
> DFSShell commands are unfamiliar. Instead it is more convenient for Hadoop 
> users to use a mountable network drive. A friendly interface to HDFS will be 
> used both for casual browsing of data and for bulk import/export. 
> The FUSE provider for HDFS is already available ( 
> http://issues.apache.org/jira/browse/HADOOP-17 )  but it had scalability 
> problems. WebDAV is a popular alternative. 
> The typical licensing terms for WebDAV tools are also attractive: 
> GPL for Linux client tools that Hadoop would not redistribute anyway. 
> More importantly, Apache Project/Apache license for Java tools and for server 
> components. 
> This allows for a tighter integration with the HDFS code base.
> There are some interesting Apache projects that support WebDAV.
> But these are probably too heavyweight for the needs of Hadoop:
> Tomcat servlet: 
> http://tomcat.apache.org/tomcat-4.1-doc/catalina/docs/api/org/apache/catalina/servlets/WebdavServlet.html
> Slide:  http://jakarta.apache.org/slide/
> Being HTTP-based and "backwards-compatible" with Web Browser clients, the 
> WebDAV server protocol could even be piggy-backed on the existing Web UI 
> ports of the Hadoop name node / data nodes. WebDAV can be hosted as (Jetty) 
> servlets. This minimizes server code bloat and this avoids additional network 
> traffic between HDFS and the WebDAV server.
> General Clients (read-only):
> Any web browser
> Linux Clients: 
> Mountable GPL davfs2  http://dav.sourceforge.net/
> FTP-like  GPL Cadaver http://www.webdav.org/cadaver/
> Server Protocol compliance tests:
> http://www.webdav.org/neon/litmus/  
> A goal is for Hadoop HDFS to pass this test (minus support for Properties)
> Pure Java clients:
> DAV Explorer Apache lic. http://www.ics.uci.edu/~webdav/  
> WebDAV also makes it convenient to add advanced features in an incremental 
> fashion:
> file locking, access control lists, hard links, symbolic links.
> New WebDAV standards get accepted and more or less featured WebDAV clients 
> exist.
> core  http://www.webdav.org/specs/rfc2518.html
> ACLs  http://www.webdav.org/specs/rfc3744.html
> redirects "soft links" http://greenbytes.de/tech/webdav/rfc4437.html
> BIND "hard links" http://www.webdav.org/bind/
> quota http://tools.ietf.org/html/rfc4331

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Doug Cutting (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547347
 ] 

Doug Cutting commented on HADOOP-2229:
--

+1 This looks good to me.

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, ugi4.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2325) Require Java 6 for release 0.16.

2007-11-30 Thread Doug Cutting (JIRA)
Require Java 6 for release 0.16.


 Key: HADOOP-2325
 URL: https://issues.apache.org/jira/browse/HADOOP-2325
 Project: Hadoop
  Issue Type: Improvement
  Components: build
Reporter: Doug Cutting
 Fix For: 0.16.0


We should require Java 6 for release 0.16.  Java 6 is now available for OS/X.  
Hadoop performs much better on Java 6.  And, finally, there are features of 
Java 6 (like 'df') that would be nice to use.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-496) Expose HDFS as a WebDAV store

2007-11-30 Thread Anurag Sharma (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-496?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547345
 ] 

Anurag Sharma commented on HADOOP-496:
--

hi,

We revived the old fuse-hadoop project (a FUSE-J based plugin that lets you 
mount Hadoop-FS).  We have tried this on a small cluster (10 nodes) and basic 
functionality works (mount, ls, cat,cp, mkdir, rm, mv, ...).

The main changes include some bug fixes to FUSE-J and changing the previous 
fuse-hadoop implementation to enforce write-once.  We found the FUSE framework 
to be straightforward and simple.

We have seen several mentions of using FUSE with Hadoop, so if there is a 
better place to post these files, please let me know.

Attachments to follow...

-thanks




> Expose HDFS as a WebDAV store
> -
>
> Key: HADOOP-496
> URL: https://issues.apache.org/jira/browse/HADOOP-496
> Project: Hadoop
>  Issue Type: New Feature
>  Components: dfs
>Reporter: Michel Tourn
>Assignee: Enis Soztutar
> Attachments: hadoop-496-3.patch, hadoop-496-4.patch, 
> hadoop-496-spool-cleanup.patch, hadoop-webdav.zip, jetty-slide.xml, 
> lib.webdav.tar.gz, screenshot-1.jpg, slideusers.properties, 
> webdav_wip1.patch, webdav_wip2.patch
>
>
> WebDAV stands for Distributed Authoring and Versioning. It is a set of 
> extensions to the HTTP protocol that lets users collaboratively edit and 
> manage files on a remote web server. It is often considered as a replacement 
> for NFS or SAMBA
> HDFS (Hadoop Distributed File System) needs a friendly file system interface. 
> DFSShell commands are unfamiliar. Instead it is more convenient for Hadoop 
> users to use a mountable network drive. A friendly interface to HDFS will be 
> used both for casual browsing of data and for bulk import/export. 
> The FUSE provider for HDFS is already available ( 
> http://issues.apache.org/jira/browse/HADOOP-17 )  but it had scalability 
> problems. WebDAV is a popular alternative. 
> The typical licensing terms for WebDAV tools are also attractive: 
> GPL for Linux client tools that Hadoop would not redistribute anyway. 
> More importantly, Apache Project/Apache license for Java tools and for server 
> components. 
> This allows for a tighter integration with the HDFS code base.
> There are some interesting Apache projects that support WebDAV.
> But these are probably too heavyweight for the needs of Hadoop:
> Tomcat servlet: 
> http://tomcat.apache.org/tomcat-4.1-doc/catalina/docs/api/org/apache/catalina/servlets/WebdavServlet.html
> Slide:  http://jakarta.apache.org/slide/
> Being HTTP-based and "backwards-compatible" with Web Browser clients, the 
> WebDAV server protocol could even be piggy-backed on the existing Web UI 
> ports of the Hadoop name node / data nodes. WebDAV can be hosted as (Jetty) 
> servlets. This minimizes server code bloat and this avoids additional network 
> traffic between HDFS and the WebDAV server.
> General Clients (read-only):
> Any web browser
> Linux Clients: 
> Mountable GPL davfs2  http://dav.sourceforge.net/
> FTP-like  GPL Cadaver http://www.webdav.org/cadaver/
> Server Protocol compliance tests:
> http://www.webdav.org/neon/litmus/  
> A goal is for Hadoop HDFS to pass this test (minus support for Properties)
> Pure Java clients:
> DAV Explorer Apache lic. http://www.ics.uci.edu/~webdav/  
> WebDAV also makes it convenient to add advanced features in an incremental 
> fashion:
> file locking, access control lists, hard links, symbolic links.
> New WebDAV standards get accepted and more or less featured WebDAV clients 
> exist.
> core  http://www.webdav.org/specs/rfc2518.html
> ACLs  http://www.webdav.org/specs/rfc3744.html
> redirects "soft links" http://greenbytes.de/tech/webdav/rfc4437.html
> BIND "hard links" http://www.webdav.org/bind/
> quota http://tools.ietf.org/html/rfc4331

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2322:
--

Fix Version/s: 0.16.0

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2322:
--

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed (The was passed by hudson when it was originally added mistakenly to 
hadoop-2224).  Resolving.

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



Re: hard link count for a file

2007-11-30 Thread Raghu Angadi


I am back to writing this code as part of periodic block verification.

Another approach I am thinking of is to check curFile.lastModified() and 
backupFile.lastModified(). As long as these two are different, we could 
assume they are different files. This is simpler than trying to lock the 
files.


Raghu.

Raghu Angadi wrote:


Thanks Nigel.

Another hack I can think of is, if DataNode is not finalized we can try 
to lock current/subdir.../blk_id and previous/subdir../blk_id. If second 
lock fails, we can assume they are the same.


Raghu.

Nigel Daley wrote:

Raghu,

LGPL not allowed in distro:
http://people.apache.org/~cliffs/3party.html

Nige

On Nov 6, 2007, at 11:32 PM, Raghu Angadi wrote:







[jira] Commented: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Doug Cutting (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547343
 ] 

Doug Cutting commented on HADOOP-2288:
--

+1 This looks good to me.


> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130b.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547342
 ] 

Hadoop QA commented on HADOOP-2322:
---

-1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370706/ts-regionserver.patch
against trunk revision r599951.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests -1.  The patch failed core unit tests.

contrib tests +1.  The patch passed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1222/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1222/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1222/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1222/console

This message is automatically generated.

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2297) [Hbase Shell] System.exit() Handling in Jar command

2007-11-30 Thread Michael Bieniosek (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2297?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael Bieniosek updated HADOOP-2297:
--

Attachment: Capture.java

Hey Edward,

I think I figured out how to suppress System.exit and capture stdout/stderr.  

Here, System.exit throws a SecurityException, which can be caught above.  In 
this case, I catch in Thread.UncaughtExceptionHandler since I am running 
misbehaved threads.  

I also wrote a class that captures stdout/stderr.  Since java only lets me set 
one printStream to capture stdout per jvm, I have to check 
Thread.currentThread, and then decide where to write the captured output.

I am hoping to incorporate some of this code into my custom jetty server that 
submits hadoop jobs.

> [Hbase Shell] System.exit() Handling in Jar command
> ---
>
> Key: HADOOP-2297
> URL: https://issues.apache.org/jira/browse/HADOOP-2297
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: Edward Yoon
>Assignee: Edward Yoon
> Fix For: 0.16.0
>
> Attachments: 2297_v02.patch, 2297_v03.patch, Capture.java
>
>
> I'd like to block the exitVM by System.exit().
> Shell should terminate by "quit" command.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2323) JobTracker.close() prints stack traces for exceptions that are not errors

2007-11-30 Thread Jim Kellerman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Kellerman updated HADOOP-2323:
--

Status: Patch Available  (was: Open)

> JobTracker.close() prints stack traces for exceptions that are not errors
> -
>
> Key: HADOOP-2323
> URL: https://issues.apache.org/jira/browse/HADOOP-2323
> Project: Hadoop
>  Issue Type: Bug
>  Components: mapred
>Affects Versions: 0.16.0
>Reporter: Jim Kellerman
>Assignee: Jim Kellerman
> Fix For: 0.16.0
>
> Attachments: patch.txt
>
>
> JobTracker.close() prints a stack trace for an interrupted exception even 
> though it was the method that interrupted the thread that threw the 
> exception. For example:
> {code}
>   this.expireTrackers.stopTracker();
>   try {
> this.expireTrackersThread.interrupt();
> this.expireTrackersThread.join();
>   } catch (InterruptedException ex) {
> ex.printStackTrace();
>   }
> {code}
> Well of course it is going to catch an InterruptedException after it just 
> interrupted the thread!
> This is *not* an error and should  *not* be dumped to the logs!
> In other circumstances, catching InterruptedException is entirely 
> appropriate. Just not in close where you've told the thread to shutdown and 
> then interrupted it to ensure it does!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2323) JobTracker.close() prints stack traces for exceptions that are not errors

2007-11-30 Thread Jim Kellerman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2323?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Kellerman updated HADOOP-2323:
--

Attachment: patch.txt

> JobTracker.close() prints stack traces for exceptions that are not errors
> -
>
> Key: HADOOP-2323
> URL: https://issues.apache.org/jira/browse/HADOOP-2323
> Project: Hadoop
>  Issue Type: Bug
>  Components: mapred
>Affects Versions: 0.16.0
>Reporter: Jim Kellerman
>Assignee: Jim Kellerman
> Fix For: 0.16.0
>
> Attachments: patch.txt
>
>
> JobTracker.close() prints a stack trace for an interrupted exception even 
> though it was the method that interrupted the thread that threw the 
> exception. For example:
> {code}
>   this.expireTrackers.stopTracker();
>   try {
> this.expireTrackersThread.interrupt();
> this.expireTrackersThread.join();
>   } catch (InterruptedException ex) {
> ex.printStackTrace();
>   }
> {code}
> Well of course it is going to catch an InterruptedException after it just 
> interrupted the thread!
> This is *not* an error and should  *not* be dumped to the logs!
> In other circumstances, catching InterruptedException is entirely 
> appropriate. Just not in close where you've told the thread to shutdown and 
> then interrupted it to ensure it does!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-1841) IPC server should write repsonses asynchronously

2007-11-30 Thread Koji Noguchi (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-1841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547337
 ] 

Koji Noguchi commented on HADOOP-1841:
--

Just wanted to note that we had couple of occasions when one user 
1) Started many dfsclients on a single node
2) Load on that client node became too high that its rpc read speed slowed down 
significantly
3) Namenode became unresponsive 

Rebooting that client node brought back the namenode. 

I'm hoping this patch would solve such cases.

> IPC server should write repsonses asynchronously
> 
>
> Key: HADOOP-1841
> URL: https://issues.apache.org/jira/browse/HADOOP-1841
> Project: Hadoop
>  Issue Type: Improvement
>  Components: ipc
>Reporter: Doug Cutting
>Assignee: dhruba borthakur
> Attachments: asyncRPC-2.patch, asyncRPC.patch, asyncRPC.patch
>
>
> Hadoop's IPC Server currently writes responses from request handler threads 
> using blocking writes.  Performance and scalability might be improved if 
> responses were written asynchronously.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2185) Server ports: to roll or not to roll.

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547336
 ] 

dhruba borthakur commented on HADOOP-2185:
--

+1 code looks good.


> Server ports: to roll or not to roll.
> -
>
> Key: HADOOP-2185
> URL: https://issues.apache.org/jira/browse/HADOOP-2185
> Project: Hadoop
>  Issue Type: Improvement
>  Components: conf, dfs, mapred
>Affects Versions: 0.15.0
>Reporter: Konstantin Shvachko
>Assignee: Konstantin Shvachko
> Fix For: 0.16.0
>
> Attachments: FixedPorts.patch, FixedPorts2.patch
>
>
> Looked at the issues related to port rolling. My impression is that port 
> rolling is required only for the unit tests to run.
> Even the name-node port should roll there, which we don't have now, in order 
> to be able to start 2 cluster for testing say dist cp.
> For real clusters on the contrary port rolling is not desired and some times 
> even prohibited.
> So we should have a way of to ban port rolling. My proposition is to
> # use ephemeral port 0 if port rolling is desired
> # if a specific port is specified then port rolling should not happen at all, 
> meaning that a 
> server is either able or not able to start on that particular port.
> The desired port is specified via configuration parameters.
> - Name-node: fs.default.name = host:port
> - Data-node: dfs.datanode.port
> - Job-tracker: mapred.job.tracker = host:port
> - Task-tracker: mapred.task.tracker.report.bindAddress = host
>   Task-tracker currently does not have an option to specify port, it always 
> uses the ephemeral port 0, 
>   and therefore I propose to add one.
> - Secondary node does not need a port to listen on.
> For info servers we have two sets of config variables *.info.bindAddress and 
> *.info.port
> except for the task tracker, which calls them *.http.bindAddress and 
> *.http.port instead of "info".
> With respect to the info servers I propose to completely eliminate the port 
> parameters, and form 
> *.info.bindAddress = host:port
> Info servers should do the same thing, namely start or fail on the specified 
> port if it is not 0,
> and start on any free port if it is ephemeral.
> For the task-tracker I would rename tasktracker.http.bindAddress to 
> mapred.task.tracker.info.bindAddress
> For the data-node the info dfs.datanode.info.bindAddress should be included 
> into the default config.
> Is there a reason why it is not there?
> This is the summary of proposed changes:
> || Server || current name = value || proposed name = value ||
> | NameNode | fs.default.name = host:port | same |
> | | dfs.info.bindAddress = host | dfs.http.bindAddress = host:port |
> | DataNode | dfs.datanode.bindAddress = host | dfs.datanode.bindAddress = 
> host:port |
> | | dfs.datanode.port = port | eliminate |
> | | dfs.datanode.info.bindAddress = host | dfs.datanode.http.bindAddress = 
> host:port |
> | | dfs.datanode.info.port = port | eliminate |
> | JobTracker | mapred.job.tracker = host:port | same |
> | | mapred.job.tracker.info.bindAddress = host | 
> mapred.job.tracker.http.bindAddress = host:port |
> | | mapred.job.tracker.info.port = port | eliminate |
> | TaskTracker | mapred.task.tracker.report.bindAddress = host | 
> mapred.task.tracker.report.bindAddress = host:port |
> | | tasktracker.http.bindAddress = host | 
> mapred.task.tracker.http.bindAddress = host:port |
> | | tasktracker.http.port = port | eliminate |
> | SecondaryNameNode | dfs.secondary.info.bindAddress = host | 
> dfs.secondary.http.bindAddress = host:port |
> | | dfs.secondary.info.port = port | eliminate |
> Do we also want to set some uniform naming convention for the configuration 
> variables?
> Like having hdfs instead of dfs, or info instead of http, or systematically 
> using either datanode
> or data.node would make that look better in my opinion.
> So these are all +*api*+ changes. I would +*really*+ like some feedback on 
> this, especially from 
> people who deal with configuration issues on practice.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-1841) IPC server should write repsonses asynchronously

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-1841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547335
 ] 

dhruba borthakur commented on HADOOP-1841:
--

As expected, this patch does not decrease the performance of the namenode. But 
I agree with owen that we should determine that it does not affect sort 
performance too.

> IPC server should write repsonses asynchronously
> 
>
> Key: HADOOP-1841
> URL: https://issues.apache.org/jira/browse/HADOOP-1841
> Project: Hadoop
>  Issue Type: Improvement
>  Components: ipc
>Reporter: Doug Cutting
>Assignee: dhruba borthakur
> Attachments: asyncRPC-2.patch, asyncRPC.patch, asyncRPC.patch
>
>
> Hadoop's IPC Server currently writes responses from request handler threads 
> using blocking writes.  Performance and scalability might be improved if 
> responses were written asynchronously.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: 2288_20071130b.patch

2288_20071130b.patch:
In FileSystem, the original create(Path f, boolean overwrite, ...) is no longer 
abstract.  It call the new create(Path f, FsPermission permission, boolean 
overwrite, ...) by default.  Similar thing is done for mkdirs.



> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130b.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: (was: 2288_20071130.patch)

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2229) Provide a simple login implementation

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2229:
--

Attachment: ugi4.patch

Ok, now the proposal for umask to remove umask from both UGI and UnixUGI. 
Instead FSPermission (defined in HADOOP-2288) is going to provide get/set umask 
methods. The approach is that rather than getting a user's umask from UNIX, 
umask is going to be read from hadoop configuration. If it is not set in the 
configuration, it is set to a default value 0022.

This new patch reflects the proposed change.

> Provide a simple login implementation
> -
>
> Key: HADOOP-2229
> URL: https://issues.apache.org/jira/browse/HADOOP-2229
> Project: Hadoop
>  Issue Type: New Feature
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: ugi.patch, ugi1.patch, ugi2.patch, ugi3.patch, ugi4.patch
>
>
> Give a simple implementation of HADOOP-1701.Hadoop clients are assumed to 
> be started within a Unix-like network which provides user and group 
> management.  This implementation read user information from the OS and send 
> them to the NameNode in plaintexts through RPC (see also HADOOP-2184).  
> NameNode trusts all information given and uses them for permission checking.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-2303) [hbase] patch-build test failures (#1194, #1193, #1192)

2007-11-30 Thread Jim Kellerman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2303?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Kellerman resolved HADOOP-2303.
---

Resolution: Duplicate

This issue will be addressed by HADOOP-2324

> [hbase] patch-build test failures (#1194, #1193, #1192)
> ---
>
> Key: HADOOP-2303
> URL: https://issues.apache.org/jira/browse/HADOOP-2303
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: stack
>Priority: Minor
>
> TestRegionServerExit failes in #1194 like this:
> {code}
> ...
> [junit] 2007-11-29 00:20:53,775 INFO  [RegionServer:0] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:816): aborting 
> server at: 140.211.11.75:61555
> [junit] 2007-11-29 00:20:53,791 INFO  [RegionServer:2] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:809): On abort, 
> closed hlog
> [junit] 2007-11-29 00:20:53,791 INFO  [RegionServer:2] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:816): aborting 
> server at: 140.211.11.75:61754
> [junit] 2007-11-29 00:20:53,801 INFO  [HMaster] 
> org.apache.hadoop.hbase.HLog.splitLog(HLog.java:205): log file splitting 
> completed for /hbase/log_140.211.11.75_1196295616136_61556
> [junit] 2007-11-29 00:20:53,801 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,801 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,802 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,802 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,803 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,803 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,803 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,803 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,803 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,804 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,804 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,804 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 14

[jira] Created: (HADOOP-2324) [hbase] Fix assertion failures in TestTableMapReduce

2007-11-30 Thread Jim Kellerman (JIRA)
[hbase] Fix assertion failures in TestTableMapReduce


 Key: HADOOP-2324
 URL: https://issues.apache.org/jira/browse/HADOOP-2324
 Project: Hadoop
  Issue Type: Bug
  Components: contrib/hbase
Affects Versions: 0.16.0
Reporter: Jim Kellerman
Assignee: Jim Kellerman
 Fix For: 0.16.0


TestTableMapReduce consists of two tests: a very simple one that verifies that 
it is working, and a larger test that verifies that it works when the table 
consists of multiple regions.

There are two failure modes:

1) an assertion failure in the verify method

2) the test times out because something does not shut down cleanly.

to figure out what is causing these failures, the first step will be to do 
copious logging in the test. Once the root cause is determined, we can address 
it.

Unfortunately, some of the testing will have to be done via Hudson, because 
that is the only environment in which it fails. We routinely test under 
Windows, Fedora Core 6, MacOsX, and Ubuntu. Windows and Ubuntu have both single 
and dual processor configurations. FC6 and MacOS run on dual core machines.

Only Hudson seems crippled enough to demonstrate these problems. (A good thing 
since they are real problems)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2314:
--

Status: Patch Available  (was: Open)

Raghu, yes, this is the real fix for HADOOP-2256.

Dhruba, thanks a million for you help with the testing. :-)

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>  Components: dfs
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HADOOP-2314:
-

Component/s: dfs

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>  Components: dfs
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547322
 ] 

dhruba borthakur commented on HADOOP-2314:
--

This fix works great. +1


> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>  Components: dfs
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2303) [hbase] patch-build test failures (#1194, #1193, #1192)

2007-11-30 Thread Jim Kellerman (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2303?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547318
 ] 

Jim Kellerman commented on HADOOP-2303:
---

TestRegionServerExit fails in #1194 should now be fixed.

#1193 fails in TestTableMapReduce because following assertion fails: 
 {code}
junit.framework.AssertionFailedError
at 
org.apache.hadoop.hbase.mapred.TestTableMapReduce.verify(TestTableMapReduce.java:389)
at 
org.apache.hadoop.hbase.mapred.TestTableMapReduce.localTestSingleRegionTable(TestTableMapReduce.java:270)
{code}

in build #1217 it failed from another assertion error:

{code}
junit.framework.AssertionFailedError
at 
org.apache.hadoop.hbase.mapred.TestTableMapReduce.verify(TestTableMapReduce.java:389)
at 
org.apache.hadoop.hbase.mapred.TestTableMapReduce.localTestMultiRegionTable(TestTableMapReduce.java:322)
{code}

The exceptions:
{code}
[junit] 2007-11-30 04:23:11,357 ERROR [expireTrackers] 
org.apache.hadoop.mapred.JobTracker$ExpireTrackers.run(JobTracker.java:308): 
Tracker Expiry Thread got exception: java.lang.InterruptedException: sleep 
interrupted
[junit] at java.lang.Thread.sleep(Native Method)
[junit] at 
org.apache.hadoop.mapred.JobTracker$ExpireTrackers.run(JobTracker.java:263)
[junit] at java.lang.Thread.run(Thread.java:595)

[junit] 2007-11-30 04:23:11,358 WARN  [Task Commit Thread] 
org.apache.hadoop.mapred.JobTracker$TaskCommitQueue.run(JobTracker.java:2017): 
Task Commit Thread exiting, got interrupted: java.lang.InterruptedException
[junit] at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.java:1815)
[junit] at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1850)
[junit] at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:359)
[junit] at 
org.apache.hadoop.mapred.JobTracker$TaskCommitQueue.run(JobTracker.java:1919)
{code}
are red herrings because of the the way JobTracker shuts down. (See HADOOP-2323)

> [hbase] patch-build test failures (#1194, #1193, #1192)
> ---
>
> Key: HADOOP-2303
> URL: https://issues.apache.org/jira/browse/HADOOP-2303
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: stack
>Priority: Minor
>
> TestRegionServerExit failes in #1194 like this:
> {code}
> ...
> [junit] 2007-11-29 00:20:53,775 INFO  [RegionServer:0] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:816): aborting 
> server at: 140.211.11.75:61555
> [junit] 2007-11-29 00:20:53,791 INFO  [RegionServer:2] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:809): On abort, 
> closed hlog
> [junit] 2007-11-29 00:20:53,791 INFO  [RegionServer:2] 
> org.apache.hadoop.hbase.HRegionServer.run(HRegionServer.java:816): aborting 
> server at: 140.211.11.75:61754
> [junit] 2007-11-29 00:20:53,801 INFO  [HMaster] 
> org.apache.hadoop.hbase.HLog.splitLog(HLog.java:205): log file splitting 
> completed for /hbase/log_140.211.11.75_1196295616136_61556
> [junit] 2007-11-29 00:20:53,801 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,801 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,802 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29 00:20:53,802 INFO  [HMaster] 
> org.apache.hadoop.hbase.HMaster$ProcessServerShutdown.process(HMaster.java:1999):
>  process shutdown of server 140.211.11.75:61556: logSplit: true, rootChecked: 
> false, rootRescanned: false, numberOfMetaRegions: 1, 
> onlineMetaRegions.size(): 1
> [junit] 2007-11-29 00:20:53,802 DEBUG [HMaster] 
> org.apache.hadoop.hbase.HMaster.run(HMaster.java:1068): Main processing loop: 
> ProcessServerShutdown of 140.211.11.75:61556
> [junit] 2007-11-29

[jira] Created: (HADOOP-2323) JobTracker.close() prints stack traces for exceptions that are not errors

2007-11-30 Thread Jim Kellerman (JIRA)
JobTracker.close() prints stack traces for exceptions that are not errors
-

 Key: HADOOP-2323
 URL: https://issues.apache.org/jira/browse/HADOOP-2323
 Project: Hadoop
  Issue Type: Bug
  Components: mapred
Affects Versions: 0.16.0
Reporter: Jim Kellerman
Assignee: Jim Kellerman
 Fix For: 0.16.0


JobTracker.close() prints a stack trace for an interrupted exception even 
though it was the method that interrupted the thread that threw the exception. 
For example:

{code}
  this.expireTrackers.stopTracker();
  try {
this.expireTrackersThread.interrupt();
this.expireTrackersThread.join();
  } catch (InterruptedException ex) {
ex.printStackTrace();
  }
{code}

Well of course it is going to catch an InterruptedException after it just 
interrupted the thread!

This is *not* an error and should  *not* be dumped to the logs!

In other circumstances, catching InterruptedException is entirely appropriate. 
Just not in close where you've told the thread to shutdown and then interrupted 
it to ensure it does!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547313
 ] 

Hadoop QA commented on HADOOP-2224:
---

+1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370699/ts-regionserver.patch
against trunk revision r599951.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests +1.  The patch passed core unit tests.

contrib tests +1.  The patch passed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1221/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1221/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1221/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1221/console

This message is automatically generated.

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2278) Streaming: better conrol over input splits

2007-11-30 Thread arkady borkovsky (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547312
 ] 

arkady borkovsky commented on HADOOP-2278:
--

Some split conditions are hard to catch by a fixed patter.

A very important split condition is "key switch" -- (need to be fiels as a 
separate issue?)

Quite often, the mapper input is grouped by key and the mapper is actually a 
reducer.  Therefore, it expects that all the values for given key go to the 
same task.
Currently, the split happens between any two records, so "key runs" are usually 
broken at split boundaries.

The work around is to have infinite split size -- which creates bad granularity.

> Streaming: better conrol over input splits
> --
>
> Key: HADOOP-2278
> URL: https://issues.apache.org/jira/browse/HADOOP-2278
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/streaming
>Reporter: arkady borkovsky
>
> In steaming, the map command usually expect to receive it's input 
> uninterpreted -- just as it is stored in DFS.
> However, the split (the beginning and the end of the portion of data that 
> goes to a single map task) is often important and is not "any line break".
> Often the input consists of multi-line docments -- e.g. in XML.
> There should be a way to specify a pattern that separates logical records.
> Existing "Streaming XML record reader" kind of provides this functionality.  
> However, it is accepted that "Streaming XML" is a hack and needs to be 
> replaced 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2322:
--

Status: Patch Available  (was: Open)

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2224:
--

Attachment: (was: ts-regionserver.patch)

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury resolved HADOOP-2224.
---

Resolution: Fixed

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2224:
--

Status: Open  (was: Patch Available)

Decided to open a new issue for the fix instead of reopening the existing one. 
See HADOOP-2322

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury reassigned HADOOP-2322:
-

Assignee: Bryan Duxbury

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread Bryan Duxbury (JIRA)
[hbase] getRow(row, TS) client interface not properly connected
---

 Key: HADOOP-2322
 URL: https://issues.apache.org/jira/browse/HADOOP-2322
 Project: Hadoop
  Issue Type: Bug
  Components: contrib/hbase
Reporter: Bryan Duxbury
Priority: Minor
 Attachments: ts-regionserver.patch

There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2321) Streaming: better support for command lines or streaming command

2007-11-30 Thread arkady borkovsky (JIRA)
Streaming: better support for command lines or streaming command


 Key: HADOOP-2321
 URL: https://issues.apache.org/jira/browse/HADOOP-2321
 Project: Hadoop
  Issue Type: Improvement
  Components: contrib/streaming
Reporter: arkady borkovsky


Quite often, the command line for streaming mapper or reducer needs to use one 
or two levels of quotes.
This make it inconvenient or impossible to pass the commands in the streaming 
command line.
It would be good to have streaming take its specification from a file -- 
especially as longer streaming commands are not typed in, but are either run 
from files (shell scripts) or generated by other processors.

The current work around is to separate files for the mapper command, for the 
reducer command, and for the streaming command itself.  This works, but is 
inconvenient and quite error-prone.
Having just one file with all three would be good.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2322) [hbase] getRow(row, TS) client interface not properly connected

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2322?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2322:
--

Attachment: ts-regionserver.patch

This patch resolves the problem.

> [hbase] getRow(row, TS) client interface not properly connected
> ---
>
> Key: HADOOP-2322
> URL: https://issues.apache.org/jira/browse/HADOOP-2322
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: ts-regionserver.patch
>
>
> There was a bug in the patch for HADOOP-2224 that was causing getRow(row, ts) 
> calls to always return the most recent cells.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2224:
--

Status: Patch Available  (was: Reopened)

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch, 
> ts-regionserver.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2224:
--

Attachment: ts-regionserver.patch

This patch adds the correct overloaded getRow call to HRegionServer.java.

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch, 
> ts-regionserver.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Issue Comment Edited: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547292
 ] 

bryanduxbury edited comment on HADOOP-2224 at 11/30/07 1:17 PM:
-

The HRegionServer class had a broken implementation of the new method. 

  was (Author: bryanduxbury):
The HRegionServer class had a broken implementation of the new class. 
  
> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Reopened: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury reopened HADOOP-2224:
---


The HRegionServer class had a broken implementation of the new class. 

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547287
 ] 

dhruba borthakur commented on HADOOP-2314:
--

I am testing this right now.

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-1792) df command doesn't exist under windows

2007-11-30 Thread Olivier Dagenais (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-1792?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547282
 ] 

Olivier Dagenais commented on HADOOP-1792:
--

I got this problem too:  It appears DF is being used to compute available disk 
space before creating files.  While Java does not appears to have support for 
this (see [bug 
4057701|http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=4057701]), it does 
appear that another Apache project (sort of) does:

[org.apache.commons.FileSystemUtils.freeSpaceKb|http://commons.apache.org/io/api-release/org/apache/commons/io/FileSystemUtils.html#freeSpaceKb(java.lang.String)]

...at least it wouldn't require that Hadoop users on Windows install Cygwin.

> df command doesn't exist under windows
> --
>
> Key: HADOOP-1792
> URL: https://issues.apache.org/jira/browse/HADOOP-1792
> Project: Hadoop
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 0.14.0
>Reporter: Benjamin Francisoud
>
> My code use to work with previous version of hadoop, I upgraded to 0.14 and 
> now:
> java.io.IOException: CreateProcess: df -k "C:\Documents and 
> Settings\Benjamin\Local Settings\Temp\test14906test\mapredLocal" error=2
>   at java.lang.ProcessImpl.create(Native Method)
>   at java.lang.ProcessImpl.(Unknown Source)
>   at java.lang.ProcessImpl.start(Unknown Source)
>   at java.lang.ProcessBuilder.start(Unknown Source)
>   at java.lang.Runtime.exec(Unknown Source)
>   at java.lang.Runtime.exec(Unknown Source)
>   at org.apache.hadoop.fs.DF.doDF(DF.java:60)
>   at org.apache.hadoop.fs.DF.(DF.java:53)
>   at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(LocalDirAllocator.java:198)
>   at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:235)
>   at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124)
>   at 
> org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:88)
>   at 
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpillToDisk(MapTask.java:373)
>   at 
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:593)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:190)
>   at 
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:137)
>   at 
> org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:283)
>   at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:397)
> ...

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-2320) [hbase] Committed TestGet2 is managled (breaks build).

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack resolved HADOOP-2320.
---

   Resolution: Fixed
Fix Version/s: 0.16.0

Committed.

> [hbase] Committed TestGet2 is managled (breaks build).
> --
>
> Key: HADOOP-2320
> URL: https://issues.apache.org/jira/browse/HADOOP-2320
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: stack
> Fix For: 0.16.0
>
> Attachments: tg2.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2320) [hbase] Committed TestGet2 is managled (breaks build).

2007-11-30 Thread stack (JIRA)
[hbase] Committed TestGet2 is managled (breaks build).
--

 Key: HADOOP-2320
 URL: https://issues.apache.org/jira/browse/HADOOP-2320
 Project: Hadoop
  Issue Type: Bug
  Components: contrib/hbase
Reporter: stack
 Attachments: tg2.patch



-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2320) [hbase] Committed TestGet2 is managled (breaks build).

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2320?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2320:
--

Attachment: tg2.patch

Add license.  Removed doubled instance of class.

> [hbase] Committed TestGet2 is managled (breaks build).
> --
>
> Key: HADOOP-2320
> URL: https://issues.apache.org/jira/browse/HADOOP-2320
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: stack
> Attachments: tg2.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2311) [hbase] Could not complete hdfs write out to flush file forcing regionserver restart

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2311:
--

Attachment: delete-logging.patch

Log all deletes.

> [hbase] Could not complete hdfs write out to flush file forcing regionserver 
> restart
> 
>
> Key: HADOOP-2311
> URL: https://issues.apache.org/jira/browse/HADOOP-2311
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: stack
>Priority: Minor
> Attachments: delete-logging.patch
>
>
> I've spent some time looking into this issue but there are not enough clues 
> in the logs to tell where the problem is. Here's what I know.
> Two region servers went down last night, a minute apart, during Paul Saab's 
> 6hr run inserting 300million rows into hbase. The regionservers went down to 
> force rerun of hlog and avoid possible data loss after a failure writing 
> memory flushes to hdfs.
> Here is the lead up to the failed flush:
> ...
> 2007-11-28 22:40:02,231 INFO  hbase.HRegionServer - MSG_REGION_OPEN : 
> regionname: postlog,img149/4699/133lm0.jpg,1196318393738, startKey: 
> , tableDesc: {name: postlog, families: 
> {cookie:={name: cookie, max versions: 1, compression: NONE, in memory: false, 
> max length: 2147483647, bloom filter: none}, ip:={name: ip, max versions: 1, 
> compression: NONE, in memory: false, max length: 2147483647, bloom filter: 
> none}}}
> 2007-11-28 22:40:02,242 DEBUG hbase.HStore - starting 1703405830/cookie (no 
> reconstruction log)
> 2007-11-28 22:40:02,741 DEBUG hbase.HStore - maximum sequence id for hstore 
> 1703405830/cookie is 29077708
> 2007-11-28 22:40:03,094 DEBUG hbase.HStore - starting 1703405830/ip (no 
> reconstruction log)
> 2007-11-28 22:40:03,852 DEBUG hbase.HStore - maximum sequence id for hstore 
> 1703405830/ip is 29077708
> 2007-11-28 22:40:04,138 DEBUG hbase.HRegion - Next sequence id for region 
> postlog,img149/4699/133lm0.jpg,1196318393738 is 29077709
> 2007-11-28 22:40:04,141 INFO  hbase.HRegion - region 
> postlog,img149/4699/133lm0.jpg,1196318393738 available
> 2007-11-28 22:40:04,141 DEBUG hbase.HLog - changing sequence number from 
> 21357623 to 29077709
> 2007-11-28 22:40:04,141 INFO  hbase.HRegionServer - MSG_REGION_OPEN : 
> regionname: postlog,img149/7512/dscnlightenedfi3.jpg,1196318393739, 
> startKey: , tableDesc: {name: postlog, 
> families: {cookie:={name: cookie, max versions: 1, compression: NONE, in 
> memory: false, max length: 2147483647, bloom filter: none}, ip:={name: ip, 
> max versions: 1, compression: NONE, in memory: false, max length: 2147483647, 
> bloom filter: none}}}
> 2007-11-28 22:40:04,145 DEBUG hbase.HStore - starting 376748222/cookie (no 
> reconstruction log)
> 2007-11-28 22:40:04,223 DEBUG hbase.HStore - maximum sequence id for hstore 
> 376748222/cookie is 29077708
> 2007-11-28 22:40:04,277 DEBUG hbase.HStore - starting 376748222/ip (no 
> reconstruction log)
> 2007-11-28 22:40:04,353 DEBUG hbase.HStore - maximum sequence id for hstore 
> 376748222/ip is 29077708
> 2007-11-28 22:40:04,699 DEBUG hbase.HRegion - Next sequence id for region 
> postlog,img149/7512/dscnlightenedfi3.jpg,1196318393739 is 29077709
> 2007-11-28 22:40:04,701 INFO  hbase.HRegion - region 
> postlog,img149/7512/dscnlightenedfi3.jpg,1196318393739 available
> 2007-11-28 22:40:34,427 DEBUG hbase.HRegionServer - flushing region 
> postlog,img143/1310/yashrk3.jpg,1196317258704
> 2007-11-28 22:40:34,428 DEBUG hbase.HRegion - Not flushing cache for region 
> postlog,img143/1310/yashrk3.jpg,1196317258704: snapshotMemcaches() determined 
> that there was nothing to do
> 2007-11-28 22:40:55,745 DEBUG hbase.HRegionServer - flushing region 
> postlog,img142/8773/1001417zc4.jpg,1196317258703
> 2007-11-28 22:40:55,745 DEBUG hbase.HRegion - Not flushing cache for region 
> postlog,img142/8773/1001417zc4.jpg,1196317258703: snapshotMemcaches() 
> determined that there was nothing to do
> 2007-11-28 22:41:04,144 DEBUG hbase.HRegionServer - flushing region 
> postlog,img149/4699/133lm0.jpg,1196318393738
> 2007-11-28 22:41:04,144 DEBUG hbase.HRegion - Started memcache flush for 
> region postlog,img149/4699/133lm0.jpg,1196318393738. Size 74.7k
> 2007-11-28 22:41:04,764 DEBUG hbase.HStore - Added 
> 1703405830/ip/610047924323344967 with sequence id 29081563 and size 53.8k
> 2007-11-28 22:41:04,902 DEBUG hbase.HStore - Added 
> 1703405830/cookie/3147798053949544972 with sequence id 29081563 and size 41.3k
> 2007-11-28 22:41:04,902 DEBUG hbase.HRegion - Finished memcache flush for 
> region postlog,img149/4699/133lm0.jpg,1196318393738 in 758ms, 
> sequenceid=29081563
> 2007-11-28 22:41:04,902 DEBUG hbase.HStore - compaction for HStore 
> postlog,img149/4699/133lm0.jpg,1196318393738/ip needed.
> 2007-11-28 22:41:04,903 DEBUG hbase.H

[jira] Commented: (HADOOP-2068) [hbase] RESTful interface

2007-11-30 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2068?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547268
 ] 

stack commented on HADOOP-2068:
---

Thanks for the patch Bryan.

> [hbase] RESTful interface
> -
>
> Key: HADOOP-2068
> URL: https://issues.apache.org/jira/browse/HADOOP-2068
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Assignee: Bryan Duxbury
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: rest-11-27-07-v2.patch, rest-11-27-07.3.patc, 
> rest-11-27-07.patch, rest-11-28-07.2.patch, rest-11-28-07.3.patch, 
> rest-11-28-07.patch, rest.patch
>
>
> A RESTful interface would be one means of making hbase accessible to clients 
> that are not java.  It might look something like the below:
> + An HTTP GET of  http://MASTER:PORT/ outputs the master's attributes: online 
> meta regions, list of tables, etc.: i.e. what you see now when you go to 
> http://MASTER:PORT/master.jsp.
> + An HTTP GET of http://MASTER:PORT/TABLENAME: 200 if tables exists and 
> HTableDescription (mimetype: text/plain or text/xml) or 401 if no such table. 
>  HTTP DELETE would drop the table.  HTTP PUT would add one.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW: 200 if row exists and 401 
> if not.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNFAMILY: 
> HColumnDescriptor (mimetype: text/plain or text/xml) or 401 if no such table.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/: 200 and latest 
> version (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE 
> would delete the cell.  HTTP PUT would add a new version.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/TIMESTAMP: 200 
> (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE would 
> remove.  HTTP PUT would put this record.
> + Browser originally goes against master but master then redirects to the 
> hosting region server to serve, update, delete, etc. the addressed cell

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2068) [hbase] RESTful interface

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2068:
--

   Resolution: Fixed
Fix Version/s: 0.16.0
   Status: Resolved  (was: Patch Available)

Committed (Failed tests were unrelated to this patch which doesn't add any new 
tests and is code that doesn't run at unit test time).  Resolving.

> [hbase] RESTful interface
> -
>
> Key: HADOOP-2068
> URL: https://issues.apache.org/jira/browse/HADOOP-2068
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Assignee: Bryan Duxbury
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: rest-11-27-07-v2.patch, rest-11-27-07.3.patc, 
> rest-11-27-07.patch, rest-11-28-07.2.patch, rest-11-28-07.3.patch, 
> rest-11-28-07.patch, rest.patch
>
>
> A RESTful interface would be one means of making hbase accessible to clients 
> that are not java.  It might look something like the below:
> + An HTTP GET of  http://MASTER:PORT/ outputs the master's attributes: online 
> meta regions, list of tables, etc.: i.e. what you see now when you go to 
> http://MASTER:PORT/master.jsp.
> + An HTTP GET of http://MASTER:PORT/TABLENAME: 200 if tables exists and 
> HTableDescription (mimetype: text/plain or text/xml) or 401 if no such table. 
>  HTTP DELETE would drop the table.  HTTP PUT would add one.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW: 200 if row exists and 401 
> if not.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNFAMILY: 
> HColumnDescriptor (mimetype: text/plain or text/xml) or 401 if no such table.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/: 200 and latest 
> version (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE 
> would delete the cell.  HTTP PUT would add a new version.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/TIMESTAMP: 200 
> (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE would 
> remove.  HTTP PUT would put this record.
> + Browser originally goes against master but master then redirects to the 
> hosting region server to serve, update, delete, etc. the addressed cell

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2068) [hbase] RESTful interface

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2068?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547263
 ] 

Hadoop QA commented on HADOOP-2068:
---

-1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370483/rest-11-28-07.3.patch
against trunk revision r599879.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests +1.  The patch passed core unit tests.

contrib tests -1.  The patch failed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1220/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1220/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1220/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1220/console

This message is automatically generated.

> [hbase] RESTful interface
> -
>
> Key: HADOOP-2068
> URL: https://issues.apache.org/jira/browse/HADOOP-2068
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: rest-11-27-07-v2.patch, rest-11-27-07.3.patc, 
> rest-11-27-07.patch, rest-11-28-07.2.patch, rest-11-28-07.3.patch, 
> rest-11-28-07.patch, rest.patch
>
>
> A RESTful interface would be one means of making hbase accessible to clients 
> that are not java.  It might look something like the below:
> + An HTTP GET of  http://MASTER:PORT/ outputs the master's attributes: online 
> meta regions, list of tables, etc.: i.e. what you see now when you go to 
> http://MASTER:PORT/master.jsp.
> + An HTTP GET of http://MASTER:PORT/TABLENAME: 200 if tables exists and 
> HTableDescription (mimetype: text/plain or text/xml) or 401 if no such table. 
>  HTTP DELETE would drop the table.  HTTP PUT would add one.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW: 200 if row exists and 401 
> if not.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNFAMILY: 
> HColumnDescriptor (mimetype: text/plain or text/xml) or 401 if no such table.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/: 200 and latest 
> version (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE 
> would delete the cell.  HTTP PUT would add a new version.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/TIMESTAMP: 200 
> (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE would 
> remove.  HTTP PUT would put this record.
> + Browser originally goes against master but master then redirects to the 
> hosting region server to serve, update, delete, etc. the addressed cell

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2298) ant target without source and docs

2007-11-30 Thread Doug Cutting (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2298?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547258
 ] 

Doug Cutting commented on HADOOP-2298:
--

We could provide an ant target that builds a binary tarball, but still release 
the compound tarball.  Folks could then run the binary-only target from within 
a release if they want to build a binary-only tarball.  Might that suffice?

Producing only a single file per release simplifies things.  But we don't want 
users to have to recompile releases, and writing a binary-only tarball target 
that doesn't recompile might be tricky.  I don't yet have a strong sense about 
the best way to proceed...


> ant target without source and docs 
> ---
>
> Key: HADOOP-2298
> URL: https://issues.apache.org/jira/browse/HADOOP-2298
> Project: Hadoop
>  Issue Type: Improvement
>  Components: build
>Reporter: Gautam Kowshik
>
> Can we have an ant target or a -D option to build the hadoop tar without the 
> source and documentation? This brings down the tar size from 11.5 MB to 5.6 
> MB. This would speed up distribution. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2319) Build both 32 and 64 bit native libraries when compiling with a 64 bit JVM

2007-11-30 Thread Nigel Daley (JIRA)
Build both 32 and 64 bit native libraries when compiling with a 64 bit JVM
--

 Key: HADOOP-2319
 URL: https://issues.apache.org/jira/browse/HADOOP-2319
 Project: Hadoop
  Issue Type: Improvement
  Components: build
Reporter: Nigel Daley
Priority: Minor


When a 32 bit JVM is used to build Hadoop, the 32 bit  native libraries are 
build (providing -Dcompile.native=true is present).  Likewise, a 64 bit JVM 
automatically builds a 64 bit native library.

It would be helpful if a 64 bit JVM built both 32 and 64 bit native libraries.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Raghu Angadi (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547255
 ] 

Raghu Angadi commented on HADOOP-2314:
--

Does this fix the problem seen in HADOOP-2256?

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2318) All C++ builds should use the autoconf tools

2007-11-30 Thread Nigel Daley (JIRA)
All C++ builds should use the autoconf tools


 Key: HADOOP-2318
 URL: https://issues.apache.org/jira/browse/HADOOP-2318
 Project: Hadoop
  Issue Type: Improvement
  Components: build
Reporter: Nigel Daley
Priority: Minor


Currently we have -Dcompile.native and -Dcompile.c++ build flags.  In addition, 
builds for pipes and libhadoop use autoconf tools, but libhdfs does not, nor 
does 64bit libhdfs compile work.

All these builds should use autoconf tools, support 64bit compilation, and 
should occur when a single flag is present (-Dcompile.c++ seems like the better 
choice).



-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HADOOP-2068) [hbase] RESTful interface

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury reassigned HADOOP-2068:
-

Assignee: Bryan Duxbury

> [hbase] RESTful interface
> -
>
> Key: HADOOP-2068
> URL: https://issues.apache.org/jira/browse/HADOOP-2068
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Assignee: Bryan Duxbury
>Priority: Minor
> Attachments: rest-11-27-07-v2.patch, rest-11-27-07.3.patc, 
> rest-11-27-07.patch, rest-11-28-07.2.patch, rest-11-28-07.3.patch, 
> rest-11-28-07.patch, rest.patch
>
>
> A RESTful interface would be one means of making hbase accessible to clients 
> that are not java.  It might look something like the below:
> + An HTTP GET of  http://MASTER:PORT/ outputs the master's attributes: online 
> meta regions, list of tables, etc.: i.e. what you see now when you go to 
> http://MASTER:PORT/master.jsp.
> + An HTTP GET of http://MASTER:PORT/TABLENAME: 200 if tables exists and 
> HTableDescription (mimetype: text/plain or text/xml) or 401 if no such table. 
>  HTTP DELETE would drop the table.  HTTP PUT would add one.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW: 200 if row exists and 401 
> if not.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNFAMILY: 
> HColumnDescriptor (mimetype: text/plain or text/xml) or 401 if no such table.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/: 200 and latest 
> version (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE 
> would delete the cell.  HTTP PUT would add a new version.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/TIMESTAMP: 200 
> (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE would 
> remove.  HTTP PUT would put this record.
> + Browser originally goes against master but master then redirects to the 
> hosting region server to serve, update, delete, etc. the addressed cell

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2315) [hbase] REST servlet doesn't treat / characters in row key correctly

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2315?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2315:
--

Priority: Trivial  (was: Major)

> [hbase] REST servlet doesn't treat / characters in row key correctly
> 
>
> Key: HADOOP-2315
> URL: https://issues.apache.org/jira/browse/HADOOP-2315
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Trivial
>
> Using row keys like "com.site.www/:http" currently doesn't work. We've 
> tracked it down to the use of request.getPathInfo() instead of 
> request.getRequestURI() in Dispatcher.getPathSegments.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Resolved: (HADOOP-2109) [hbase] TestRegionServerExit failures in patch build #1004, #1003, #991, and #971

2007-11-30 Thread Jim Kellerman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Kellerman resolved HADOOP-2109.
---

Resolution: Fixed

This test hasn't failed in a while. Resolving issue.

Should it start failing again,  we'll open a new issue.

> [hbase] TestRegionServerExit failures in patch build #1004, #1003, #991, and 
> #971
> -
>
> Key: HADOOP-2109
> URL: https://issues.apache.org/jira/browse/HADOOP-2109
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Affects Versions: 0.16.0
>Reporter: stack
> Fix For: 0.16.0
>
> Attachments: patch.txt, patch.txt, patch.txt, patch.txt, patch.txt
>
>
> Figure why the failures.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2316) [hbase] Run REST servlet outside of master

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2316:
--

Priority: Minor  (was: Major)

> [hbase] Run REST servlet outside of master
> --
>
> Key: HADOOP-2316
> URL: https://issues.apache.org/jira/browse/HADOOP-2316
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>Priority: Minor
>
> In order to support desired deployment strategy, we need to be able to run 
> the REST servlet independently of the master info server. We should add an 
> new option to the bin/hbase command ("rest"?) that optionally takes a port 
> and bind address and starts the servlet outside of any other project.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2317) [hbase] Support multipart/related instead of xml input/output

2007-11-30 Thread Bryan Duxbury (JIRA)
[hbase] Support multipart/related instead of xml input/output
-

 Key: HADOOP-2317
 URL: https://issues.apache.org/jira/browse/HADOOP-2317
 Project: Hadoop
  Issue Type: New Feature
  Components: contrib/hbase
Reporter: Bryan Duxbury
Assignee: Bryan Duxbury
Priority: Trivial


XML is bulky and slow, but the only way currently supported for multi-column 
gets/puts/etc. We should add support for multipart/related encoded entity 
bodies since that will be all binary and presumably faster and more compact.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HADOOP-2316) [hbase] Run REST servlet outside of master

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2316?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury reassigned HADOOP-2316:
-

Assignee: Bryan Duxbury

> [hbase] Run REST servlet outside of master
> --
>
> Key: HADOOP-2316
> URL: https://issues.apache.org/jira/browse/HADOOP-2316
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>
> In order to support desired deployment strategy, we need to be able to run 
> the REST servlet independently of the master info server. We should add an 
> new option to the bin/hbase command ("rest"?) that optionally takes a port 
> and bind address and starts the servlet outside of any other project.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2316) [hbase] Run REST servlet outside of master

2007-11-30 Thread Bryan Duxbury (JIRA)
[hbase] Run REST servlet outside of master
--

 Key: HADOOP-2316
 URL: https://issues.apache.org/jira/browse/HADOOP-2316
 Project: Hadoop
  Issue Type: New Feature
  Components: contrib/hbase
Reporter: Bryan Duxbury


In order to support desired deployment strategy, we need to be able to run the 
REST servlet independently of the master info server. We should add an new 
option to the bin/hbase command ("rest"?) that optionally takes a port and bind 
address and starts the servlet outside of any other project.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2314:
--

Attachment: block.patch

Here is a patch for fixing the problem. I've run this for 50 times in a row and 
havenot see any timeout.

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2315) [hbase] REST servlet doesn't treat / characters in row key correctly

2007-11-30 Thread Bryan Duxbury (JIRA)
[hbase] REST servlet doesn't treat / characters in row key correctly


 Key: HADOOP-2315
 URL: https://issues.apache.org/jira/browse/HADOOP-2315
 Project: Hadoop
  Issue Type: Bug
  Components: contrib/hbase
Reporter: Bryan Duxbury


Using row keys like "com.site.www/:http" currently doesn't work. We've tracked 
it down to the use of request.getPathInfo() instead of request.getRequestURI() 
in Dispatcher.getPathSegments.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: 2288_20071130.patch

2288_20071130.patch: Change applyMask(...) to applyUMask(...) since it is 
useful to HADOOP-2229.

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
> Attachments: 2288_20071130.patch
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HADOOP-2315) [hbase] REST servlet doesn't treat / characters in row key correctly

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2315?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury reassigned HADOOP-2315:
-

Assignee: Bryan Duxbury

> [hbase] REST servlet doesn't treat / characters in row key correctly
> 
>
> Key: HADOOP-2315
> URL: https://issues.apache.org/jira/browse/HADOOP-2315
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Reporter: Bryan Duxbury
>Assignee: Bryan Duxbury
>
> Using row keys like "com.site.www/:http" currently doesn't work. We've 
> tracked it down to the use of request.getPathInfo() instead of 
> request.getRequestURI() in Dispatcher.getPathSegments.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HADOOP-2314:
--

Fix Version/s: (was: 0.15.2)
   0.16.0

> TestBlockReplacement occasionally get into an infinite loop
> ---
>
> Key: HADOOP-2314
> URL: https://issues.apache.org/jira/browse/HADOOP-2314
> Project: Hadoop
>  Issue Type: Bug
>Affects Versions: 0.15.1
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.16.0
>
> Attachments: block.patch
>
>
> It turns out that in the case that tests an invalid deletion hint, either the 
> newNode or source may be choosen to be deleted as an exessive replica since 
> both of the nodes are on the same rack. The test assumes that only newNode 
> will be deleted and wait for its deletion. This causes an infinite loop when 
> source is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HADOOP-2314) TestBlockReplacement occasionally get into an infinite loop

2007-11-30 Thread Hairong Kuang (JIRA)
TestBlockReplacement occasionally get into an infinite loop
---

 Key: HADOOP-2314
 URL: https://issues.apache.org/jira/browse/HADOOP-2314
 Project: Hadoop
  Issue Type: Bug
Affects Versions: 0.15.1
Reporter: Hairong Kuang
Assignee: Hairong Kuang
 Fix For: 0.15.2


It turns out that in the case that tests an invalid deletion hint, either the 
newNode or source may be choosen to be deleted as an exessive replica since 
both of the nodes are on the same rack. The test assumes that only newNode will 
be deleted and wait for its deletion. This causes an infinite loop when source 
is chosen to be deleted.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2288) Change FileSystem API to support access control.

2007-11-30 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2288?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HADOOP-2288:
---

Attachment: (was: 2288_20071129b.patch)

> Change FileSystem API to support access control.
> 
>
> Key: HADOOP-2288
> URL: https://issues.apache.org/jira/browse/HADOOP-2288
> Project: Hadoop
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 0.15.0
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.16.0
>
>
> - Some FileSystem methods like create and mkdir need an additional parameter 
> for permission.
> - FileSystem has to provide methods for setting permission, changing 
> ownership, etc.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2068) [hbase] RESTful interface

2007-11-30 Thread Bryan Duxbury (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bryan Duxbury updated HADOOP-2068:
--

Status: Patch Available  (was: Open)

Submitting preliminary REST implementation.

> [hbase] RESTful interface
> -
>
> Key: HADOOP-2068
> URL: https://issues.apache.org/jira/browse/HADOOP-2068
> Project: Hadoop
>  Issue Type: New Feature
>  Components: contrib/hbase
>Reporter: stack
>Priority: Minor
> Attachments: rest-11-27-07-v2.patch, rest-11-27-07.3.patc, 
> rest-11-27-07.patch, rest-11-28-07.2.patch, rest-11-28-07.3.patch, 
> rest-11-28-07.patch, rest.patch
>
>
> A RESTful interface would be one means of making hbase accessible to clients 
> that are not java.  It might look something like the below:
> + An HTTP GET of  http://MASTER:PORT/ outputs the master's attributes: online 
> meta regions, list of tables, etc.: i.e. what you see now when you go to 
> http://MASTER:PORT/master.jsp.
> + An HTTP GET of http://MASTER:PORT/TABLENAME: 200 if tables exists and 
> HTableDescription (mimetype: text/plain or text/xml) or 401 if no such table. 
>  HTTP DELETE would drop the table.  HTTP PUT would add one.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW: 200 if row exists and 401 
> if not.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNFAMILY: 
> HColumnDescriptor (mimetype: text/plain or text/xml) or 401 if no such table.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/: 200 and latest 
> version (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE 
> would delete the cell.  HTTP PUT would add a new version.
> + An HTTP GET of http://MASTER:PORT/TABLENAME/ROW/COLUMNNAME/TIMESTAMP: 200 
> (mimetype: binary/octet-stream) or 401 if no such cell. HTTP DELETE would 
> remove.  HTTP PUT would put this record.
> + Browser originally goes against master but master then redirects to the 
> hosting region server to serve, update, delete, etc. the addressed cell

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2313) build does not fail when libhdfs build fails

2007-11-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2313?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547239
 ] 

Hadoop QA commented on HADOOP-2313:
---

+1 overall.  Here are the results of testing the latest attachment 
http://issues.apache.org/jira/secure/attachment/12370680/HADOOP-2313.patch
against trunk revision r599879.

@author +1.  The patch does not contain any @author tags.

javadoc +1.  The javadoc tool did not generate any warning messages.

javac +1.  The applied patch does not generate any new compiler warnings.

findbugs +1.  The patch does not introduce any new Findbugs warnings.

core tests +1.  The patch passed core unit tests.

contrib tests +1.  The patch passed contrib unit tests.

Test results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1219/testReport/
Findbugs warnings: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1219/artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Checkstyle results: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1219/artifact/trunk/build/test/checkstyle-errors.html
Console output: 
http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Patch/1219/console

This message is automatically generated.

> build does not fail when libhdfs build fails
> 
>
> Key: HADOOP-2313
> URL: https://issues.apache.org/jira/browse/HADOOP-2313
> Project: Hadoop
>  Issue Type: Bug
>  Components: libhdfs
>Affects Versions: 0.15.1
>Reporter: Nigel Daley
>Assignee: Nigel Daley
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: HADOOP-2313.patch
>
>
> compile-libhdfs:
> [mkdir] Created dir: 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs
>  [exec] gcc -g -Wall -O2 -fPIC -m32 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include/linux -c hdfs.c -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfs.o
>  [exec] gcc -g -Wall -O2 -fPIC -m32 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include/linux -c 
> hdfsJniHelper.c -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfsJniHelper.o
>  [exec] gcc 
> -L/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/jre/lib/amd64/server -ljvm 
> -shared -m32 -Wl,-x  -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1
>  -Wl,-soname,libhdfs.so 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfs.o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfsJniHelper.o
>  \
>  [exec] && ln -sf 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1
>  
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so
>  [exec] /usr/bin/ld: skipping incompatible 
> /home/hadoopqa/tools/java/jdk1.5.0_11-64bit/jre/lib/amd64/server/libjvm.so 
> when searching for -ljvm
>  [exec] /usr/bin/ld: cannot find -ljvm
>  [exec] collect2: ld returned 1 exit status
>  [exec] make: *** 
> [/home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1]
>  Error 1
>  [exec] Result: 2
> ...
> BUILD SUCCESSFUL

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-2158) hdfsListDirectory in libhdfs does not scale

2007-11-30 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-2158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547224
 ] 

dhruba borthakur commented on HADOOP-2158:
--

Good patch. Code looks good. + 1.

II think the listPaths API on a directory used to return the size of the entire 
directory subtree. However, the listStatus API on a directory does not do so. 
If your application is not replying the original behaviour of listPaths then 
this change makes sense.

> hdfsListDirectory in libhdfs does not scale
> ---
>
> Key: HADOOP-2158
> URL: https://issues.apache.org/jira/browse/HADOOP-2158
> Project: Hadoop
>  Issue Type: Bug
>  Components: libhdfs
>Affects Versions: 0.15.0
>Reporter: Christian Kunz
>Assignee: Christian Kunz
>Priority: Blocker
> Fix For: 0.15.2
>
> Attachments: 2158.patch
>
>
> hdfsListDirectory makes one rpc call using deprecated 
> fs.FileSystem.listPaths, and then two rpc calls for every entry in the 
> returned array. When running a job with more than 3000 mappers each running a 
> pipes application using libhdfs to scan a dfs directory with about 100-200 
> entries, this results in about 1M rpc calls to the namenode server 
> overwhelming it.
> hdfsListDirectory should call fs.FileSystem.listStatus instead.
> I will submit a patch.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HADOOP-1841) IPC server should write repsonses asynchronously

2007-11-30 Thread Owen O'Malley (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-1841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12547222
 ] 

Owen O'Malley commented on HADOOP-1841:
---

I think on RPC patches, it is a very good idea to run a sort on 500 nodes to 
make sure that the system remains stable.

> IPC server should write repsonses asynchronously
> 
>
> Key: HADOOP-1841
> URL: https://issues.apache.org/jira/browse/HADOOP-1841
> Project: Hadoop
>  Issue Type: Improvement
>  Components: ipc
>Reporter: Doug Cutting
>Assignee: dhruba borthakur
> Attachments: asyncRPC-2.patch, asyncRPC.patch, asyncRPC.patch
>
>
> Hadoop's IPC Server currently writes responses from request handler threads 
> using blocking writes.  Performance and scalability might be improved if 
> responses were written asynchronously.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2313) build does not fail when libhdfs build fails

2007-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2313?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated HADOOP-2313:
--

Affects Version/s: 0.15.1

+1

> build does not fail when libhdfs build fails
> 
>
> Key: HADOOP-2313
> URL: https://issues.apache.org/jira/browse/HADOOP-2313
> Project: Hadoop
>  Issue Type: Bug
>  Components: libhdfs
>Affects Versions: 0.15.1
>Reporter: Nigel Daley
>Assignee: Nigel Daley
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: HADOOP-2313.patch
>
>
> compile-libhdfs:
> [mkdir] Created dir: 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs
>  [exec] gcc -g -Wall -O2 -fPIC -m32 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include/linux -c hdfs.c -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfs.o
>  [exec] gcc -g -Wall -O2 -fPIC -m32 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include 
> -I/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/include/linux -c 
> hdfsJniHelper.c -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfsJniHelper.o
>  [exec] gcc 
> -L/home/hadoopqa/tools/java/jdk1.5.0_11-64bit/jre/lib/amd64/server -ljvm 
> -shared -m32 -Wl,-x  -o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1
>  -Wl,-soname,libhdfs.so 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfs.o 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/hdfsJniHelper.o
>  \
>  [exec] && ln -sf 
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1
>  
> /home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so
>  [exec] /usr/bin/ld: skipping incompatible 
> /home/hadoopqa/tools/java/jdk1.5.0_11-64bit/jre/lib/amd64/server/libjvm.so 
> when searching for -ljvm
>  [exec] /usr/bin/ld: cannot find -ljvm
>  [exec] collect2: ld returned 1 exit status
>  [exec] make: *** 
> [/home/hadoopqa/workspace/Hadoop-LinuxTest-0.15/branch/build/libhdfs/libhdfs.so.1]
>  Error 1
>  [exec] Result: 2
> ...
> BUILD SUCCESSFUL

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2224) [hbase] Add HTable.getRow(ROW, ts)

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2224:
--

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed (Failure was in the unrelated, currently-erratic TestTableMapReduce). 
 Resolving.  Thanks for the patch Bryan.

> [hbase] Add HTable.getRow(ROW, ts)
> --
>
> Key: HADOOP-2224
> URL: https://issues.apache.org/jira/browse/HADOOP-2224
> Project: Hadoop
>  Issue Type: Improvement
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: stack
>Assignee: stack
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: get-full-timestamp.patch, get-full-ts-11-26-07.patch, 
> get-full-ts-11-29-07.patch, getfullts-v3.patch, testget2.patch
>
>
> There is HTable.getRow(ROW).  Need to add HTable.getRow(ROW, ts)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2304) [Hbase shell] Abbreviated symbol parsing error of dir path in jar command

2007-11-30 Thread stack (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

stack updated HADOOP-2304:
--

Resolution: Fixed
Status: Resolved  (was: Patch Available)

Committed.  Closing.  Thanks for the patch Edward.

> [Hbase shell] Abbreviated symbol parsing error of dir path in jar command
> -
>
> Key: HADOOP-2304
> URL: https://issues.apache.org/jira/browse/HADOOP-2304
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Affects Versions: 0.15.0
>Reporter: Edward Yoon
>Assignee: Edward Yoon
> Fix For: 0.16.0
>
> Attachments: 2304.patch
>
>
> I found a abbreviated symbol parsing error of dir path.
> Of course, arguments will have same error.
> ./build/hadoop-*0.16.0*-examples.jar pi 10 10   (FLOATING_POINT_LITERAL, 
> DOT...)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HADOOP-2309) [hbase] ConcurrentModificationException doing get of all region start keys

2007-11-30 Thread Jim Kellerman (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-2309?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jim Kellerman updated HADOOP-2309:
--

Status: Open  (was: Patch Available)

TestTableMapReduce failed in build 1217 - investigating

> [hbase] ConcurrentModificationException doing get of all region start keys
> --
>
> Key: HADOOP-2309
> URL: https://issues.apache.org/jira/browse/HADOOP-2309
> Project: Hadoop
>  Issue Type: Bug
>  Components: contrib/hbase
>Affects Versions: 0.16.0
>Reporter: stack
>Assignee: Jim Kellerman
>Priority: Minor
> Fix For: 0.16.0
>
> Attachments: patch.txt, patch.txt
>
>
> Reported by Paul Saab:
> {code}
> I'm getting the following exception when getting the start key from every 
> table.  If I try to grab it again, it will succeed.  Is this a bug or do you 
> need to catch these sort of exceptions in my code?...
> Looking for key: img529/6013/46da7e3f111daje2.jpg
> Exception in thread "main" java.io.IOException: java.io.IOException: 
> java.util.ConcurrentModificationException
> at java.util.TreeMap$NavigableSubMap$SubMapIterator.nextEntry 
> (TreeMap.java:1572)
> at 
> java.util.TreeMap$NavigableSubMap$SubMapEntryIterator.next(TreeMap.java:1620)
> at 
> java.util.TreeMap$NavigableSubMap$SubMapEntryIterator.next(TreeMap.java:1614)
> at org.apache.hadoop.hbase.HStore$Memcache.internalGet (HStore.java:223)
> at org.apache.hadoop.hbase.HStore$Memcache.get(HStore.java:152)
> at org.apache.hadoop.hbase.HStore.get(HStore.java:1490)
> at org.apache.hadoop.hbase.HRegion.get(HRegion.java:1002)
> at org.apache.hadoop.hbase.HRegion.get(HRegion.java:959)
> at org.apache.hadoop.hbase.HRegionServer.get(HRegionServer.java:1233)
> at sun.reflect.GeneratedMethodAccessor20.invoke(Unknown Source)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke 
> (DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java :596)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> at sun.reflect.DelegatingConstructorAccessorImpl.newInstance 
> (DelegatingConstructorAccessorImpl.java:27)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
> at 
> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java
>  :82)
> at org.apache.hadoop.hbase.HTable.get(HTable.java:202)
> at us.imageshack.hbase.ListRegions.run(ListRegions.java:25)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> at us.imageshack.hbase.ListRegions.main (ListRegions.java:38)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> at sun.reflect.DelegatingMethodAccessorImpl.invoke 
> (DelegatingMethodAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:597)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:155)
> {code}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



  1   2   >