[jira] Created: (HDFS-1249) with fuse-dfs, chown which only has owner (or only group) argument fails with Input/output error.

2010-06-18 Thread matsusaka kentaro (JIRA)
with fuse-dfs, chown which only has owner (or only group) argument fails with 
Input/output error.
-

 Key: HDFS-1249
 URL: https://issues.apache.org/jira/browse/HDFS-1249
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: contrib/fuse-dfs
Affects Versions: 0.20.2, 0.20.1
 Environment: x86 linux (ubuntu 10.04)
Reporter: matsusaka kentaro
Priority: Minor


with fuse-dfs, chown which only has owner (or only group) argument fails with 
Input/output error.
--
/mnt/hdfs/tmp# chown root file1
chown: changing ownership of `file1': Input/output error
/mnt/hdfs/tmp# chown root:root file1
/mnt/hdfs/tmp# chown :root file1
chown: changing group of `file1': Input/output error
--
I think it should be treated as unchanged for missing part(owner or group) 
instead of returning an error.

I took fuse_dfs log and it is saying
--
unique: 25, opcode: SETATTR (4), nodeid: 14, insize: 128
chown /tmp/file1 0 4294967295
could not lookup group -1
   unique: 25, error: -5 (Input/output error), outsize: 16
unique: 26, opcode: SETATTR (4), nodeid: 14, insize: 128
chown /tmp/file1 0 0
getattr /tmp/file1
   unique: 26, success, outsize: 120
unique: 27, opcode: SETATTR (4), nodeid: 14, insize: 128
chown /tmp/file1 4294967295 0
could not lookup userid -1
   unique: 27, error: -5 (Input/output error), outsize: 16
--

therefore this should happen because dfs_chown() in 
src/contrib/fuse-dfs/src/fuse_impls_chown.c has following
--
...
  user = getUsername(uid);
  if (NULL == user) {
syslog(LOG_ERR,"Could not lookup the user id string %d\n",(int)uid); 
fprintf(stderr, "could not lookup userid %d\n", (int)uid); 
ret = -EIO;
  }

  if (0 == ret) {
group = getGroup(gid);
if (group == NULL) {
  syslog(LOG_ERR,"Could not lookup the group id string %d\n",(int)gid); 
  fprintf(stderr, "could not lookup group %d\n", (int)gid); 
  ret = -EIO;
} 
  }
...
--

but actually, hdfsChown() in src/c++/libhdfs/hdfs.c has this
--
...
if (owner == NULL && group == NULL) {
  fprintf(stderr, "Both owner and group cannot be null in chown");
  errno = EINVAL;
  return -1;
}
...
--

and also, setOwner seems allowing NULL
--
username - If it is null, the original username remains unchanged.
groupname - If it is null, the original groupname remains unchanged.
--
according to the api document.

therefore, I think fuse_impls_chown.c should not treat only user(or only group) 
lookup fail as an error.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1249) with fuse-dfs, chown which only has owner (or only group) argument fails with Input/output error.

2010-06-18 Thread matsusaka kentaro (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

matsusaka kentaro updated HDFS-1249:


Attachment: HDFS1249.1

I think it should behave like this attached diff.

> with fuse-dfs, chown which only has owner (or only group) argument fails with 
> Input/output error.
> -
>
> Key: HDFS-1249
> URL: https://issues.apache.org/jira/browse/HDFS-1249
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: contrib/fuse-dfs
>Affects Versions: 0.20.1, 0.20.2
> Environment: x86 linux (ubuntu 10.04)
>Reporter: matsusaka kentaro
>Priority: Minor
> Attachments: HDFS1249.1
>
>
> with fuse-dfs, chown which only has owner (or only group) argument fails with 
> Input/output error.
> --
> /mnt/hdfs/tmp# chown root file1
> chown: changing ownership of `file1': Input/output error
> /mnt/hdfs/tmp# chown root:root file1
> /mnt/hdfs/tmp# chown :root file1
> chown: changing group of `file1': Input/output error
> --
> I think it should be treated as unchanged for missing part(owner or group) 
> instead of returning an error.
> I took fuse_dfs log and it is saying
> --
> unique: 25, opcode: SETATTR (4), nodeid: 14, insize: 128
> chown /tmp/file1 0 4294967295
> could not lookup group -1
>unique: 25, error: -5 (Input/output error), outsize: 16
> unique: 26, opcode: SETATTR (4), nodeid: 14, insize: 128
> chown /tmp/file1 0 0
> getattr /tmp/file1
>unique: 26, success, outsize: 120
> unique: 27, opcode: SETATTR (4), nodeid: 14, insize: 128
> chown /tmp/file1 4294967295 0
> could not lookup userid -1
>unique: 27, error: -5 (Input/output error), outsize: 16
> --
> therefore this should happen because dfs_chown() in 
> src/contrib/fuse-dfs/src/fuse_impls_chown.c has following
> --
> ...
>   user = getUsername(uid);
>   if (NULL == user) {
> syslog(LOG_ERR,"Could not lookup the user id string %d\n",(int)uid); 
> fprintf(stderr, "could not lookup userid %d\n", (int)uid); 
> ret = -EIO;
>   }
>   if (0 == ret) {
> group = getGroup(gid);
> if (group == NULL) {
>   syslog(LOG_ERR,"Could not lookup the group id string %d\n",(int)gid); 
>   fprintf(stderr, "could not lookup group %d\n", (int)gid); 
>   ret = -EIO;
> } 
>   }
> ...
> --
> but actually, hdfsChown() in src/c++/libhdfs/hdfs.c has this
> --
> ...
> if (owner == NULL && group == NULL) {
>   fprintf(stderr, "Both owner and group cannot be null in chown");
>   errno = EINVAL;
>   return -1;
> }
> ...
> --
> and also, setOwner seems allowing NULL
> --
> username - If it is null, the original username remains unchanged.
> groupname - If it is null, the original groupname remains unchanged.
> --
> according to the api document.
> therefore, I think fuse_impls_chown.c should not treat only user(or only 
> group) lookup fail as an error.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-752:
-

Attachment: HDFS-752.patch

This patch adds interface classification to HDFS classes. Contrib and tests are 
excluded from this. Approach:
- All protocol related classes, Implementations of FileSystem and 
AbstractFileSystem are tagged private and evolving
- All exceptions exposed in protocol classes are marked private and evolving. 
These classes should perhaps move to protocol package.
- All the other public classes are marked private

Tagged DatanodeDescriptor.BlockTargetPair as private and evolving, since it is 
used in BlockCommand. This should be moved to protocol directory, to ensure all 
the classes used in protocols is in protocol packages. Will open a bug to track 
this.

Tom, can you please review this.


> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 0.22.0
>
> Attachments: HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1057) Concurrent readers hit ChecksumExceptions if following a writer to very end of file

2010-06-18 Thread sam rash (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880278#action_12880278
 ] 

sam rash commented on HDFS-1057:


So removing setBytesOnDisk() means:


{code}
  if (replicaInfo instanceof ReplicaBeingWritten) {
((ReplicaBeingWritten) replicaInfo)
  .setLastChecksumAndDataLen(offsetInBlock, lastChunkChecksum);
  }
  
  replicaInfo.setBytesOnDisk(offsetInBlock);
{code}

will not have the latter.  So all other implementations of Replica will have a 
valid value for getByteOnDIsk()?
Does this also mean that the impl of getBytesOnDisk for ReplicaInPipeline will 
move to ReplicaBeingWritten ?

> Concurrent readers hit ChecksumExceptions if following a writer to very end 
> of file
> ---
>
> Key: HDFS-1057
> URL: https://issues.apache.org/jira/browse/HDFS-1057
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node
>Affects Versions: 0.20-append, 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Assignee: sam rash
>Priority: Blocker
> Attachments: conurrent-reader-patch-1.txt, 
> conurrent-reader-patch-2.txt, conurrent-reader-patch-3.txt, 
> hdfs-1057-trunk-1.txt, hdfs-1057-trunk-2.txt, hdfs-1057-trunk-3.txt
>
>
> In BlockReceiver.receivePacket, it calls replicaInfo.setBytesOnDisk before 
> calling flush(). Therefore, if there is a concurrent reader, it's possible to 
> race here - the reader will see the new length while those bytes are still in 
> the buffers of BlockReceiver. Thus the client will potentially see checksum 
> errors or EOFs. Additionally, the last checksum chunk of the file is made 
> accessible to readers even though it is not stable.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1071) savenamespace should write the fsimage to all configured fs.name.dir in parallel

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov updated HDFS-1071:


Status: Patch Available  (was: Open)

> savenamespace should write the fsimage to all configured fs.name.dir in 
> parallel
> 
>
> Key: HDFS-1071
> URL: https://issues.apache.org/jira/browse/HDFS-1071
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: dhruba borthakur
>Assignee: Dmytro Molkov
> Attachments: HDFS-1071.2.patch, HDFS-1071.3.patch, HDFS-1071.4.patch, 
> HDFS-1071.5.patch, HDFS-1071.patch
>
>
> If you have a large number of files in HDFS, the fsimage file is very big. 
> When the namenode restarts, it writes a copy of the fsimage to all 
> directories configured in fs.name.dir. This takes a long time, especially if 
> there are many directories in fs.name.dir. Make the NN write the fsimage to 
> all these directories in parallel.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1071) savenamespace should write the fsimage to all configured fs.name.dir in parallel

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov updated HDFS-1071:


Attachment: HDFS-1071.5.patch

Attaching new patch that applies to trunk since there are some conflicts with 
the older version.

> savenamespace should write the fsimage to all configured fs.name.dir in 
> parallel
> 
>
> Key: HDFS-1071
> URL: https://issues.apache.org/jira/browse/HDFS-1071
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: dhruba borthakur
>Assignee: Dmytro Molkov
> Attachments: HDFS-1071.2.patch, HDFS-1071.3.patch, HDFS-1071.4.patch, 
> HDFS-1071.5.patch, HDFS-1071.patch
>
>
> If you have a large number of files in HDFS, the fsimage file is very big. 
> When the namenode restarts, it writes a copy of the fsimage to all 
> directories configured in fs.name.dir. This takes a long time, especially if 
> there are many directories in fs.name.dir. Make the NN write the fsimage to 
> all these directories in parallel.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1057) Concurrent readers hit ChecksumExceptions if following a writer to very end of file

2010-06-18 Thread sam rash (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880291#action_12880291
 ] 

sam rash commented on HDFS-1057:


another way to ask this:  only ReplicaBeingWritten needs to have the bytes on 
disk set in BlockRecevier?

> Concurrent readers hit ChecksumExceptions if following a writer to very end 
> of file
> ---
>
> Key: HDFS-1057
> URL: https://issues.apache.org/jira/browse/HDFS-1057
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node
>Affects Versions: 0.20-append, 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Assignee: sam rash
>Priority: Blocker
> Attachments: conurrent-reader-patch-1.txt, 
> conurrent-reader-patch-2.txt, conurrent-reader-patch-3.txt, 
> hdfs-1057-trunk-1.txt, hdfs-1057-trunk-2.txt, hdfs-1057-trunk-3.txt
>
>
> In BlockReceiver.receivePacket, it calls replicaInfo.setBytesOnDisk before 
> calling flush(). Therefore, if there is a concurrent reader, it's possible to 
> race here - the reader will see the new length while those bytes are still in 
> the buffers of BlockReceiver. Thus the client will potentially see checksum 
> errors or EOFs. Additionally, the last checksum chunk of the file is made 
> accessible to readers even though it is not stable.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1000) libhdfs needs to be updated to use the new UGI

2010-06-18 Thread Devaraj Das (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Devaraj Das updated HDFS-1000:
--

   Status: Resolved  (was: Patch Available)
Fix Version/s: (was: 0.22.0)
   Resolution: Fixed

I just committed this.

> libhdfs needs to be updated to use the new UGI
> --
>
> Key: HDFS-1000
> URL: https://issues.apache.org/jira/browse/HDFS-1000
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Devaraj Das
>Assignee: Devaraj Das
>Priority: Blocker
> Fix For: 0.21.0
>
> Attachments: fs-javadoc.patch, hdfs-1000-bp20.3.patch, 
> hdfs-1000-bp20.4.patch, hdfs-1000-bp20.patch, hdfs-1000-trunk.1.patch
>
>
> libhdfs needs to be updated w.r.t the APIs in the new UGI.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880319#action_12880319
 ] 

Tom White commented on HDFS-752:


+1 This looks fine to me.
 
A few public classes seem to have been missed, including HdfsConstants, 
Upgradeable, FSDatasetInterface, Replica, FSDatasetMBean, FSClusterStats, 
FSInodeInfo, FSNamesystemMBean, DatanodeProtocol, InterDatanodeProtocol, 
NamenodeProtocol, NamenodeProtocols, NodeRegistration, GSet, LightWeightGSet. 
Do you want to add annotations to these? Thanks.

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 0.22.0
>
> Attachments: HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1140) Speedup INode.getPathComponents

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov updated HDFS-1140:


Issue Type: Improvement  (was: Bug)
  Priority: Minor  (was: Major)

> Speedup INode.getPathComponents
> ---
>
> Key: HDFS-1140
> URL: https://issues.apache.org/jira/browse/HDFS-1140
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
>Priority: Minor
> Attachments: HDFS-1140.2.patch, HDFS-1140.patch
>
>
> When the namenode is loading the image there is a significant amount of time 
> being spent in the DFSUtil.string2Bytes. We have a very specific workload 
> here. The path that namenode does getPathComponents for shares N - 1 
> component with the previous path this method was called for (assuming current 
> path has N components).
> Hence we can improve the image load time by caching the result of previous 
> conversion.
> We thought of using some simple LRU cache for components, but the reality is, 
> String.getBytes gets optimized during runtime and LRU cache doesn't perform 
> as well, however using just the latest path components and their translation 
> to bytes in two arrays gives quite a performance boost.
> I could get another 20% off of the time to load the image on our cluster (30 
> seconds vs 24) and I wrote a simple benchmark that tests performance with and 
> without caching.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HDFS-1245) Plugable block id generation

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1245?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov reassigned HDFS-1245:
---

Assignee: Dmytro Molkov

> Plugable block id generation 
> -
>
> Key: HDFS-1245
> URL: https://issues.apache.org/jira/browse/HDFS-1245
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
>
> The idea is to have a way to easily create block id generation engines that 
> may fit a certain purpose. One of them could be HDFS-898 started by 
> Konstantin, but potentially others.
> We chatted with Dhruba about this for a while and came up with the following 
> approach:
> There should be a BlockIDGenerator interface that has following methods:
> void blockAdded(Block)
> void blockRemoved(Block)
> Block nextBlock()
> First two methods are needed for block generation engines that hold a certain 
> state. During the restart, when namenode reads the fsimage it will notify 
> generator about all the blocks it reads from the image and during runtime 
> namenode will notify the generator about block removals on file deletion.
> The instance of the generator will also have a reference to the block 
> registry, the interface that BlockManager implements. The only method there 
> is __blockExists(Block)__, so that the current random block id generation can 
> be implemented, since it needs to check with the block manager if the id is 
> already present.
> What does the community think about this proposal?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1057) Concurrent readers hit ChecksumExceptions if following a writer to very end of file

2010-06-18 Thread Hairong Kuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880321#action_12880321
 ] 

Hairong Kuang commented on HDFS-1057:
-

I meant removing setBytesOnDisk and moving setLastChecksumAndDataLen to 
ReplicaPipelineInterface.

> Concurrent readers hit ChecksumExceptions if following a writer to very end 
> of file
> ---
>
> Key: HDFS-1057
> URL: https://issues.apache.org/jira/browse/HDFS-1057
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node
>Affects Versions: 0.20-append, 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Assignee: sam rash
>Priority: Blocker
> Attachments: conurrent-reader-patch-1.txt, 
> conurrent-reader-patch-2.txt, conurrent-reader-patch-3.txt, 
> hdfs-1057-trunk-1.txt, hdfs-1057-trunk-2.txt, hdfs-1057-trunk-3.txt
>
>
> In BlockReceiver.receivePacket, it calls replicaInfo.setBytesOnDisk before 
> calling flush(). Therefore, if there is a concurrent reader, it's possible to 
> race here - the reader will see the new length while those bytes are still in 
> the buffers of BlockReceiver. Thus the client will potentially see checksum 
> errors or EOFs. Additionally, the last checksum chunk of the file is made 
> accessible to readers even though it is not stable.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-752:
-

Attachment: HDFS-752.1.patch

Good catch... I had missed tagging public interfaces. Attached patch fixes all 
the interfaces. One addition to your list StorageDirType.

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880336#action_12880336
 ] 

Tom White commented on HDFS-752:


+1

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-752:
-

Status: Patch Available  (was: Open)

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.21.0, 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HDFS-1250) Namenode accepts block report from dead datanodes

2010-06-18 Thread Suresh Srinivas (JIRA)
Namenode accepts block report from dead datanodes
-

 Key: HDFS-1250
 URL: https://issues.apache.org/jira/browse/HDFS-1250
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: name-node
Affects Versions: 0.20.2, 0.22.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


When a datanode heartbeat times out namenode marks it dead. The subsequent 
heartbeat from the datanode is rejected with a command to datanode to 
re-register. However namenode accepts block report from the datanode although 
it is marked dead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1250) Namenode accepts block report from dead datanodes

2010-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880342#action_12880342
 ] 

Suresh Srinivas commented on HDFS-1250:
---

This creates several problems:
# Blocks from dead datanode is added to block map. Opening a file returns 
blocks from the dead datanode. The client experience failure to read it 
repeatedly.
# Decommissioning the datanode does not resolve this problem as the blocks 
continue to remain in the block map. Only way to solve this problem is by 
restarting namenode.

This problem was observed on a yahoo cluster, where a datanode was marked dead, 
however a block report from it made it to namenode.

> Namenode accepts block report from dead datanodes
> -
>
> Key: HDFS-1250
> URL: https://issues.apache.org/jira/browse/HDFS-1250
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.20.2, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>
> When a datanode heartbeat times out namenode marks it dead. The subsequent 
> heartbeat from the datanode is rejected with a command to datanode to 
> re-register. However namenode accepts block report from the datanode although 
> it is marked dead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1114) Reducing NameNode memory usage by an alternate hash table

2010-06-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1114?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-1114:
-

Attachment: benchmark20100618.patch

benchmark20100618.patch: additional codes for benchmark and memory footprint 
measurement.

> Reducing NameNode memory usage by an alternate hash table
> -
>
> Key: HDFS-1114
> URL: https://issues.apache.org/jira/browse/HDFS-1114
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.22.0
>
> Attachments: benchmark20100618.patch, GSet20100525.pdf, 
> gset20100608.pdf, h1114_20100607.patch, h1114_20100614b.patch, 
> h1114_20100615.patch, h1114_20100616b.patch, h1114_20100617.patch, 
> h1114_20100617b.patch
>
>
> NameNode uses a java.util.HashMap to store BlockInfo objects.  When there are 
> many blocks in HDFS, this map uses a lot of memory in the NameNode.  We may 
> optimize the memory usage by a light weight hash table implementation.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1057) Concurrent readers hit ChecksumExceptions if following a writer to very end of file

2010-06-18 Thread sam rash (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880346#action_12880346
 ] 

sam rash commented on HDFS-1057:


got it.  i will make the changes and get a patch soon

> Concurrent readers hit ChecksumExceptions if following a writer to very end 
> of file
> ---
>
> Key: HDFS-1057
> URL: https://issues.apache.org/jira/browse/HDFS-1057
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: data-node
>Affects Versions: 0.20-append, 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Assignee: sam rash
>Priority: Blocker
> Attachments: conurrent-reader-patch-1.txt, 
> conurrent-reader-patch-2.txt, conurrent-reader-patch-3.txt, 
> hdfs-1057-trunk-1.txt, hdfs-1057-trunk-2.txt, hdfs-1057-trunk-3.txt
>
>
> In BlockReceiver.receivePacket, it calls replicaInfo.setBytesOnDisk before 
> calling flush(). Therefore, if there is a concurrent reader, it's possible to 
> race here - the reader will see the new length while those bytes are still in 
> the buffers of BlockReceiver. Thus the client will potentially see checksum 
> errors or EOFs. Additionally, the last checksum chunk of the file is made 
> accessible to readers even though it is not stable.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1192) refreshSuperUserGroupsConfiguration should use server side configuration for the refresh (for HADOOP-6815)

2010-06-18 Thread Boris Shkolnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1192?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Boris Shkolnik updated HDFS-1192:
-

Attachment: HDFS-1192-9.patch

> refreshSuperUserGroupsConfiguration should use server side configuration for 
> the refresh (for HADOOP-6815)
> --
>
> Key: HDFS-1192
> URL: https://issues.apache.org/jira/browse/HDFS-1192
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Boris Shkolnik
>Assignee: Boris Shkolnik
> Attachments: HDFS-1192-1.patch, HDFS-1192-9.patch
>
>
> Currently refreshSuperUserGroupsConfiguration is using client side 
> Configuration.
> One of the issues with this is that if the cluster is restarted it will loose 
> the "refreshed' values (unless they are copied to the NameNode/JobTracker 
> machine).

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-599) Improve Namenode robustness by prioritizing datanode heartbeats over client requests

2010-06-18 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-599:
---

Status: Resolved  (was: Patch Available)
Resolution: Fixed

I just committed this. Thanks Dmytro!

> Improve Namenode robustness by prioritizing datanode heartbeats over client 
> requests
> 
>
> Key: HDFS-599
> URL: https://issues.apache.org/jira/browse/HDFS-599
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: dhruba borthakur
>Assignee: Dmytro Molkov
> Fix For: 0.22.0
>
> Attachments: HDFS-599.3.patch, HDFS-599.patch
>
>
> The namenode processes RPC requests from clients that are reading/writing to 
> files as well as heartbeats/block reports from datanodes.
> Sometime, because of various reasons (Java GC runs, inconsistent performance 
> of NFS filer that stores HDFS transacttion logs, etc), the namenode 
> encounters transient slowness. For example, if the device that stores the 
> HDFS transaction logs becomes sluggish, the Namenode's ability to process 
> RPCs slows down to a certain extent. During this time, the RPCs from clients 
> as well as the RPCs from datanodes suffer in similar fashion. If the 
> underlying problem becomes worse, the NN's ability to process a heartbeat 
> from a DN is severly impacted, thus causing the NN to declare that the DN is 
> dead. Then the NN starts replicating blocks that used to reside on the 
> now-declared-dead datanode. This adds extra load to the NN. Then the 
> now-declared-datanode finally re-establishes contact with the NN, and sends a 
> block report. The block report processing on the NN is another heavyweight 
> activity, thus casing more load to the already overloaded namenode. 
> My proposal is tha the NN should try its best to continue processing RPCs 
> from datanodes and give lesser priority to serving client requests. The 
> Datanode RPCs are integral to the consistency and performance of the Hadoop 
> file system, and it is better to protect it at all costs. This will ensure 
> that NN  recovers from the hiccup much faster than what it does now.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1110) Namenode heap optimization - reuse objects for commonly used file names

2010-06-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880355#action_12880355
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-1110:
--

Hi Suresh, new commit should be appended at end of the list in CHANGE.txt but 
not inserted in the beginning.

{noformat}
--- hadoop/hdfs/trunk/CHANGES.txt   2010/06/11 18:09:48 953802
+++ hadoop/hdfs/trunk/CHANGES.txt   2010/06/11 18:15:17 953803
@@ -9,6 +9,9 @@
 
   IMPROVEMENTS
 
+HDFS-1110. Reuses objects for commonly used file names in namenode to
+reduce the heap usage. (suresh)
+
 HDFS-1096. fix for prev. commit. (boryas)
{noformat}


> Namenode heap optimization - reuse objects for commonly used file names
> ---
>
> Key: HDFS-1110
> URL: https://issues.apache.org/jira/browse/HDFS-1110
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: hdfs-1110.2.patch, hdfs-1110.3.patch, hdfs-1110.4.patch, 
> hdfs-1110.5.patch, hdfs-1110.6.patch, hdfs-1110.patch, hdfs-1110.y20.patch
>
>
> There are a lot of common file names used in HDFS, mainly created by 
> mapreduce, such as file names starting with "part". Reusing byte[] 
> corresponding to these recurring file names will save significant heap space 
> used for storing the file names in millions of INodeFile objects.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1096) allow dfsadmin/mradmin refresh of superuser proxy group mappings

2010-06-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1096?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880356#action_12880356
 ] 

Tsz Wo (Nicholas), SZE commented on HDFS-1096:
--

Hi Boris, 
- new commit should be appended at end of the list in CHANGE.txt but not 
inserted in the beginning.
- you don't now have to add new item in CHANGE.txt for fixing previous commit.
{noformat}
HDFS-1096. fix for prev. commit. (boryas)

HDFS-1096. allow dfsadmin/mradmin refresh of superuser proxy group
 mappings (boryas)
{noformat}


> allow dfsadmin/mradmin refresh of superuser proxy group mappings
> 
>
> Key: HDFS-1096
> URL: https://issues.apache.org/jira/browse/HDFS-1096
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Reporter: Boris Shkolnik
>Assignee: Boris Shkolnik
> Attachments: HDFS-1096-3.patch, HDFS-1096-BP20-4.patch, 
> HDFS-1096-BP20-6-fix.patch, HDFS-1096-BP20-6.patch, HDFS-1096-BP20-7.patch, 
> HDFS-1096-fix.patch
>
>


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1110) Namenode heap optimization - reuse objects for commonly used file names

2010-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880358#action_12880358
 ] 

Suresh Srinivas commented on HDFS-1110:
---

Sure. Next time I commit some thing I will fix it. 

> Namenode heap optimization - reuse objects for commonly used file names
> ---
>
> Key: HDFS-1110
> URL: https://issues.apache.org/jira/browse/HDFS-1110
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: hdfs-1110.2.patch, hdfs-1110.3.patch, hdfs-1110.4.patch, 
> hdfs-1110.5.patch, hdfs-1110.6.patch, hdfs-1110.patch, hdfs-1110.y20.patch
>
>
> There are a lot of common file names used in HDFS, mainly created by 
> mapreduce, such as file names starting with "part". Reusing byte[] 
> corresponding to these recurring file names will save significant heap space 
> used for storing the file names in millions of INodeFile objects.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1119) Refactor BlocksMap with GettableSet

2010-06-18 Thread Tsz Wo (Nicholas), SZE (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1119?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsz Wo (Nicholas), SZE updated HDFS-1119:
-

Attachment: h1119_20100525_y0.20.1xx.patch

h1119_20100525_y0.20.1xx.patch: for y0.20.1xx

> Refactor BlocksMap with GettableSet
> ---
>
> Key: HDFS-1119
> URL: https://issues.apache.org/jira/browse/HDFS-1119
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.22.0
>
> Attachments: h1119_20100429.patch, h1119_20100506.patch, 
> h1119_20100521.patch, h1119_20100525.patch, h1119_20100525_y0.20.1xx.patch
>
>
> The data structure required in BlocksMap is a GettableSet.  See also [this 
> comment|https://issues.apache.org/jira/browse/HDFS-1114?focusedCommentId=12862118&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#action_12862118].

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1250) Namenode accepts block report from dead datanodes

2010-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1250?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880366#action_12880366
 ] 

Suresh Srinivas commented on HDFS-1250:
---

Hudson is stuck. I ran testpatch; here is the result:
 [exec] -1 overall.  
 [exec] 
 [exec] +1 @author.  The patch does not contain any @author tags.
 [exec] 
 [exec] -1 tests included.  The patch doesn't appear to include any new 
or modified tests.
 [exec] Please justify why no new tests are needed 
for this patch.
 [exec] Also please list what manual steps were 
performed to verify this patch.
 [exec] 
 [exec] +1 javadoc.  The javadoc tool did not generate any warning 
messages.
 [exec] 
 [exec] +1 javac.  The applied patch does not increase the total number 
of javac compiler warnings.
 [exec] 
 [exec] +1 findbugs.  The patch does not introduce any new Findbugs 
warnings.
 [exec] 
 [exec] +1 release audit.  The applied patch does not increase the 
total number of release audit warnings.


This change just tags the code with interface classification. Hence tests are 
not included. I also ran unit tests and it ran without failures.

> Namenode accepts block report from dead datanodes
> -
>
> Key: HDFS-1250
> URL: https://issues.apache.org/jira/browse/HDFS-1250
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Affects Versions: 0.20.2, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>
> When a datanode heartbeat times out namenode marks it dead. The subsequent 
> heartbeat from the datanode is rejected with a command to datanode to 
> re-register. However namenode accepts block report from the datanode although 
> it is marked dead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-752:
-

   Status: Resolved  (was: Patch Available)
 Hadoop Flags: [Reviewed]
Fix Version/s: (was: 0.21.0)
   Resolution: Fixed

Committed the patch to trunk.

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1140) Speedup INode.getPathComponents

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov updated HDFS-1140:


Attachment: HDFS-1140.3.patch

I removed one array copy from the routine. This should speed it up even more.
I will submit this for tests right away.

> Speedup INode.getPathComponents
> ---
>
> Key: HDFS-1140
> URL: https://issues.apache.org/jira/browse/HDFS-1140
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
>Priority: Minor
> Attachments: HDFS-1140.2.patch, HDFS-1140.3.patch, HDFS-1140.patch
>
>
> When the namenode is loading the image there is a significant amount of time 
> being spent in the DFSUtil.string2Bytes. We have a very specific workload 
> here. The path that namenode does getPathComponents for shares N - 1 
> component with the previous path this method was called for (assuming current 
> path has N components).
> Hence we can improve the image load time by caching the result of previous 
> conversion.
> We thought of using some simple LRU cache for components, but the reality is, 
> String.getBytes gets optimized during runtime and LRU cache doesn't perform 
> as well, however using just the latest path components and their translation 
> to bytes in two arrays gives quite a performance boost.
> I could get another 20% off of the time to load the image on our cluster (30 
> seconds vs 24) and I wrote a simple benchmark that tests performance with and 
> without caching.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1140) Speedup INode.getPathComponents

2010-06-18 Thread Dmytro Molkov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmytro Molkov updated HDFS-1140:


Status: Patch Available  (was: Open)

> Speedup INode.getPathComponents
> ---
>
> Key: HDFS-1140
> URL: https://issues.apache.org/jira/browse/HDFS-1140
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
>Priority: Minor
> Attachments: HDFS-1140.2.patch, HDFS-1140.3.patch, HDFS-1140.patch
>
>
> When the namenode is loading the image there is a significant amount of time 
> being spent in the DFSUtil.string2Bytes. We have a very specific workload 
> here. The path that namenode does getPathComponents for shares N - 1 
> component with the previous path this method was called for (assuming current 
> path has N components).
> Hence we can improve the image load time by caching the result of previous 
> conversion.
> We thought of using some simple LRU cache for components, but the reality is, 
> String.getBytes gets optimized during runtime and LRU cache doesn't perform 
> as well, however using just the latest path components and their translation 
> to bytes in two arrays gives quite a performance boost.
> I could get another 20% off of the time to load the image on our cluster (30 
> seconds vs 24) and I wrote a simple benchmark that tests performance with and 
> without caching.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1245) Plugable block id generation

2010-06-18 Thread dhruba borthakur (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1245?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880379#action_12880379
 ] 

dhruba borthakur commented on HDFS-1245:


Maybe you can write up just the proposed Interface class (with empty methods) 
along with an imbedded javadoc docs? That might help in understanding this 
proposal.

> Plugable block id generation 
> -
>
> Key: HDFS-1245
> URL: https://issues.apache.org/jira/browse/HDFS-1245
> Project: Hadoop HDFS
>  Issue Type: New Feature
>  Components: name-node
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
>
> The idea is to have a way to easily create block id generation engines that 
> may fit a certain purpose. One of them could be HDFS-898 started by 
> Konstantin, but potentially others.
> We chatted with Dhruba about this for a while and came up with the following 
> approach:
> There should be a BlockIDGenerator interface that has following methods:
> void blockAdded(Block)
> void blockRemoved(Block)
> Block nextBlock()
> First two methods are needed for block generation engines that hold a certain 
> state. During the restart, when namenode reads the fsimage it will notify 
> generator about all the blocks it reads from the image and during runtime 
> namenode will notify the generator about block removals on file deletion.
> The instance of the generator will also have a reference to the block 
> registry, the interface that BlockManager implements. The only method there 
> is __blockExists(Block)__, so that the current random block id generation can 
> be implemented, since it needs to check with the block manager if the id is 
> already present.
> What does the community think about this proposal?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HDFS-752:
-

Attachment: HDFS-752.rel21.patch

Patch for branch 0.21

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, HDFS-752.rel21.patch, 
> hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880385#action_12880385
 ] 

Suresh Srinivas commented on HDFS-752:
--

Tom, should this be committed 0.21 branch?

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, HDFS-752.rel21.patch, 
> hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880412#action_12880412
 ] 

Tom White commented on HDFS-752:


Yes, please.

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, HDFS-752.rel21.patch, 
> hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1114) Reducing NameNode memory usage by an alternate hash table

2010-06-18 Thread Scott Carey (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880423#action_12880423
 ] 

Scott Carey commented on HDFS-1114:
---

bq. This is a good point. Is there a way to determine if UseCompressedOops is 
set in runtime?

Well, there is ManagementFactory.getRuntimeMXBean().getInputArguments(), but 
later versions of Java are going to be making +UseCompressedOops the default.

There is also a way to check if the VM is 64 bit or 32 bit, either its out of 
ManagementFactory or one of the system properties.  Digging around I don't see 
it, but I have used it before.  I think it is vendor specific though.





> Reducing NameNode memory usage by an alternate hash table
> -
>
> Key: HDFS-1114
> URL: https://issues.apache.org/jira/browse/HDFS-1114
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: Tsz Wo (Nicholas), SZE
>Assignee: Tsz Wo (Nicholas), SZE
> Fix For: 0.22.0
>
> Attachments: benchmark20100618.patch, GSet20100525.pdf, 
> gset20100608.pdf, h1114_20100607.patch, h1114_20100614b.patch, 
> h1114_20100615.patch, h1114_20100616b.patch, h1114_20100617.patch, 
> h1114_20100617b.patch
>
>
> NameNode uses a java.util.HashMap to store BlockInfo objects.  When there are 
> many blocks in HDFS, this map uses a lot of memory in the NameNode.  We may 
> optimize the memory usage by a light weight hash table implementation.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Created: (HDFS-1251) Unnecessary log messages are printed in info level

2010-06-18 Thread Tsz Wo (Nicholas), SZE (JIRA)
Unnecessary log messages are printed in info level
--

 Key: HDFS-1251
 URL: https://issues.apache.org/jira/browse/HDFS-1251
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs client, security
Reporter: Tsz Wo (Nicholas), SZE
Priority: Minor


When I run a job on a secure cluster, the following messages are printed on the 
console.
{noformat}
10/06/19 05:05:18 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 
125033 for tsz
10/06/19 05:05:18 INFO security.TokenCache: Got dt for
 
hdfs://namenode/user/tsz/.staging/job_201006091641_23763;uri=xx.xx.xx.xx:8020;t.service=xx.xx.xx.xx:8020
{noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-752) Add interface classification stable & scope to HDFS

2010-06-18 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12880430#action_12880430
 ] 

Suresh Srinivas commented on HDFS-752:
--

Committed the patch to 0.21 branch.

> Add interface classification stable & scope to HDFS
> ---
>
> Key: HDFS-752
> URL: https://issues.apache.org/jira/browse/HDFS-752
> Project: Hadoop HDFS
>  Issue Type: New Feature
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
> Fix For: 0.22.0
>
> Attachments: HDFS-752.1.patch, HDFS-752.patch, HDFS-752.rel21.patch, 
> hdfs.interface.txt
>
>
> This jira addresses adding interface classification for the classes in hadoop 
> hdfs, based on the mechanism described in Hadoop-5073.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-947) The namenode should redirect a hftp request to read a file to the datanode that has the maximum number of local replicas

2010-06-18 Thread dhruba borthakur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-947?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dhruba borthakur updated HDFS-947:
--

   Status: Resolved  (was: Patch Available)
 Hadoop Flags: [Reviewed]
Fix Version/s: 0.22.0
   Resolution: Fixed

I just committed this. Thanks Dmytro!

> The namenode should redirect a hftp request to read a file to the datanode 
> that has the maximum number of local replicas
> 
>
> Key: HDFS-947
> URL: https://issues.apache.org/jira/browse/HDFS-947
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: dhruba borthakur
>Assignee: Dmytro Molkov
> Fix For: 0.22.0
>
> Attachments: HDFS-947.2.patch, HDFS-947.patch, hftpRedirection.patch
>
>
> A client that uses the Hftp protocol to read a file is redirected by the 
> namenode to a random datanode. It would be nice if the client gets redirected 
> to a datanode that has the maximum number of local replicas of the blocks of 
> the file.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.