Re: New Hadoop Eclipse functionality

2013-06-11 Thread Henry Junyoung Kim
Can anybody download the plugin from the update-site?

the body from the url is empty.
(people.apache.org/~srimanth/hadoop-eclipse/update-site/)

html
head
titleSrimanth Gunturi/title
/head

body
/body
/html


2013. 6. 11., 오후 4:27, Srimanth Gunturi srima...@gmail.com 작성:

 Hello,
 I wanted to invite all users and developers to try out the new Hadoop 
 functionality in Eclipse. Initial features include a HDFS Explorer (built on 
 top of Eclipse File System), and a ZooKeeper explorer. The project can be 
 accessed at http://people.apache.org/~srimanth/hadoop-eclipse/.
 
 Please open new feature requests, issues, comments and questions at 
 https://github.com/srimanth-gunturi/hadoop-eclipse/issues.
 
 This project is created under the Apache 2.0 License. Developers are welcome 
 to contribute and extend on this project. The project has been created with 
 extensibility and support for multiple Hadoop versions in mind.
 
 Requesting the community to use and strengthen this project, to provide 
 better tools in the Hadoop environment.
 Best regards,
 Srimanth



Re: are we able to decommission multi nodes at one time?

2013-04-03 Thread Henry Junyoung Kim
thanks for all. 

my strategies for removing from 15 DN to 8 DN.

1. kill two DNs at same time. : NN will detect nodes' down and he will try to 
retain replication factors of lost blocks.
2. check your NN web UI. there is an info to let you know counts of 
under-replicated blocks.
3. if it is returned to 0 again, you can do remove other 2 DNs again. let's 
start from first again.

this scenario is what I did. my cluster is working well. ;)

thanks for all.

2013. 4. 3., 오후 5:18, Azuryy Yu azury...@gmail.com 작성:

 not at all.  so don't worry about that.
 
 
 On Wed, Apr 3, 2013 at 2:04 PM, Yanbo Liang yanboha...@gmail.com wrote:
 It means that may be some replicas will be stay in under replica state?
 
 
 2013/4/3 Azuryy Yu azury...@gmail.com
 bq. then namenode start to copy block replicates on DN-2 to another DN, 
 supposed DN-2. 
 
 sorry for typo.
 
 Correct for it:
 then namenode start to copy block replicates on DN-1 to another DN, supposed 
 DN-2.
 
 
 On Wed, Apr 3, 2013 at 9:51 AM, Azuryy Yu azury...@gmail.com wrote:
 It's different.
 If you just want to stop DN-1 a short time, just kill the DataNode process on 
 DN-1. then do what you want. during this time, Namenode  cannot receive the 
 heart beat from DN-1, then namenode start to copy block replicates on DN-2 to 
 another DN, supposed DN-2. 
 
 But when you start DN-1 again, Namenode receive the DN-1 registration, then 
 namenode stop to copy the DN-1's block replicates even if NN doesn't finish 
 coping.
 
 Am I explain clearly?
 
 
 
 On Wed, Apr 3, 2013 at 9:43 AM, Henry Junyoung Kim henry.jy...@gmail.com 
 wrote:
 @Harsh
 
 What's the reasons to make big gaps for removing nodes between decommission 
 and just down nodes?
 In my understanding, both are necessary to copy un-replicated blocks to 
 another alive nodes.
 If main costs of  them are this one, total elapsed time couldn't be big 
 different.
 
 Could you share some articles or documents to understand about 
 decommissioning procedures?
 - explaining is always thanks ;)
 
 
 2013. 4. 2., 오후 5:37, Harsh J ha...@cloudera.com 작성:
 
  Yes, you can do the downtime work in steps of 2 DNs at a time,
  especially since you mentioned the total work would be only ~30mins at
  most.
 
  On Tue, Apr 2, 2013 at 1:46 PM, Henry Junyoung Kim
  henry.jy...@gmail.com wrote:
  the rest of nodes to be alive has enough size to store.
 
  for this one that you've mentioned.
  its easier to do so in a rolling manner without need of a
  decommission.
 
  to check my understanding, just shutting down 2 of them and then 2 more 
  and then 2 more without decommissions.
 
  is this correct?
 
 
  2013. 4. 2., 오후 4:54, Harsh J ha...@cloudera.com 작성:
 
  Note though that its only possible to decommission 7 nodes at the same
  time and expect it to finish iff the remaining 8 nodes have adequate
  free space for the excess replicas.
 
  If you're just going to take them down for a short while (few mins
  each), its easier to do so in a rolling manner without need of a
  decommission. You can take upto two down at a time on a replication
  average of 3 or 3+, and put it back in later without too much data
  movement impact.
 
  On Tue, Apr 2, 2013 at 1:06 PM, Yanbo Liang yanboha...@gmail.com wrote:
  It's reasonable to decommission 7 nodes at the same time.
  But may be it also takes long time to finish it.
  Because all the replicas in these 7 nodes need to be copied to remaining 
  8
  nodes.
  The size of transfer from these nodes to the remaining nodes is equal.
 
 
  2013/4/2 Henry Junyoung Kim henry.jy...@gmail.com
 
  :)
 
  currently, I  have 15 data nodes.
  for some tests, I am trying to decommission until 8 nodes.
 
  Now, the total dfs used size is 52 TB which is including all replicated
  blocks.
  from 15 to 8, total spent time is almost 4 days long. ;(
 
  someone mentioned that I don't need to decommission node by node.
  for this case, is there no problems if I decommissioned 7 nodes at the
  same time?
 
 
  2013. 4. 2., 오후 12:14, Azuryy Yu azury...@gmail.com 작성:
 
  I can translate it to native English: how many nodes you want to
  decommission?
 
 
  On Tue, Apr 2, 2013 at 11:01 AM, Yanbo Liang yanboha...@gmail.com 
  wrote:
 
  You want to decommission how many nodes?
 
 
  2013/4/2 Henry JunYoung KIM henry.jy...@gmail.com
 
  15 for datanodes and 3 for replication factor.
 
  2013. 4. 1., 오후 3:23, varun kumar varun@gmail.com 작성:
 
  How many nodes do you have and replication factor for it.
 
 
 
 
 
 
 
 
  --
  Harsh J
 
 
 
 
  --
  Harsh J
 
 
 
 
 



Re: are we able to decommission multi nodes at one time?

2013-04-02 Thread Henry Junyoung Kim
the rest of nodes to be alive has enough size to store.

for this one that you've mentioned.
 its easier to do so in a rolling manner without need of a
 decommission.

to check my understanding, just shutting down 2 of them and then 2 more and 
then 2 more without decommissions.

is this correct?


2013. 4. 2., 오후 4:54, Harsh J ha...@cloudera.com 작성:

 Note though that its only possible to decommission 7 nodes at the same
 time and expect it to finish iff the remaining 8 nodes have adequate
 free space for the excess replicas.
 
 If you're just going to take them down for a short while (few mins
 each), its easier to do so in a rolling manner without need of a
 decommission. You can take upto two down at a time on a replication
 average of 3 or 3+, and put it back in later without too much data
 movement impact.
 
 On Tue, Apr 2, 2013 at 1:06 PM, Yanbo Liang yanboha...@gmail.com wrote:
 It's reasonable to decommission 7 nodes at the same time.
 But may be it also takes long time to finish it.
 Because all the replicas in these 7 nodes need to be copied to remaining 8
 nodes.
 The size of transfer from these nodes to the remaining nodes is equal.
 
 
 2013/4/2 Henry Junyoung Kim henry.jy...@gmail.com
 
 :)
 
 currently, I  have 15 data nodes.
 for some tests, I am trying to decommission until 8 nodes.
 
 Now, the total dfs used size is 52 TB which is including all replicated
 blocks.
 from 15 to 8, total spent time is almost 4 days long. ;(
 
 someone mentioned that I don't need to decommission node by node.
 for this case, is there no problems if I decommissioned 7 nodes at the
 same time?
 
 
 2013. 4. 2., 오후 12:14, Azuryy Yu azury...@gmail.com 작성:
 
 I can translate it to native English: how many nodes you want to
 decommission?
 
 
 On Tue, Apr 2, 2013 at 11:01 AM, Yanbo Liang yanboha...@gmail.com wrote:
 
 You want to decommission how many nodes?
 
 
 2013/4/2 Henry JunYoung KIM henry.jy...@gmail.com
 
 15 for datanodes and 3 for replication factor.
 
 2013. 4. 1., 오후 3:23, varun kumar varun@gmail.com 작성:
 
 How many nodes do you have and replication factor for it.
 
 
 
 
 
 
 
 
 -- 
 Harsh J



Re: are we able to decommission multi nodes at one time?

2013-04-02 Thread Henry Junyoung Kim
@Harsh

What's the reasons to make big gaps for removing nodes between decommission and 
just down nodes?
In my understanding, both are necessary to copy un-replicated blocks to another 
alive nodes.
If main costs of  them are this one, total elapsed time couldn't be big 
different.

Could you share some articles or documents to understand about decommissioning 
procedures?
- explaining is always thanks ;)


2013. 4. 2., 오후 5:37, Harsh J ha...@cloudera.com 작성:

 Yes, you can do the downtime work in steps of 2 DNs at a time,
 especially since you mentioned the total work would be only ~30mins at
 most.
 
 On Tue, Apr 2, 2013 at 1:46 PM, Henry Junyoung Kim
 henry.jy...@gmail.com wrote:
 the rest of nodes to be alive has enough size to store.
 
 for this one that you've mentioned.
 its easier to do so in a rolling manner without need of a
 decommission.
 
 to check my understanding, just shutting down 2 of them and then 2 more and 
 then 2 more without decommissions.
 
 is this correct?
 
 
 2013. 4. 2., 오후 4:54, Harsh J ha...@cloudera.com 작성:
 
 Note though that its only possible to decommission 7 nodes at the same
 time and expect it to finish iff the remaining 8 nodes have adequate
 free space for the excess replicas.
 
 If you're just going to take them down for a short while (few mins
 each), its easier to do so in a rolling manner without need of a
 decommission. You can take upto two down at a time on a replication
 average of 3 or 3+, and put it back in later without too much data
 movement impact.
 
 On Tue, Apr 2, 2013 at 1:06 PM, Yanbo Liang yanboha...@gmail.com wrote:
 It's reasonable to decommission 7 nodes at the same time.
 But may be it also takes long time to finish it.
 Because all the replicas in these 7 nodes need to be copied to remaining 8
 nodes.
 The size of transfer from these nodes to the remaining nodes is equal.
 
 
 2013/4/2 Henry Junyoung Kim henry.jy...@gmail.com
 
 :)
 
 currently, I  have 15 data nodes.
 for some tests, I am trying to decommission until 8 nodes.
 
 Now, the total dfs used size is 52 TB which is including all replicated
 blocks.
 from 15 to 8, total spent time is almost 4 days long. ;(
 
 someone mentioned that I don't need to decommission node by node.
 for this case, is there no problems if I decommissioned 7 nodes at the
 same time?
 
 
 2013. 4. 2., 오후 12:14, Azuryy Yu azury...@gmail.com 작성:
 
 I can translate it to native English: how many nodes you want to
 decommission?
 
 
 On Tue, Apr 2, 2013 at 11:01 AM, Yanbo Liang yanboha...@gmail.com wrote:
 
 You want to decommission how many nodes?
 
 
 2013/4/2 Henry JunYoung KIM henry.jy...@gmail.com
 
 15 for datanodes and 3 for replication factor.
 
 2013. 4. 1., 오후 3:23, varun kumar varun@gmail.com 작성:
 
 How many nodes do you have and replication factor for it.
 
 
 
 
 
 
 
 
 --
 Harsh J
 
 
 
 
 -- 
 Harsh J



Re: are we able to decommission multi nodes at one time?

2013-04-01 Thread Henry JunYoung KIM
15 for datanodes and 3 for replication factor.

2013. 4. 1., 오후 3:23, varun kumar varun@gmail.com 작성:

 How many nodes do you have and replication factor for it.



Re: are we able to decommission multi nodes at one time?

2013-04-01 Thread Henry Junyoung Kim
:) 

currently, I  have 15 data nodes. 
for some tests, I am trying to decommission until 8 nodes. 

Now, the total dfs used size is 52 TB which is including all replicated blocks.
from 15 to 8, total spent time is almost 4 days long. ;(

someone mentioned that I don't need to decommission node by node. 
for this case, is there no problems if I decommissioned 7 nodes at the same 
time?


2013. 4. 2., 오후 12:14, Azuryy Yu azury...@gmail.com 작성:

 I can translate it to native English: how many nodes you want to decommission?
 
 
 On Tue, Apr 2, 2013 at 11:01 AM, Yanbo Liang yanboha...@gmail.com wrote:
 You want to decommission how many nodes?
 
 
 2013/4/2 Henry JunYoung KIM henry.jy...@gmail.com
 15 for datanodes and 3 for replication factor.
 
 2013. 4. 1., 오후 3:23, varun kumar varun@gmail.com 작성:
 
  How many nodes do you have and replication factor for it.
 
 
 



are we able to decommission multi nodes at one time?

2013-03-31 Thread Henry JunYoung KIM
hi, hadoop users.

to decommission a single node, there is necessary to wait to remove a node from 
a cluster for awhile. (in my case, 20 ~ 30 minutes)
for the safety, I am decommissioning a node at a time. 

for the performance, am I able to remove multi nodes at same time?



problems under same hosts and different ip addresses

2013-02-19 Thread Henry JunYoung KIM
hi, hadoopers.

Recently, we've moved our clusters to another idc center.
We keep the same host-names, but, they have now different ip addresses.

Without any configuration changes, we got the following error after starting 
cluster.

13.110.239.218 -- old ip 
13.271.6.54 -- new ip 

2013-02-20 10:26:10,536 FATAL org.apache.hadoop.hdfs.server.datanode.DataNode: 
Initialization failed for block pool Block pool 
BP-846907911-13.110.239.218-1359529186091 (storage id 
DS-2127506481-13.110.239.155-50010-1359529245747) service to 
search-ddm-test2.daum.net/13.271.5.233:8020
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException):
 Datanode denied communication with namenode: DatanodeRegistration(13.271.6.54, 
storageID=DS-2127506481-13.110.239.155-50010-1359529245747, infoPort=50075, 
ipcPort=50020, 
storageInfo=lv=-40;cid=CID-c497f9b4-77e1-4b04-acfe-31aceea9b0b1;nsid=582785493;c=0)
at 
org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager.registerDatanode(DatanodeManager.java:566)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:3358)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.registerDatanode(NameNodeRpcServer.java:854)
at 
org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolServerSideTranslatorPB.registerDatanode(DatanodeProtocolServerSideTranslatorPB.java:91)


any suggestions to resolve this problem?
thanks for your concerns.