Re: Python Hadoop Example

2019-06-16 Thread Artem Ervits
https://www.michael-noll.com/tutorials/writing-an-hadoop-mapreduce-program-in-python/

On Sun, Jun 16, 2019, 9:18 AM Mike IT Expert  wrote:

> Please let me know where I can find a good/simple example of mapreduce
> Python code running on Hadoop. Like tutorial or sth.
>
> Thank you
>
>
>


Re: hadoop mapreduce job rest api

2015-12-23 Thread Artem Ervits
Take a look at webhcat api
On Dec 24, 2015 12:50 AM, "ram kumar"  wrote:

> Hi,
>
> I want to submit a mapreduce job using rest api,
> and get the status of the job every n interval.
> Is there a way to do it?
>
> Thanks
>


Re: Start/stop scripts - particularly start-dfs.sh - in Hortonworks Data Platform 2.3.X

2015-10-24 Thread Artem Ervits
Look in /usr/hdp/2.3
On Oct 23, 2015 10:07 AM, "Stephen Boesch"  wrote:

>
> We are setting up automated deployments on a headless system: so using the
> GUI is not an option here.  When we search for those scripts under HDP
> they are not found:
>
> $ pwd
> /usr/hdp/current
>
> Which scripts exist in HDP ?
>
> [stack@s1-639016 current]$ find -L . -name \*.sh
> ...
>
> There are ZERO start/stop sh scripts..
>
> In particular I am interested in the *start-dfs.sh* script that starts
> the namenode(s) , journalnode, and datanodes.
>
>


Re: Active Namenode keeps crashing

2015-08-12 Thread Artem Ervits
Check whether connectivity between servers is stable. Error says it can't
reach one node. Also check that time is synched between nodes.
On Aug 9, 2015 9:31 PM, "Suresh V"  wrote:

> In our HA setup, the active namenode keeps crashing once a week or so. The
> cluster is quite idle without many jobs running and not much user activity.
>
> Below is logs from journal nodes. Can someone help us with this please?
>
>
> 2015-08-04 13:00:20,054 INFO  server.Journal
> (Journal.java:updateLastPromisedEpoch(315)) - Updating lastPromisedEpoch
> from 9 to 10 for client /172.26.44.133
>
> 2015-08-04 13:00:20,175 INFO  server.Journal
> (Journal.java:scanStorageForLatestEdits(188)) - Scanning storage
> FileJournalManager(root=/hadoop/hdfs/journal/HDPPROD)
>
> 2015-08-04 13:00:20,220 INFO  server.Journal
> (Journal.java:scanStorageForLatestEdits(194)) - Latest log is
> EditLogFile(file=/hadoop/hdfs/journal/HDPPROD/current/edits_inprogress_0523903,first=0523903,last=0523925,inProgress=true,hasCorruptHeader=false)
>
> 2015-08-04 13:00:20,891 INFO  server.Journal
> (Journal.java:getSegmentInfo(687)) - getSegmentInfo(523903):
> EditLogFile(file=/hadoop/hdfs/journal/HDPPROD/current/edits_inprogress_0523903,first=0523903,last=0523925,inProgress=true,hasCorruptHeader=false)
> -> startTxId: 523903 endTxId: 523925 isInProgress: true
>
> 2015-08-04 13:00:20,891 INFO  server.Journal
> (Journal.java:prepareRecovery(731)) - Prepared recovery for segment 523903:
> segmentState { startTxId: 523903 endTxId: 523925 isInProgress: true }
> lastWriterEpoch: 9 lastCommittedTxId: 523924
>
> 2015-08-04 13:00:20,956 INFO  server.Journal
> (Journal.java:getSegmentInfo(687)) - getSegmentInfo(523903):
> EditLogFile(file=/hadoop/hdfs/journal/HDPPROD/current/edits_inprogress_0523903,first=0523903,last=0523925,inProgress=true,hasCorruptHeader=false)
> -> startTxId: 523903 endTxId: 523925 isInProgress: true
>
> 2015-08-04 13:00:20,956 INFO  server.Journal
> (Journal.java:acceptRecovery(817)) - Skipping download of log startTxId:
> 523903 endTxId: 523925 isInProgress: true: already have up-to-date logs
>
> 2015-08-04 13:00:20,989 INFO  server.Journal
> (Journal.java:acceptRecovery(850)) - Accepted recovery for segment 523903:
> segmentState { startTxId: 523903 endTxId: 523925 isInProgress: true }
> acceptedInEpoch: 10
>
> 2015-08-04 13:00:21,791 INFO  server.Journal
> (Journal.java:finalizeLogSegment(584)) - Validating log segment
> /hadoop/hdfs/journal/HDPPROD/current/edits_inprogress_0523903
> about to be finalized
>
> 2015-08-04 13:00:21,805 INFO  namenode.FileJournalManager
> (FileJournalManager.java:finalizeLogSegment(133)) - Finalizing edits file
> /hadoop/hdfs/journal/HDPPROD/current/edits_inprogress_0523903
> ->
> /hadoop/hdfs/journal/HDPPROD/current/edits_0523903-0523925
>
> 2015-08-04 13:00:22,257 INFO  server.Journal
> (Journal.java:startLogSegment(532)) - Updating lastWriterEpoch from 9 to 10
> for client /172.26.44.133
>
> 2015-08-04 13:00:23,699 INFO  ipc.Server (Server.java:run(2060)) - IPC
> Server handler 4 on 8485, call
> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocol.journal from
> 172.26.44.135:43678 Call#304302 Retry#0
>
> java.io.IOException: IPC's epoch 9 is less than the last promised epoch 10
>
> at
> org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:414)
>
> at
> org.apache.hadoop.hdfs.qjournal.server.Journal.checkWriteRequest(Journal.java:442)
>
> at
> org.apache.hadoop.hdfs.qjournal.server.Journal.journal(Journal.java:342)
>
> at
> org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.journal(JournalNodeRpcServer.java:148)
>
> at
> org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.journal(QJournalProtocolServerSideTranslatorPB.java:158)
>
> at
> org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25421)
>
> at
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
>
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
>
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2039)
>
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2035)
>
> at java.security.AccessController.doPrivileged(Native Method)
>
> at javax.security.auth.Subject.doAs(Subject.java:415)
>
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2033)
>
> 2015-08-06 19:13:14,012 INFO  httpclient.HttpMethodDirector
> (HttpMethodDirector.java:executeWithRetry(439)) - I/O exception
> (org.apache.commons.httpclient.NoHttpResponseException) caught when
> processing request: The server a

Re: Backup of individual component of Hadoop ecosystem

2015-02-26 Thread Artem Ervits
There are several approaches. I would check hdfs trash folder of the user
deleting a file. Expiration of items in trash is controlled by
fs.trash.interval property on core-site.xml.
Artem Ervits
On Feb 26, 2015 1:31 PM, "Krish Donald"  wrote:

> Hi,
>
> As per my understanding we don't take backup of Hadoop cluster as the size
> is very large generally .
>
> However in case if somebody has dropped a table by mistake then how should
> we recover the data ?
>
> How to take backup of Hadoop ecosystem individual component.
>
> Thanks
> Krish
>


Re: Running MapReduce jobs in batch mode on different data sets

2015-02-21 Thread Artem Ervits
Take a look at Apache Oozie

Artem Ervits
On Feb 21, 2015 6:35 AM, "tesm...@gmail.com"  wrote:

> Hi,
>
> Is it possible to run jobs on Hadoop in batch mode?
>
> I have 5 different datasets in HDFS and need to run the same MapReduce
> application on these datasets sets one after the other.
>
> Right now I am doing it manually How can I automate this?
>
> How can I save the log of each execution in text files for later
> processing?
>
> Regards,
> Tariq
>


Re: Home for Apache Big Data Solutions?

2015-02-09 Thread Artem Ervits
I believe Apache Bigtop is what you're looking for.

Artem Ervits
On Feb 9, 2015 8:15 AM, "Jean-Baptiste Onofré"  wrote:

> Hi Amir,
>
> thanks for the update.
>
> Please, let me know if you need some help on the proposal and to "qualify"
> your ideas.
>
> Regards
> JB
>
> On 02/09/2015 02:05 PM, MrAsanjar . wrote:
>
>> Hi Chris,
>> thanks for the information, will get on it ...
>>
>> Hi JB
>> Glad that you are familiar with Juju, however my personal goal is not to
>> promote any tool but
>> to take the next step, which is to build a community for apache big data
>> solutions.
>>
>>  >>do you already have a kind of proposal/description of your projects ?
>> working on it :) I got the idea while flying back from South Africa on
>> Saturday. During my trip I noticed most of the communities spending
>> their precious resources on solution plumbing, without much of emphasis
>> on solution best practices due to the lack of expertise. By the time
>> Big Data solution framework becomes operational, funding has diminished
>> enough to limit solution activity (i.e data analytic payload
>> development). I am sure we could find
>> similar scenarios with  other institutions and SMB (small and
>> medium-size businesses) anywhere.
>> In the nutshell my goals are as follow:
>> 1) Make Big Data solutions available to everyone
>> 2) Encapsulate the best practices
>> 3) All Orchestration tools are welcomed - Some solutions could have
>> hybrid tooling model
>> 4) Enforce automated testing and quality control.
>> 5) Share analytic payloads (i.e mapreduce apps, storm topology, Pig
>> scripts,...)
>>
>>
>>  >>Is it like distribution, or tooling ?
>> Good question, I envision to have a distribution model as it has
>> dependency on Apache hadoop projects distributions.
>>
>>  >>What's the current license ?
>> Charms/Bundles are moving to Apache 2.0 license, target data 2/27.
>>
>> Regards
>> Amir Sanjar
>> Big Data Solution Lead
>> Canonical
>>
>> On Sun, Feb 8, 2015 at 10:46 AM, Mattmann, Chris A (3980)
>> mailto:chris.a.mattm...@jpl.nasa.gov>>
>> wrote:
>>
>> Dear Amir,
>>
>> Thank you for your interest in contributing these projects
>> to the ASF! Sincerely appreciate it.
>>
>> My suggestion would be to look into the Apache Incubator,
>> which is the home for incoming projects at the ASF. The
>> TL;DR answer is:
>>
>> 1. You’ll need to create a proposal for each project
>> that you would like to bring in using:
>> http://incubator.apache.org/guides/proposal.html
>>
>>
>> 2. You should put your proposal up on a public wiki
>> for each project:
>> http://wiki.apache.org/incubator/
>> create a new page e.g., YourProjectProposal, which would in
>> turn become http://wiki.apache.org/incubator/YouProjectProposal
>> You will need to request permissions to add the page on the
>> wiki
>>
>> 3. Recruit at least 3 IPMC/ASF members to mentor your project:
>> http://people.apache.org/committers-by-project.html#incubator-pmc
>>
>> http://people.apache.org/committers-by-project.html#member
>>
>>
>> 4. Submit your proposal for consideration at the Incubator
>> 5. Enjoy!
>>
>> Cheers and good luck.
>>
>> Cheers,
>> Chris
>>
>> ++
>> Chris Mattmann, Ph.D.
>> Chief Architect
>> Instrument Software and Science Data Systems Section (398)
>> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
>> Office: 168-519, Mailstop: 168-527
>> Email: chris.a.mattm...@nasa.gov <mailto:chris.a.mattm...@nasa.gov>
>> WWW: http://sunset.usc.edu/~mattmann/
>> ++
>> Adjunct Associate Professor, Computer Science Department
>> University of Southern California, Los Angeles, CA 90089 USA
>> ++
>>
>>
>>
>>
>>
>>
>> -Original Message-
>> From: "MrAsanjar ." mailto:afsan...@gmail.com>>
>> Reply-To: "user@hadoop.apache.org <mailto:user@hadoop.apache.org>"
>> mailto:user@hadoop.apache.org>>
>> Date: Sunday, February 8, 2015 at 8:36 AM
>> To: "

Re: Adding datanodes to Hadoop cluster - Will data redistribute?

2015-02-08 Thread Artem Ervits
Look at hdfs balancer

Artem Ervits
On Feb 6, 2015 5:54 PM, "Manoj Venkatesh"  wrote:

> Dear Hadoop experts,
>
> I have a Hadoop cluster of 8 nodes, 6 were added during cluster creation
> and 2 additional nodes were added later to increase disk and CPU capacity.
> What i see is that processing is shared amongst all the nodes whereas the
> storage is reaching capacity on the original 6 nodes whereas the newly
> added machines have relatively large amount of storage still unoccupied.
>
> I was wondering if there is an automated or any way of redistributing data
> so that all the nodes are equally utilized. I have checked for the
> configuration parameter - *dfs.datanode.fsdataset.volume.choosing.policy*
> have options 'Round Robin' or 'Available Space', are there any other
> configurations which need to be reviewed.
>
> Thanks,
> Manoj
>


Re: tools.DistCp: Invalid arguments

2015-02-03 Thread Artem Ervits
Another good option is hftp.

Artem Ervits
On Feb 3, 2015 6:42 AM, "xeonmailinglist"  wrote:

>  I have found the problem. I started to use `webhdfs` and everything is ok.
>
>
> On 03-02-2015 10:40, xeonmailinglist wrote:
>
>  What do you mean by no path is given? Even if I launch this command, I
> get the same error…. What path should I put here?
>
> $ hadoop distcp hdfs://hadoop-coc-1:50070/input1
> hdfs://hadoop-coc-2:50070/input1
>
> Thanks,
>
> On 02-02-2015 19:59, Alexander Alten-Lorenz wrote:
>
> Have a closer look:
>
>hdfs://hadoop-coc-2:50070/
>
>
>  No Path is given.
>
>
>  On 02 Feb 2015, at 20:52, xeonmailinglist 
> wrote:
>
>   Hi,
>
> I am trying to copy data using distcp but I get this error. Both hadoop
> runtime are working properly. Why is this happening?
>
>
> vagrant@hadoop-coc-1:~/Programs/hadoop$ hadoop distcp 
> hdfs://hadoop-coc-1:50070/input1 hdfs://hadoop-coc-2:50070/
> 15/02/02 19:46:37 ERROR tools.DistCp: Invalid arguments:
> java.io.IOException: Failed on local exception: 
> com.google.protobuf.InvalidProtocolBufferException: Protocol message 
> end-group tag did not match expected tag.; Host Details : local host is: 
> "hadoop-coc-1/127.0.1.1"; destination host is: "hadoop-coc-2":50070;
> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
> at org.apache.hadoop.ipc.Client.call(Client.java:1472)
> at org.apache.hadoop.ipc.Client.call(Client.java:1399)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
> at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
> at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
> at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
> at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
> at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
> at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
> at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:188)
> at org.apache.hadoop.tools.DistCp.run(DistCp.java:111)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> at org.apache.hadoop.tools.DistCp.main(DistCp.java:401)
> Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol 
> message end-group tag did not match expected tag.
> at 
> com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94)
> at 
> com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124)
> at 
> com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202)
> at 
> com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
> at 
> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
> at 
> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
> at 
> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
> at 
> org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:3167)
> at 
> org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1072)
> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:966)
> Invalid arguments: Failed on local exception: 
> com.google.protobuf.InvalidProtocolBufferException: Protocol message 
> end-group tag did not match expected tag.; Host Details : local host is: 
> "hadoop-coc-1/127.0.1.1"; destination host is: "hadoop-coc-2":50070;
> usage: distcp OPTIONS [source_path...] 
>
> Thanks,
> ​
>
>
>  ​
>
>
>


Re: Copy data between clusters during the job execution.

2015-02-02 Thread Artem Ervits
take a look at oozie, once first job completes you can distcp to another
server.

Artem Ervits
On Feb 2, 2015 5:46 AM, "Daniel Haviv"  wrote:

> It should run after your job finishes.
> You can create the flow using a simple bash script
>
> Daniel
>
> On 2 בפבר׳ 2015, at 12:31, xeonmailinglist 
> wrote:
>
> But can I use discp inside my job, or I need to program something that
> executes distcp after executing my job?
>
>
> On 02-02-2015 10:20, Daniel Haviv wrote:
>
>  an use distcp
>
> Daniel
>
>  On 2 בפבר׳ 2015, at 11:12,
>
>
>


Re: Any working VM of Apache Hadoop ?

2015-01-19 Thread Artem Ervits
+1 for Bigtop

Artem Ervits
On Jan 18, 2015 4:10 PM, "Jay Vyas"  wrote:

> Also BigTop has a very flexible vagrant infrastructure:
>
>
> https://github.com/apache/bigtop/tree/master/bigtop-deploy/vm/vagrant-puppet
>
> On Jan 18, 2015, at 3:37 PM, Andre Kelpe  wrote:
>
> Try our vagrant setup:
> https://github.com/Cascading/vagrant-cascading-hadoop-cluster
>
> - André
>
> On Sat, Jan 17, 2015 at 10:07 PM, Krish Donald 
> wrote:
>
>> Hi,
>>
>> I am looking for working VM of Apache Hadoop.
>> Not looking for cloudera or Horton works VMs.
>> If anybody has it and if they can share that would be great .
>>
>> Thanks
>> Krish
>>
>
>
>
> --
> André Kelpe
> an...@concurrentinc.com
> http://concurrentinc.com
>
>


Re: ExportSnapshot webhdfs problems

2014-10-06 Thread Artem Ervits
Does it work with a small table? I prefer to use hftp instead of webhdfs.

From: Brian Jeltema [mailto:brian.jelt...@digitalenvoy.net]
Sent: Friday, October 03, 2014 11:01 AM
To: user@hadoop.apache.org 
Subject: ExportSnapshot webhdfs problems

I posted this on users@hbase, but got not response, so I thought I’d try here:

I’m trying to use ExportSnapshot to copy a snapshot from a Hadoop 1 to a Hadoop 
2 cluster using the webhdfs protocol.
I’ve done this successfully before, though there are always mapper failures and 
retries in the job log. However, I’m not
having success with a rather large table due to an excessive number of 
failures. The exceptions in the job log are always:

14/09/29 20:28:11 INFO mapred.JobClient: Task Id : 
attempt_201409241055_0024_m_05_1, Status : FAILED
org.apache.hadoop.ipc.RemoteException
at org.apache.hadoop.hdfs.web.JsonUtil.toRemoteException(JsonUtil.java:114)
at 
org.apache.hadoop.hdfs.web.WebHdfsFileSystem.validateResponse(WebHdfsFileSystem.java:290)
at 
org.apache.hadoop.hdfs.web.WebHdfsFileSystem.access$500(WebHdfsFileSystem.java:98)
at 
org.apache.hadoop.hdfs.web.WebHdfsFileSystem$2.close(WebHdfsFileSystem.java:653)
at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:204)
at 
org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:146)
…
…
…


So I presume the real exception is taking place on the target system. However, 
examining the namenode logs and a handful of
the datanode logs has not revealed any exceptions that correlate with those in 
the job log. Is there some other log I should be looking at?
I reduced the number of mappers to 6 and the target cluster has 10 datanodes, 
so it’s hard to believe its a capacity problem.

Thanks
Brian

This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.


Re: Using Ambari to deploy Apache hadoop

2013-10-18 Thread Artem Ervits
For offline you need to set up local mirror, there's a guide in the manual. As 
far as bits, its an Apache release not Hortonworks, same as if you were to 
install from BigTop. 


Artem Ervits
Data Analyst
New York Presbyterian Hospital

- Original Message -
From: Siddharth Tiwari [mailto:siddharth.tiw...@live.com]
Sent: Friday, October 18, 2013 05:13 AM
To: user@hadoop.apache.org 
Subject: Using Ambari to deploy Apache hadoop

Hi team,

Is it possible to deploy hadoop from Apache via Ambari ? Also is there a link 
for full offline installation ? We do not have access to outside world and we 
want to use Ambari for reploying Hadoop ( not hortonworks release though )

Sent from my iPhone

This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.


Re: Hadoop Metrics Issue in ganglia.

2013-09-12 Thread Artem Ervits
Check firewall and /etc/hosts also make sure hosts lines up with result of 
hostname -f command. Both hostname -f and hosts entries should have fqdn names. 
I use ambari to install my cluster, including ganglia metrics and I had 
identical issue. Once I corrected that it started working. 


Artem Ervits
Data Analyst
New York Presbyterian Hospital

- Original Message -
From: Chris Embree [mailto:cemb...@gmail.com]
Sent: Wednesday, September 11, 2013 02:40 PM
To: user@hadoop.apache.org 
Subject: Re: Hadoop Metrics Issue in ganglia.

Did you try ganglia forums/lists?


On 9/11/13, orahad bigdata  wrote:
> Hi All,
>
> Can somebody help me please?
>
> Thanks
> On 9/11/13, orahad bigdata  wrote:
>> Hi All,
>>
>> I'm facing an issue while showing Hadoop metrics in ganglia, Though I
>> have installed ganglia on my master/slaves nodes and I'm able to see
>> all the default metrics on ganglia UI from all the nodes but I'm not
>> able to see Hadoop metrics in metrics section.
>>
>> versions:-
>> Hadoop 1.1.1
>> ganglia  3.0.7
>>
>> hadoop-metrics2.properties -
>>
>> dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
>> dfs.period=10
>> dfs.servers=master:8649
>> mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext31
>> mapred.period=10
>> mapred.servers=master:8649
>>
>> please guide.
>>
>> Thanks
>>
>

This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.


protect from accidental deletes

2013-04-01 Thread Artem Ervits
Hello all,

I'd like to know what users are doing to protect themselves from accidental 
deletes of files and directories in HDFS? Any suggestions are appreciated.

Thanks.



Hadoop 1.1.2 is it beta or stable?

2013-03-14 Thread Artem Ervits
Hello all,

I'd like to know whether version 1.1.2 is the new stable release. The download 
page stills says it's beta.

Thank you.

Artem Ervits
New York Presbyterian Hospital





This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: How to install Oozie 3.3.1 on Hadoop 1.1.1

2013-02-20 Thread Artem Ervits
It may not necessarily be correct but I built the project using 0.20.200 and am 
running it without problems "so far" on 1.0.3 hadoop.


Artem Ervits
Data Analyst
New York Presbyterian Hospital

From: Jagat Singh [mailto:jagatsi...@gmail.com]
Sent: Friday, February 15, 2013 09:42 PM
To: user@hadoop.apache.org 
Subject: Re: How to install Oozie 3.3.1 on Hadoop 1.1.1

Hi,

I can see that in pom.xml the supported hadoop version is

1.0.1

You can try to build your self with version you want to see if it works.

Also try to ask your question on oozie mailing list.

Regards,

Jagat Singh




On Sat, Feb 16, 2013 at 12:45 PM, Hemanth Yamijala 
mailto:yhema...@thoughtworks.com>> wrote:
Hi,

It may be useful to post this question on the oozie user mailing list. There 
are likely to be more expert users there. 
u...@oozie.apache.org<mailto:u...@oozie.apache.org>

Thanks
Hemanth

On Friday, February 15, 2013, anand verma wrote:
Hi,

I am struggling for many days to install Oozie 3.3.1 on Hadoop 1.1.1. Oozie 
documentation is very poorly written I am not able to figure it out. While 
installing I got an error saying it doesn't support Hadoop v1.1.1. Please help 
me out.

--
Regards
Ananda Prakash Verma






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Dell Hardware

2013-01-31 Thread Artem Ervits
Hello all,

Does anyone run Hadoop on Dell R720 model of servers? Dell site lists C2100 
model of servers as best fit for Hadoop workloads. What does community 
recommend?

Artem Ervits
New York Presbyterian Hospital





This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: JobCache directory cleanup

2013-01-10 Thread Artem Ervits
As soon as job completes, your jobcache should be cleared. Check your 
mapred-site.xml for mapred.local.dir setting and make sure job cleanup step is 
successful in web UI. Setting your job's intermediate output setting to true 
will keep the jobcache folder smaller.



Artem Ervits
Data Analyst
New York Presbyterian Hospital

From: Hemanth Yamijala [mailto:yhema...@thoughtworks.com]
Sent: Thursday, January 10, 2013 07:37 AM
To: user@hadoop.apache.org 
Subject: Re: JobCache directory cleanup

Hi,

On Thu, Jan 10, 2013 at 5:17 PM, Ivan Tretyakov 
mailto:itretya...@griddynamics.com>> wrote:
Thanks for replies!

Hemanth,
I could see following exception in TaskTracker log: 
https://issues.apache.org/jira/browse/MAPREDUCE-5
But I'm not sure if it is related to this issue.

> Now, when a job completes, the directories under the jobCache must get 
> automatically cleaned up. However it doesn't look like this is happening in 
> your case.

So, If I've no running jobs, jobcache directory should be empty. Is it correct?


That is correct. I just verified it with my Hadoop 1.0.2 version

Thanks
Hemanth



On Thu, Jan 10, 2013 at 8:18 AM, Hemanth Yamijala 
mailto:yhema...@thoughtworks.com>> wrote:
Hi,

The directory name you have provided is 
/data?/mapred/local/taskTracker/persona/jobcache/. This directory is used by 
the TaskTracker (slave) daemons to localize job files when the tasks are run on 
the slaves.

Hence, I don't think this is related to the parameter 
"mapreduce.jobtracker.retiredjobs.cache.size", which is a parameter related to 
the jobtracker process.

Now, when a job completes, the directories under the jobCache must get 
automatically cleaned up. However it doesn't look like this is happening in 
your case.

Could you please look at the logs of the tasktracker machine where this has 
gotten filled up to see if there are any errors that could give clues ?
Also, since this is a CDH release, it could be a problem specific to that - and 
maybe reaching out on the CDH mailing lists will also help

Thanks
hemanth

On Wed, Jan 9, 2013 at 8:11 PM, Ivan Tretyakov 
mailto:itretya...@griddynamics.com>> wrote:
Hello!

I've found that jobcache directory became very large on our cluster, e.g.:

# du -sh /data?/mapred/local/taskTracker/user/jobcache
465G/data1/mapred/local/taskTracker/user/jobcache
464G/data2/mapred/local/taskTracker/user/jobcache
454G/data3/mapred/local/taskTracker/user/jobcache

And it stores information for about 100 jobs:

# ls -1 /data?/mapred/local/taskTracker/persona/jobcache/  | sort | uniq | wc -l

I've found that there is following parameter:


  mapreduce.jobtracker.retiredjobs.cache.size
  1000
  The number of retired job status to keep in the cache.
  


So, if I got it right it intended to control job cache size by limiting number 
of jobs to store cache for.

Also, I've seen that some hadoop users uses cron approach to cleanup jobcache: 
http://grokbase.com/t/hadoop/common-user/102ax9bze1/cleaning-jobcache-manually 
(http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201002.mbox/%3c99484d561002100143s4404df98qead8f2cf687a7...@mail.gmail.com%3E)

Are there other approaches to control jobcache size?
What is more correct way to do it?

Thanks in advance!

P.S. We are using CDH 4.1.1.

--
Best Regards
Ivan Tretyakov

Deployment Engineer
Grid Dynamics
+7 812 640 38 76
Skype: ivan.tretyakov
www.griddynamics.com<http://www.griddynamics.com>
itretya...@griddynamics.com<mailto:itretya...@griddynamics.com>




--
Best Regards
Ivan Tretyakov

Deployment Engineer
Grid Dynamics
+7 812 640 38 76
Skype: ivan.tretyakov
www.griddynamics.com<http://www.griddynamics.com>
itretya...@griddynamics.com<mailto:itretya...@griddynamics.com>





This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: JobCache directory cleanup

2013-01-09 Thread Artem Ervits
Just ran into similar problem. If you compress intermediate data, it will kEep 
jobcache folder manageable.


Artem Ervits
Data Analyst
New York Presbyterian Hospital

From: Ivan Tretyakov [mailto:itretya...@griddynamics.com]
Sent: Wednesday, January 09, 2013 10:22 AM
To: user@hadoop.apache.org 
Subject: Re: JobCache directory cleanup

Thanks a lot Alexander!

What is mapreduce.jobtracker.retiredjobs.cache.size for?
Does cron approach safe for hadoop? Is that only way at the moment?


On Wed, Jan 9, 2013 at 6:50 PM, Alexander Alten-Lorenz 
mailto:wget.n...@gmail.com>> wrote:
Hi,

Per default (and not configurable) the logs will be persist for 30 days. This 
will be configurable in future 
(https://issues.apache.org/jira/browse/MAPREDUCE-4643).

- Alex

On Jan 9, 2013, at 3:41 PM, Ivan Tretyakov 
mailto:itretya...@griddynamics.com>> wrote:

> Hello!
>
> I've found that jobcache directory became very large on our cluster, e.g.:
>
> # du -sh /data?/mapred/local/taskTracker/user/jobcache
> 465G/data1/mapred/local/taskTracker/user/jobcache
> 464G/data2/mapred/local/taskTracker/user/jobcache
> 454G/data3/mapred/local/taskTracker/user/jobcache
>
> And it stores information for about 100 jobs:
>
> # ls -1 /data?/mapred/local/taskTracker/persona/jobcache/  | sort | uniq |
> wc -l
>
> I've found that there is following parameter:
>
> 
>  mapreduce.jobtracker.retiredjobs.cache.size
>  1000
>  The number of retired job status to keep in the cache.
>  
> 
>
> So, if I got it right it intended to control job cache size by limiting
> number of jobs to store cache for.
>
> Also, I've seen that some hadoop users uses cron approach to cleanup
> jobcache:
> http://grokbase.com/t/hadoop/common-user/102ax9bze1/cleaning-jobcache-manually
> (
> http://mail-archives.apache.org/mod_mbox/hadoop-common-user/201002.mbox/%3c99484d561002100143s4404df98qead8f2cf687a7...@mail.gmail.com%3E
> )
>
> Are there other approaches to control jobcache size?
> What is more correct way to do it?
>
> Thanks in advance!
>
> P.S. We are using CDH 4.1.1.
>
> --
> Best Regards
> Ivan Tretyakov
>
> Deployment Engineer
> Grid Dynamics
> +7 812 640 38 76
> Skype: ivan.tretyakov
> www.griddynamics.com<http://www.griddynamics.com>
> itretya...@griddynamics.com<mailto:itretya...@griddynamics.com>

--
Alexander Alten-Lorenz
http://mapredit.blogspot.com
German Hadoop LinkedIn Group: http://goo.gl/N8pCF




--
Best Regards
Ivan Tretyakov

Deployment Engineer
Grid Dynamics
+7 812 640 38 76
Skype: ivan.tretyakov
www.griddynamics.com<http://www.griddynamics.com>
itretya...@griddynamics.com<mailto:itretya...@griddynamics.com>




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Hadoop throughput question

2013-01-04 Thread Artem Ervits
John, the two programs below, one is from the Definitive Guide chapter 4 with 
slight mods and the other is in-house but similar to Hadoop in Action chap 3.

package sequencefileprocessor;

// cc SequenceFileReadDemo Reading a SequenceFile
import java.io.IOException;
import java.net.URI;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.io.SequenceFile;
import org.apache.hadoop.io.Writable;
import org.apache.hadoop.util.ReflectionUtils;

// vv SequenceFileReadDemo
public class SequenceFileProcessor
{
public static void main(String[] args) throws IOException
{
Configuration conf = new Configuration();
conf.set("mapred.map.child.java.opts", "-Xmx256m");
conf.set("mapred.reduce.child.java.opts", "-Xmx256m");
//conf.set("io.file.buffer.size", "65536");  //10mb/sec improvement, 
jumped from 26mb/sec to 36mb/sec
conf.set("io.file.buffer.size", "131072");  // 15mb/sec improvement, 
jumped from 26mb/sec to 39mb/sec

FileSystem fs = null;
Path path = null;
int total_count = 0;
int count = 0;
long start = System.currentTimeMillis();

for (String uri : args)
{
fs = FileSystem.get(URI.create(uri), conf);
path = new Path(uri);

SequenceFile.Reader reader = null;
try
{
reader = new SequenceFile.Reader(fs, path, conf);
Writable key = (Writable) 
ReflectionUtils.newInstance(reader.getKeyClass(), conf);
Writable value = (Writable) 
ReflectionUtils.newInstance(reader.getValueClass(), conf);
long position = reader.getPosition();
while (reader.next(key, value))
{
   String syncSeen = reader.syncSeen() ? "*" : "";
//System.out.printf("[%s%s]\t%s\t%s\n", position, syncSeen, 
key, value);
position = reader.getPosition(); // beginning of next record
count += 1;
//System.out.println("count is: " + count);

if((count % 100) == 0)
System.out.println("processed " + count + " records");
}
}
finally
{
IOUtils.closeStream(reader);
}
}
total_count += count;
System.out.println("Total count: " + total_count);
System.out.println("Elapsed time: " + ((System.currentTimeMillis() - 
start) / 1000) + " seconds");
}
}
// ^^ SequenceFileReadDemo


/*
* To change this template, choose Tools | Templates
* and open the template in the editor.
*/
package hdfsspeedtest;

import java.io.FileNotFoundException;
import java.io.IOException;
import java.util.Date;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.FileStatus;

/ code is */

public class HDFSSpeedTest {

public static void main(String[] args) throws Exception {

System.out.println(new Date().toString());

Path pt = new Path(args[0]);
try {

// Use this for reading the data.
byte[] buffer = new byte[32*1024*1024];

Configuration conf = new Configuration();
//conf.set(null, null);
FileSystem fs = FileSystem.get(conf);
FileStatus[] inputFiles = fs.listStatus(pt);
long total = 0;

   for(int i=0; imailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 9:04 PM
To: user@hadoop.apache.org
Subject: RE: Hadoop throughput question


Perhaps if Artem posted the presumably-simple code we could get other users to 
benchmark other 4-node systems and compare.

--John Lilley



Artem Ervits mailto:are9...@nyp.org>> wrote:


Setting the property to 64k made the throughput jump to 36mb/sec, 39mb for 128k.

Thank you for the tip.

From: Michael Katzenellenbogen 
[mailto:mich...@cloudera.com]<mailto:[mailto:mich...@cloudera.com]>
Sent: Thursday, January 03, 2013 7:28 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Re: Hadoop throughput question

What is the value of the io.file.buffer.size property? Try tuning it up to 64k 
or 128k and see if this improves performance when reading SequenceFiles.

-Michael

On Jan 3, 2013, at 7:00 PM, Artem Ervits 
mailto:are9...@nyp.org>> wrote:
I will follow up on that certainly, thank you for the information.

So further investigation showed that counting SequenceFile records takes about 
26mb/sec. If I simply read bytes on the same cluster and th

RE: Hadoop throughput question

2013-01-03 Thread Artem Ervits
Setting the property to 64k made the throughput jump to 36mb/sec, 39mb for 128k.

Thank you for the tip.

From: Michael Katzenellenbogen [mailto:mich...@cloudera.com]
Sent: Thursday, January 03, 2013 7:28 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop throughput question

What is the value of the io.file.buffer.size property? Try tuning it up to 64k 
or 128k and see if this improves performance when reading SequenceFiles.

-Michael

On Jan 3, 2013, at 7:00 PM, Artem Ervits 
mailto:are9...@nyp.org>> wrote:
I will follow up on that certainly, thank you for the information.

So further investigation showed that counting SequenceFile records takes about 
26mb/sec. If I simply read bytes on the same cluster and the same file, the 
speed is 70mb/sec. Is there a configuration for optimizing SequenceFile 
processing?

Thank you.

From: John Lilley [mailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 6:09 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: RE: Hadoop throughput question

Unless the Hadoop processing and the OneFS storage are co-located, MapReduce 
can't schedule tasks so as to take advantage of data locality.  You would 
basically be doing a distributed computation against a separate NAS, so 
throughput would be limited by the performance properties of the Insilon NAS 
and the network switch architecture.  Still, 26MB/sec in aggregate is far worse 
than what I'd expect Insilon to deliver, even over a single 1GB connection.
john

From: Artem Ervits [mailto:are9...@nyp.org]<mailto:[mailto:are9...@nyp.org]>
Sent: Thursday, January 03, 2013 4:02 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: RE: Hadoop throughput question

Hadoop is using OneFS, not HDFS in our configuration. Isilon NAS and the Hadoop 
nodes are in the same datacenter but as far as rack locations, I cannot tell.

From: John Lilley [mailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 5:15 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: RE: Hadoop throughput question

Let's suppose you are doing a read-intensive job like, for example, counting 
records.  This is will be disk bandwidth limited.  On a 4-node cluster with 2 
local SATA on each node you should easily read 400MB/sec in aggregate.  When 
you are running the Hadoop cluster, is the Hadoop processing co-located with 
the Ilsilon nodes?  Is Hadoop configured to use OneFS or HDFS?
John

From: Artem Ervits [mailto:are9...@nyp.org]<mailto:[mailto:are9...@nyp.org]>
Sent: Thursday, January 03, 2013 3:00 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Hadoop throughput question

Hello all,

I'd like to pick the community brain on average throughput speeds for a 
moderately specced 4-node Hadoop cluster with 1GigE networking. Is it 
reasonable to expect constant average speeds of 150-200mb/sec on such setup? 
Forgive me if the question is loaded but we're Hadoop cluster with HDFS served 
via EMC Isilon storage. We're getting about 30mb/sec with our machines and we 
do not see a difference in job speed between 2 node cluster and 4 node cluster.

Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.



-

RE: Hadoop throughput question

2013-01-03 Thread Artem Ervits
I will follow up on that certainly, thank you for the information.

So further investigation showed that counting SequenceFile records takes about 
26mb/sec. If I simply read bytes on the same cluster and the same file, the 
speed is 70mb/sec. Is there a configuration for optimizing SequenceFile 
processing?

Thank you.

From: John Lilley [mailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 6:09 PM
To: user@hadoop.apache.org
Subject: RE: Hadoop throughput question

Unless the Hadoop processing and the OneFS storage are co-located, MapReduce 
can't schedule tasks so as to take advantage of data locality.  You would 
basically be doing a distributed computation against a separate NAS, so 
throughput would be limited by the performance properties of the Insilon NAS 
and the network switch architecture.  Still, 26MB/sec in aggregate is far worse 
than what I'd expect Insilon to deliver, even over a single 1GB connection.
john

From: Artem Ervits [mailto:are9...@nyp.org]<mailto:[mailto:are9...@nyp.org]>
Sent: Thursday, January 03, 2013 4:02 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: RE: Hadoop throughput question

Hadoop is using OneFS, not HDFS in our configuration. Isilon NAS and the Hadoop 
nodes are in the same datacenter but as far as rack locations, I cannot tell.

From: John Lilley [mailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 5:15 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: RE: Hadoop throughput question

Let's suppose you are doing a read-intensive job like, for example, counting 
records.  This is will be disk bandwidth limited.  On a 4-node cluster with 2 
local SATA on each node you should easily read 400MB/sec in aggregate.  When 
you are running the Hadoop cluster, is the Hadoop processing co-located with 
the Ilsilon nodes?  Is Hadoop configured to use OneFS or HDFS?
John

From: Artem Ervits [mailto:are9...@nyp.org]<mailto:[mailto:are9...@nyp.org]>
Sent: Thursday, January 03, 2013 3:00 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Hadoop throughput question

Hello all,

I'd like to pick the community brain on average throughput speeds for a 
moderately specced 4-node Hadoop cluster with 1GigE networking. Is it 
reasonable to expect constant average speeds of 150-200mb/sec on such setup? 
Forgive me if the question is loaded but we're Hadoop cluster with HDFS served 
via EMC Isilon storage. We're getting about 30mb/sec with our machines and we 
do not see a difference in job speed between 2 node cluster and 4 node cluster.

Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.








This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Hadoop throughput question

2013-01-03 Thread Artem Ervits
Hadoop is using OneFS, not HDFS in our configuration. Isilon NAS and the Hadoop 
nodes are in the same datacenter but as far as rack locations, I cannot tell.

From: John Lilley [mailto:john.lil...@redpoint.net]
Sent: Thursday, January 03, 2013 5:15 PM
To: user@hadoop.apache.org
Subject: RE: Hadoop throughput question

Let's suppose you are doing a read-intensive job like, for example, counting 
records.  This is will be disk bandwidth limited.  On a 4-node cluster with 2 
local SATA on each node you should easily read 400MB/sec in aggregate.  When 
you are running the Hadoop cluster, is the Hadoop processing co-located with 
the Ilsilon nodes?  Is Hadoop configured to use OneFS or HDFS?
John

From: Artem Ervits [mailto:are9...@nyp.org]<mailto:[mailto:are9...@nyp.org]>
Sent: Thursday, January 03, 2013 3:00 PM
To: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Hadoop throughput question

Hello all,

I'd like to pick the community brain on average throughput speeds for a 
moderately specced 4-node Hadoop cluster with 1GigE networking. Is it 
reasonable to expect constant average speeds of 150-200mb/sec on such setup? 
Forgive me if the question is loaded but we're Hadoop cluster with HDFS served 
via EMC Isilon storage. We're getting about 30mb/sec with our machines and we 
do not see a difference in job speed between 2 node cluster and 4 node cluster.

Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Hadoop throughput question

2013-01-03 Thread Artem Ervits
I have a 4.5gb file with records in SequenceFile format. If I use 
SequenceFile.Reader class to count the records in this file, which amount to 
5.5million records, it takes 176seconds, or roughly 26mb/sec.

Thank you.

From: Michael Katzenellenbogen [mailto:mich...@cloudera.com]
Sent: Thursday, January 03, 2013 5:08 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop throughput question

Loaded question indeed.

How are you measuring that 30mb/s? Is that per machine / NIC? HDFS throughout? 
Some other metric?

-Michael

On Jan 3, 2013, at 5:01 PM, Artem Ervits 
mailto:are9...@nyp.org>> wrote:
Hello all,

I'd like to pick the community brain on average throughput speeds for a 
moderately specced 4-node Hadoop cluster with 1GigE networking. Is it 
reasonable to expect constant average speeds of 150-200mb/sec on such setup? 
Forgive me if the question is loaded but we're Hadoop cluster with HDFS served 
via EMC Isilon storage. We're getting about 30mb/sec with our machines and we 
do not see a difference in job speed between 2 node cluster and 4 node cluster.

Thank you.







This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.







This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.








This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Hadoop throughput question

2013-01-03 Thread Artem Ervits
Hello all,

I'd like to pick the community brain on average throughput speeds for a 
moderately specced 4-node Hadoop cluster with 1GigE networking. Is it 
reasonable to expect constant average speeds of 150-200mb/sec on such setup? 
Forgive me if the question is loaded but we're Hadoop cluster with HDFS served 
via EMC Isilon storage. We're getting about 30mb/sec with our machines and we 
do not see a difference in job speed between 2 node cluster and 4 node cluster.

Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Hadoop on Isilon problem

2012-10-18 Thread Artem Ervits
Yup, that was it. I confused this tmp with another tmp we created before.

Thank you.

-Original Message-
From: Harsh J [mailto:ha...@cloudera.com] 
Sent: Wednesday, October 17, 2012 3:36 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop on Isilon problem

Hi Artem,

I don't think this is a Isilon issue as much as its a /tmp directory permission 
issue. Ensure that /tmp has proper world+rw permissions, OR if thats not 
doable, set the JVM system property of -Djava.io.tmpdir to a local user 
writable directory and also re-point config hadoop.tmp.dir to such a thing 
thats non /tmp but writable by the user (/home/${user.name}/.tmp maybe).

Of course, easiest is just to fix the /tmp to be drwxrwxrwt,

On Thu, Oct 18, 2012 at 12:48 AM, Artem Ervits  wrote:
> Anyone using Hadoop running on Isilon NAS? I am trying to submit a job 
> with a user other than the one running Hadoop and I'm getting the 
> following
> error:
>
>
>
> Exception in thread "main" java.io.IOException: Permission denied
>
> at java.io.UnixFileSystem.createFileExclusively(Native Method)
>
> at java.io.File.checkAndCreate(File.java:1717)
>
> at java.io.File.createTempFile0(File.java:1738)
>
> at java.io.File.createTempFile(File.java:1815)
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:115)
>
>
>
>
>
> Any ideas?
>
>
>
>
>
> Artem Ervits
>
> Data Analyst
>
> New York Presbyterian Hospital
>
>
>
> 
>
> This electronic message is intended to be for the use only of the 
> named recipient, and may contain information that is confidential or 
> privileged.
> If you are not the intended recipient, you are hereby notified that 
> any disclosure, copying, distribution or use of the contents of this 
> message is strictly prohibited.  If you have received this message in 
> error or are not the named recipient, please notify us immediately by 
> contacting the sender at the electronic mail address noted above, and 
> delete and destroy all copies of this message.  Thank you.
>
> 
>
> This electronic message is intended to be for the use only of the 
> named recipient, and may contain information that is confidential or 
> privileged.
> If you are not the intended recipient, you are hereby notified that 
> any disclosure, copying, distribution or use of the contents of this 
> message is strictly prohibited.  If you have received this message in 
> error or are not the named recipient, please notify us immediately by 
> contacting the sender at the electronic mail address noted above, and 
> delete and destroy all copies of this message.  Thank you.
>
>



--
Harsh J




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: Hadoop on Isilon problem

2012-10-17 Thread Artem Ervits
With Isilon, there is no need for hdfs-site configuration file. Isilon takes 
care of replication, although you can certainly add hadoop replication. The 
biggest plus is the scalability for storage layer. We keep a lot of our data in 
Isilon and importing into hdfs will result in two locations of massive amounts 
of data. With this solution, we can read and write in one location. Performance 
wise, I cannot answer yet, so far our standard cluster runs jobs 3x faster than 
Isilon but we have many factors why it is so and results are inconclusive.




Artem Ervits
Data Analyst
New York Presbyterian Hospital

From: Rita [mailto:rmorgan...@gmail.com]
Sent: Wednesday, October 17, 2012 08:23 PM
To: user@hadoop.apache.org 
Subject: Re: Hadoop on Isilon problem

out of curiosity, what does running HDFS give you when running thru an Isilon 
cluster?


On Wed, Oct 17, 2012 at 3:59 PM, Mohit Anchlia 
mailto:mohitanch...@gmail.com>> wrote:
Look at the directory permissions?


On Wed, Oct 17, 2012 at 12:18 PM, Artem Ervits 
mailto:are9...@nyp.org>> wrote:
Anyone using Hadoop running on Isilon NAS? I am trying to submit a job with a 
user other than the one running Hadoop and I’m getting the following error:

Exception in thread "main" java.io.IOException: Permission denied
at java.io.UnixFileSystem.createFileExclusively(Native Method)
at java.io.File.checkAndCreate(File.java:1717)
at java.io.File.createTempFile0(File.java:1738)
at java.io.File.createTempFile(File.java:1815)
at org.apache.hadoop.util.RunJar.main(RunJar.java:115)


Any ideas?


Artem Ervits
Data Analyst
New York Presbyterian Hospital




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.







--
--- Get your facts first, then you can distort them as you please.--




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: Hadoop on Isilon problem

2012-10-17 Thread Artem Ervits
I will try that first thing in the morning, thank you very much!




Artem Ervits
Data Analyst
New York Presbyterian Hospital

- Original Message -
From: Harsh J [mailto:ha...@cloudera.com]
Sent: Wednesday, October 17, 2012 03:36 PM
To: user@hadoop.apache.org 
Subject: Re: Hadoop on Isilon problem

Hi Artem,

I don't think this is a Isilon issue as much as its a /tmp directory
permission issue. Ensure that /tmp has proper world+rw permissions, OR
if thats not doable, set the JVM system property of -Djava.io.tmpdir
to a local user writable directory and also re-point config
hadoop.tmp.dir to such a thing thats non /tmp but writable by the user
(/home/${user.name}/.tmp maybe).

Of course, easiest is just to fix the /tmp to be drwxrwxrwt,

On Thu, Oct 18, 2012 at 12:48 AM, Artem Ervits  wrote:
> Anyone using Hadoop running on Isilon NAS? I am trying to submit a job with
> a user other than the one running Hadoop and I’m getting the following
> error:
>
>
>
> Exception in thread "main" java.io.IOException: Permission denied
>
> at java.io.UnixFileSystem.createFileExclusively(Native Method)
>
> at java.io.File.checkAndCreate(File.java:1717)
>
> at java.io.File.createTempFile0(File.java:1738)
>
> at java.io.File.createTempFile(File.java:1815)
>
> at org.apache.hadoop.util.RunJar.main(RunJar.java:115)
>
>
>
>
>
> Any ideas?
>
>
>
>
>
> Artem Ervits
>
> Data Analyst
>
> New York Presbyterian Hospital
>
>
>
> 
>
> This electronic message is intended to be for the use only of the named
> recipient, and may contain information that is confidential or privileged.
> If you are not the intended recipient, you are hereby notified that any
> disclosure, copying, distribution or use of the contents of this message is
> strictly prohibited.  If you have received this message in error or are not
> the named recipient, please notify us immediately by contacting the sender
> at the electronic mail address noted above, and delete and destroy all
> copies of this message.  Thank you.
>
> 
>
> This electronic message is intended to be for the use only of the named
> recipient, and may contain information that is confidential or privileged.
> If you are not the intended recipient, you are hereby notified that any
> disclosure, copying, distribution or use of the contents of this message is
> strictly prohibited.  If you have received this message in error or are not
> the named recipient, please notify us immediately by contacting the sender
> at the electronic mail address noted above, and delete and destroy all
> copies of this message.  Thank you.
>
>



-- 
Harsh J




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Hadoop on Isilon problem

2012-10-17 Thread Artem Ervits
Anyone using Hadoop running on Isilon NAS? I am trying to submit a job with a 
user other than the one running Hadoop and I'm getting the following error:

Exception in thread "main" java.io.IOException: Permission denied
at java.io.UnixFileSystem.createFileExclusively(Native Method)
at java.io.File.checkAndCreate(File.java:1717)
at java.io.File.createTempFile0(File.java:1738)
at java.io.File.createTempFile(File.java:1815)
at org.apache.hadoop.util.RunJar.main(RunJar.java:115)


Any ideas?


Artem Ervits
Data Analyst
New York Presbyterian Hospital





This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Hadoop recovery test

2012-09-19 Thread Artem Ervits
Harsh,

That was actually what it was. I was messing with HBase install and edited the 
/etc/hosts file to add the 127.0.0.1 address. Once I removed the entry, 
datanodes were able to see the namenode. I was also able to successfully test 
the recovery. Hadoop fsck -blocks reports a healthy filesystem now.

Thank you very much.

-Original Message-
From: Harsh J [mailto:ha...@cloudera.com] 
Sent: Tuesday, September 18, 2012 11:37 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop recovery test

Artem,

If you check the logs of the other DNs, do you see issues with connectivity to 
NameNode?

Basic questions, but need to ask to be sure: have you checked if the firewalls 
are down or properly configured? Are you sure that your hostname of the master 
machine resolves not to the loopback address but to the external interface 
provided IP?

On Tue, Sep 18, 2012 at 10:29 PM, Artem Ervits  wrote:
> I didn't realize that I didn't edit core-site and mapred-site on all machines 
> to point to the new namenode. Although that didn't make a difference, I still 
> see only one datanode which Is also the namenode:
>
> Datanodes available: 1 (1 total, 0 dead)
>
> Name: 127.0.0.1:50010
> Decommission Status : Normal
> Configured Capacity: 105425190912 (98.18 GB) DFS Used: 1058557952 
> (1009.52 MB) Non DFS Used: 200396800 (191.11 MB) DFS Remaining: 
> 104166236160(97.01 GB) DFS Used%: 1% DFS Remaining%: 98.81% Last 
> contact: Tue Sep 18 12:58:07 EDT 2012
>
> The other strange thing is that it points to local 127.0.0.1 rather than 
> namenode's IP.
>
> -Original Message-
> From: Artem Ervits [mailto:are9...@nyp.org]
> Sent: Tuesday, September 18, 2012 9:57 AM
> To: user@hadoop.apache.org
> Cc: James Brown
> Subject: RE: Hadoop recovery test
>
> No it only sees itself. It doesn't see the rest of the nodes.
>
> -Original Message-
> From: James Brown [mailto:jb...@syndicate.net]
> Sent: Monday, September 17, 2012 5:49 PM
> To: user@hadoop.apache.org
> Subject: Re: Hadoop recovery test
>
> Does the new NameNode server see all of the DataNodes?
>
> On 9/17/2012 2:38 PM, Artem Ervits wrote:
>> Hello all,
>>
>> I am testing the Hadoop recovery as per 
>> http://wiki.apache.org/hadoop/NameNode document. But instead of using 
>> an NFS share, I am copying to another directory. Then when I shut 
>> down the cluster, I scp that directory to another server and start 
>> Hadoop cluster using that machine as the namenode. I see in the log 
>> that some blocks are corrupt and/or missing. Do I have to wait for 
>> replication to recover all blocks or am I doing something else 
>> altogether? I am using Hadoop 1.0.3. Can someone point me to a more 
>> detailed document than the wiki in case I'm doing something wrong.
>>
>> p.s. if I restart the cluster using the original namenode, filesystem 
>> reports as healthy.
>>
>> Thank you.
>>
>> .
>>
>> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: CORRUPT block
>> blk_9043419219670949307
>>
>> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks of 
>> total size 4 B...
>>
>> /user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
>> Under replicated blk_-976282286234272458_1079. Target Replicas is 3 
>> but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:
>> Under replicated blk_137658109390447967_1075. Target Replicas is 3 
>> but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen/_partition.lst:  Under replicated 
>> blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen/part-0:  Under replicated 
>> blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen/part-1:  Under replicated 
>> blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
>>
>> ..
>>
>> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
>> Under replicated blk_1137779303840586677_1089. Target Replicas is 3 
>> but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:
>> Under replicated blk_7701720691642589882_1086. Target Replicas is 3 
>> but found 1 replica(s).
>>
>> .
>>
>> /user/hduser/teragen-out/part-0: CORRUPT block
>> blk_8059469267617478950
>>
>> /user/hduser/teragen-out/part-0: MISSING 1 blocks of total size
>> 100 B..

RE: Hadoop recovery test

2012-09-18 Thread Artem Ervits
I didn't realize that I didn't edit core-site and mapred-site on all machines 
to point to the new namenode. Although that didn't make a difference, I still 
see only one datanode which Is also the namenode:

Datanodes available: 1 (1 total, 0 dead)

Name: 127.0.0.1:50010
Decommission Status : Normal
Configured Capacity: 105425190912 (98.18 GB)
DFS Used: 1058557952 (1009.52 MB)
Non DFS Used: 200396800 (191.11 MB)
DFS Remaining: 104166236160(97.01 GB)
DFS Used%: 1%
DFS Remaining%: 98.81%
Last contact: Tue Sep 18 12:58:07 EDT 2012

The other strange thing is that it points to local 127.0.0.1 rather than 
namenode's IP. 

-Original Message-
From: Artem Ervits [mailto:are9...@nyp.org] 
Sent: Tuesday, September 18, 2012 9:57 AM
To: user@hadoop.apache.org
Cc: James Brown
Subject: RE: Hadoop recovery test

No it only sees itself. It doesn't see the rest of the nodes.

-Original Message-
From: James Brown [mailto:jb...@syndicate.net]
Sent: Monday, September 17, 2012 5:49 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop recovery test

Does the new NameNode server see all of the DataNodes?

On 9/17/2012 2:38 PM, Artem Ervits wrote:
> Hello all,
>
> I am testing the Hadoop recovery as per 
> http://wiki.apache.org/hadoop/NameNode document. But instead of using 
> an NFS share, I am copying to another directory. Then when I shut down 
> the cluster, I scp that directory to another server and start Hadoop 
> cluster using that machine as the namenode. I see in the log that some 
> blocks are corrupt and/or missing. Do I have to wait for replication 
> to recover all blocks or am I doing something else altogether? I am 
> using Hadoop 1.0.3. Can someone point me to a more detailed document 
> than the wiki in case I'm doing something wrong.
>
> p.s. if I restart the cluster using the original namenode, filesystem 
> reports as healthy.
>
> Thank you.
>
> .
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: CORRUPT block
> blk_9043419219670949307
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks of 
> total size 4 B...
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
> Under replicated blk_-976282286234272458_1079. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:
> Under replicated blk_137658109390447967_1075. Target Replicas is 3 but 
> found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_partition.lst:  Under replicated 
> blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-0:  Under replicated 
> blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-1:  Under replicated 
> blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
>
> ..
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
> Under replicated blk_1137779303840586677_1089. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:
> Under replicated blk_7701720691642589882_1086. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/part-0: CORRUPT block
> blk_8059469267617478950
>
> /user/hduser/teragen-out/part-0: MISSING 1 blocks of total size
> 100 B...
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_1347458495941_hduser_TeraValidate:
> Under replicated blk_5680565744062298575_1098. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_conf.xml:
> Under replicated blk_1566253937037013126_1095. Target Replicas is 3 
> but found 1 replica(s).
>
> .Status: CORRUPT
>
> Total size:1050720258 B
>
> Total dirs:39
>
> Total files:   32
>
> Total blocks (validated):  42 (avg. block size 25017149 B)
>
>
>
>CORRUPT FILES:2
>
>MISSING BLOCKS:   2
>
>MISSING SIZE: 104 B
>
>CORRUPT BLOCKS:   2
>
>
>
> Minimally replicated blocks:   40 (95.2381 %)
>
> Over-replicated blocks:0 (0.0 %)
>
> Under-replicated blocks:   40 (95.2381 %)
>
> Mis-replicated blocks: 0 (0.0 %)
>
> Default replication factor:3
>
> Average block replication: 0.95238096
>
> Corrupt blocks:2
>
> Missing replicas:  80 (200.0 %)
>
> Number of data-nodes:  1
>
> Number of racks:   

RE: Hadoop recovery test

2012-09-18 Thread Artem Ervits
Thanks Harsh,

I'm aware of the implications of copying periodically. This is just a test 
until I get an NFS share to play with. Do you just let Hadoop write to two 
directories where one is an NFS share or is there another way?

-Original Message-
From: Harsh J [mailto:ha...@cloudera.com] 
Sent: Monday, September 17, 2012 10:44 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop recovery test

Hi Artem,

You are running 1 DN in this cluster from what I see, and hence you can ignore 
the reports that go: Under replicated blk_7701720691642589882_1086. Target 
Replicas is 3 but found 1 replica(s).

The two truly missing blocks are:

/hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks
/user/hduser/teragen-out/part-0: MISSING 1 blocks

Which may be cause of those being written at the time of your copy of the 
fsimage and edits (thats a wrong way to go about it, btw - you should configure 
for redundant writes such that you also sustain failures, not copy it 
periodically - thats not a consistent way to keep a backup, and you can rather 
go for dfsadmin methods to fetchImage instead). Does that sound likely?

On Tue, Sep 18, 2012 at 3:08 AM, Artem Ervits  wrote:
> Hello all,
>
>
>
> I am testing the Hadoop recovery as per 
> http://wiki.apache.org/hadoop/NameNode document. But instead of using 
> an NFS share, I am copying to another directory. Then when I shut down 
> the cluster, I scp that directory to another server and start Hadoop 
> cluster using that machine as the namenode. I see in the log that some 
> blocks are corrupt and/or missing. Do I have to wait for replication 
> to recover all blocks or am I doing something else altogether? I am 
> using Hadoop 1.0.3. Can someone point me to a more detailed document 
> than the wiki in case I'm doing something wrong.
>
>
>
> p.s. if I restart the cluster using the original namenode, filesystem 
> reports as healthy.
>
>
>
> Thank you.
>
>
>
> .
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: CORRUPT block
> blk_9043419219670949307
>
>
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks of 
> total size 4 B...
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
> Under replicated blk_-976282286234272458_1079. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:  
> Under replicated blk_137658109390447967_1075. Target Replicas is 3 but 
> found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_partition.lst:  Under replicated 
> blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-0:  Under replicated 
> blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-1:  Under replicated 
> blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
>
> ..
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
> Under replicated blk_1137779303840586677_1089. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:
> Under replicated blk_7701720691642589882_1086. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/part-0: CORRUPT block 
> blk_8059469267617478950
>
>
>
> /user/hduser/teragen-out/part-0: MISSING 1 blocks of total size 
> 100 B...
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_1347458495941_hduser_TeraValidate:
> Under replicated blk_5680565744062298575_1098. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_conf.xml:
> Under replicated blk_1566253937037013126_1095. Target Replicas is 3 
> but found 1 replica(s).
>
> .Status: CORRUPT
>
> Total size:1050720258 B
>
> Total dirs:39
>
> Total files:   32
>
> Total blocks (validated):  42 (avg. block size 25017149 B)
>
>   
>
>   CORRUPT FILES:2
>
>   MISSING BLOCKS:   2
>
>   MISSING SIZE: 104 B
>
>   CORRUPT BLOCKS:   2
>
>   
>
> Minimally replicated blocks:   40 (95.2381 %)
>
> Over-replicated blocks:0 (0.0 %)
>
> Under-replicated blocks:   40 (95.2381 %)
>
> Mis-replicated blocks: 0 (0.0 %)
>
> Default replication factor:3
>
> Average block replication: 0.95238096
>
> Corrupt blocks:2
>
> Missing replicas:  80 (200.0 %)
&g

RE: Hadoop recovery test

2012-09-18 Thread Artem Ervits
I tried copying two ways. Once while Hadoop was running and second time when I 
shut down the original cluster. I used scp -r command, is there a better 
option, rsync? I also tried scp with -rp4 switch but still can't get the 
folders to look identical

Original node:

total 20
drwxr-xr-x. 16 hduser hadoop 4096 Sep 17 13:31 ..
drwxrwx---.  2 hduser hadoop 4096 Sep 17 13:41 image
drwxrwx---.  2 hduser hadoop 4096 Sep 17 16:39 previous.checkpoint
drwxrwx---.  2 hduser hadoop 4096 Sep 18 09:57 current
drwxrwx---.  5 hduser hadoop 4096 Sep 18 09:59 .

new namenode:

total 20
drwxrwx---.  2 hduser hadoop 4096 Sep 17 13:41 image
drwxrwx---.  2 hduser hadoop 4096 Sep 17 16:39 previous.checkpoint
drwxrwx---.  2 hduser hadoop 4096 Sep 18 09:57 current
drwxrwx---.  5 hduser hadoop 4096 Sep 18 09:59 .
drwxr-xr-x. 17 hduser hadoop 4096 Sep 18 10:03 ..



From: Robert Molina [mailto:rmol...@hortonworks.com]
Sent: Monday, September 17, 2012 5:55 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop recovery test

Hi Artem,
At what point do you do the copy, was namenode still running? Does the copy of 
the edits file and fsimage file match up with the original (i.e filesize)?

-Robert
On Mon, Sep 17, 2012 at 2:38 PM, Artem Ervits 
mailto:are9...@nyp.org>> wrote:
Hello all,

I am testing the Hadoop recovery as per http://wiki.apache.org/hadoop/NameNode 
document. But instead of using an NFS share, I am copying to another directory. 
Then when I shut down the cluster, I scp that directory to another server and 
start Hadoop cluster using that machine as the namenode. I see in the log that 
some blocks are corrupt and/or missing. Do I have to wait for replication to 
recover all blocks or am I doing something else altogether? I am using Hadoop 
1.0.3. Can someone point me to a more detailed document than the wiki in case 
I'm doing something wrong.

p.s. if I restart the cluster using the original namenode, filesystem reports 
as healthy.

Thank you.

.
/hdfs/hadoop/tmp/mapred/system/jobtracker.info<http://jobtracker.info>: CORRUPT 
block blk_9043419219670949307

/hdfs/hadoop/tmp/mapred/system/jobtracker.info<http://jobtracker.info>: MISSING 
1 blocks of total size 4 B...
/user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
  Under replicated blk_-976282286234272458_1079. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:  Under 
replicated blk_137658109390447967_1075. Target Replicas is 3 but found 1 
replica(s).
.
/user/hduser/teragen/_partition.lst:  Under replicated 
blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
.
/user/hduser/teragen/part-0:  Under replicated 
blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
.
/user/hduser/teragen/part-1:  Under replicated 
blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
..
/user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
  Under replicated blk_1137779303840586677_1089. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:  Under 
replicated blk_7701720691642589882_1086. Target Replicas is 3 but found 1 
replica(s).
.
/user/hduser/teragen-out/part-0: CORRUPT block blk_8059469267617478950

/user/hduser/teragen-out/part-0: MISSING 1 blocks of total size 100 B...
/user/hduser/teragen-validate/_logs/history/job_201209120941_0004_1347458495941_hduser_TeraValidate:
  Under replicated blk_5680565744062298575_1098. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen-validate/_logs/history/job_201209120941_0004_conf.xml:  
Under replicated blk_1566253937037013126_1095. Target Replicas is 3 but found 1 
replica(s).
.Status: CORRUPT
Total size:1050720258 B
Total dirs:39
Total files:   32
Total blocks (validated):  42 (avg. block size 25017149 B)
  
  CORRUPT FILES:2
  MISSING BLOCKS:   2
  MISSING SIZE: 104 B
  CORRUPT BLOCKS:   2
  
Minimally replicated blocks:   40 (95.2381 %)
Over-replicated blocks:0 (0.0 %)
Under-replicated blocks:   40 (95.2381 %)
Mis-replicated blocks: 0 (0.0 %)
Default replication factor:3
Average block replication: 0.95238096
Corrupt blocks:2
Missing replicas:  80 (200.0 %)
Number of data-nodes:  1
Number of racks:   1
FSCK ended at Mon Sep 17 17:29:08 EDT 2012 in 21 milliseconds


The filesystem under path '/' is CORRUPT


Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
discl

RE: Hadoop recovery test

2012-09-18 Thread Artem Ervits
No it only sees itself. It doesn't see the rest of the nodes.

-Original Message-
From: James Brown [mailto:jb...@syndicate.net] 
Sent: Monday, September 17, 2012 5:49 PM
To: user@hadoop.apache.org
Subject: Re: Hadoop recovery test

Does the new NameNode server see all of the DataNodes?

On 9/17/2012 2:38 PM, Artem Ervits wrote:
> Hello all,
>
> I am testing the Hadoop recovery as per 
> http://wiki.apache.org/hadoop/NameNode document. But instead of using 
> an NFS share, I am copying to another directory. Then when I shut down 
> the cluster, I scp that directory to another server and start Hadoop 
> cluster using that machine as the namenode. I see in the log that some 
> blocks are corrupt and/or missing. Do I have to wait for replication 
> to recover all blocks or am I doing something else altogether? I am 
> using Hadoop 1.0.3. Can someone point me to a more detailed document 
> than the wiki in case I'm doing something wrong.
>
> p.s. if I restart the cluster using the original namenode, filesystem 
> reports as healthy.
>
> Thank you.
>
> .
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: CORRUPT block
> blk_9043419219670949307
>
> /hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks of 
> total size 4 B...
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
> Under replicated blk_-976282286234272458_1079. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:
> Under replicated blk_137658109390447967_1075. Target Replicas is 3 but 
> found 1 replica(s).
>
> .
>
> /user/hduser/teragen/_partition.lst:  Under replicated 
> blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-0:  Under replicated 
> blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
>
> .
>
> /user/hduser/teragen/part-1:  Under replicated 
> blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
>
> ..
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
> Under replicated blk_1137779303840586677_1089. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:
> Under replicated blk_7701720691642589882_1086. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-out/part-0: CORRUPT block 
> blk_8059469267617478950
>
> /user/hduser/teragen-out/part-0: MISSING 1 blocks of total size
> 100 B...
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_1347458495941_hduser_TeraValidate:
> Under replicated blk_5680565744062298575_1098. Target Replicas is 3 
> but found 1 replica(s).
>
> .
>
> /user/hduser/teragen-validate/_logs/history/job_201209120941_0004_conf.xml:
> Under replicated blk_1566253937037013126_1095. Target Replicas is 3 
> but found 1 replica(s).
>
> .Status: CORRUPT
>
> Total size:1050720258 B
>
> Total dirs:39
>
> Total files:   32
>
> Total blocks (validated):  42 (avg. block size 25017149 B)
>
>
>
>CORRUPT FILES:2
>
>MISSING BLOCKS:   2
>
>MISSING SIZE: 104 B
>
>CORRUPT BLOCKS:   2
>
>
>
> Minimally replicated blocks:   40 (95.2381 %)
>
> Over-replicated blocks:0 (0.0 %)
>
> Under-replicated blocks:   40 (95.2381 %)
>
> Mis-replicated blocks: 0 (0.0 %)
>
> Default replication factor:3
>
> Average block replication: 0.95238096
>
> Corrupt blocks:2
>
> Missing replicas:  80 (200.0 %)
>
> Number of data-nodes:  1
>
> Number of racks:   1
>
> FSCK ended at Mon Sep 17 17:29:08 EDT 2012 in 21 milliseconds
>
> The filesystem under path '/' is CORRUPT
>
> Artem Ervits
>
> Data Analyst
>
> New York Presbyterian Hospital
>
>
> --
> -- This electronic message is intended to be for the use only of the 
> named recipient, and may contain information that is confidential or 
> privileged. If you are not the intended recipient, you are hereby 
> notified that any disclosure, copying, distribution or use of the 
> contents of this message is strictly prohibited. If you have received 
> this message in error or are not the named recipient, please notify us 
> immediately by contacting the sender at the electronic mail address 
>

Hadoop recovery test

2012-09-17 Thread Artem Ervits
Hello all,

I am testing the Hadoop recovery as per http://wiki.apache.org/hadoop/NameNode 
document. But instead of using an NFS share, I am copying to another directory. 
Then when I shut down the cluster, I scp that directory to another server and 
start Hadoop cluster using that machine as the namenode. I see in the log that 
some blocks are corrupt and/or missing. Do I have to wait for replication to 
recover all blocks or am I doing something else altogether? I am using Hadoop 
1.0.3. Can someone point me to a more detailed document than the wiki in case 
I'm doing something wrong.

p.s. if I restart the cluster using the original namenode, filesystem reports 
as healthy.

Thank you.

.
/hdfs/hadoop/tmp/mapred/system/jobtracker.info: CORRUPT block 
blk_9043419219670949307

/hdfs/hadoop/tmp/mapred/system/jobtracker.info: MISSING 1 blocks of total size 
4 B...
/user/hduser/teragen/_logs/history/job_201209120941_0002_1347458152167_hduser_TeraGen:
  Under replicated blk_-976282286234272458_1079. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen/_logs/history/job_201209120941_0002_conf.xml:  Under 
replicated blk_137658109390447967_1075. Target Replicas is 3 but found 1 
replica(s).
.
/user/hduser/teragen/_partition.lst:  Under replicated 
blk_-3005280481530403302_1080. Target Replicas is 3 but found 1 replica(s).
.
/user/hduser/teragen/part-0:  Under replicated 
blk_-7008813028808832816_1077. Target Replicas is 3 but found 1 replica(s).
.
/user/hduser/teragen/part-1:  Under replicated 
blk_-5256967771026054061_1078. Target Replicas is 3 but found 1 replica(s).
..
/user/hduser/teragen-out/_logs/history/job_201209120941_0003_1347458249920_hduser_TeraSort:
  Under replicated blk_1137779303840586677_1089. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen-out/_logs/history/job_201209120941_0003_conf.xml:  Under 
replicated blk_7701720691642589882_1086. Target Replicas is 3 but found 1 
replica(s).
.
/user/hduser/teragen-out/part-0: CORRUPT block blk_8059469267617478950

/user/hduser/teragen-out/part-0: MISSING 1 blocks of total size 100 B...
/user/hduser/teragen-validate/_logs/history/job_201209120941_0004_1347458495941_hduser_TeraValidate:
  Under replicated blk_5680565744062298575_1098. Target Replicas is 3 but found 
1 replica(s).
.
/user/hduser/teragen-validate/_logs/history/job_201209120941_0004_conf.xml:  
Under replicated blk_1566253937037013126_1095. Target Replicas is 3 but found 1 
replica(s).
.Status: CORRUPT
Total size:1050720258 B
Total dirs:39
Total files:   32
Total blocks (validated):  42 (avg. block size 25017149 B)
  
  CORRUPT FILES:2
  MISSING BLOCKS:   2
  MISSING SIZE: 104 B
  CORRUPT BLOCKS:   2
  
Minimally replicated blocks:   40 (95.2381 %)
Over-replicated blocks:0 (0.0 %)
Under-replicated blocks:   40 (95.2381 %)
Mis-replicated blocks: 0 (0.0 %)
Default replication factor:3
Average block replication: 0.95238096
Corrupt blocks:2
Missing replicas:  80 (200.0 %)
Number of data-nodes:  1
Number of racks:   1
FSCK ended at Mon Sep 17 17:29:08 EDT 2012 in 21 milliseconds


The filesystem under path '/' is CORRUPT


Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named re

Re: Hadoop and MainFrame integration

2012-08-28 Thread Artem Ervits
Can you read the data off backup tapes and dump it to flat files?


Artem Ervits
Data Analyst
New York Presbyterian Hospital

From: Marcos Ortiz [mailto:mlor...@uci.cu]
Sent: Tuesday, August 28, 2012 06:51 PM
To: user@hadoop.apache.org 
Cc: Siddharth Tiwari 
Subject: Re: Hadoop and MainFrame integration

The problem with it, is that Hadoop depends on top of HDFS to storage in blocks 
of 64/128 MB of size (or the size that you determine, 64 MB is the de-facto 
size), and then make the calculations.
So, you need to move all your data to a HDFS cluster to use data in MapReduce 
jobs if you want to make the calculations with Hadoop.
Best wishes

El 28/08/2012 12:24, Siddharth Tiwari escribió:
Hi Users.

We have flat files on mainframes with around a billion records. We need to sort 
them and then use them with different jobs on mainframe for report generation. 
I was wondering was there any way I could integrate the mainframe with hadoop 
do the sorting and keep the file on the sever itself ( I do not want to ftp the 
file to a hadoop cluster and then ftp back the sorted file to Mainframe as it 
would waste MIPS and nullify the advantage ). This way I could save on MIPS and 
ultimately improve profitability.

Thank you in advance


**
Cheers !!!
Siddharth Tiwari
Have a refreshing day !!!
"Every duty is holy, and devotion to duty is the highest form of worship of 
God.�
"Maybe other people will try to limit me but I don't limit myself"

[http://universidad.uci.cu/email.gif]
<http://www.uci.cu/>



[http://universidad.uci.cu/email.gif]
<http://www.uci.cu/>

This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Re: hadoop download path missing

2012-08-25 Thread Artem Ervits
I also tried to download hadoop last night and all mirrors I'd tried returned 
error. At the end I gave up and searched my work pc for a download file.



Artem Ervits
Data Analyst
New York Presbyterian Hospital

- Original Message -
From: Harsh J [mailto:ha...@cloudera.com]
Sent: Saturday, August 25, 2012 10:11 AM
To: user@hadoop.apache.org 
Subject: Re: hadoop download path missing

Hi Steven,

Please file an INFRA issue with a list of URLs that are inaccessible.
I tried a few and found them to work at the moment. It may have been a
transient issue?

On Fri, Aug 24, 2012 at 9:10 PM, Steven Willis  wrote:
> Hi Sonal,
>
>
>
> Yes, that URL is working, but the vast majority of the mirrors listed at:
> http://www.apache.org/dyn/closer.cgi/hadoop/core/ which is what
> http://hadoop.apache.org/common/releases.html#Download points users to are
> returning 404s. In fact I don’t even see techartifact on the list of
> mirrors.
>
>
>
> -Steve Willis
>
>
>
> From: Sonal Goyal [mailto:sonalgoy...@gmail.com]
> Sent: Friday, August 24, 2012 11:35 AM
> To: user@hadoop.apache.org
> Subject: Re: hadoop download path missing
>
>
>
> I just tried and could go to
> http://apache.techartifact.com/mirror/hadoop/common/hadoop-2.0.1-alpha/
>
>
>
> Is this still happening for you?
>
>
> Best Regards,
> Sonal
> Crux: Reporting for HBase
> Nube Technologies
>
>
>
>
>
>
> On Fri, Aug 24, 2012 at 8:59 PM, Steven Willis  wrote:
>
> All the links at: http://www.apache.org/dyn/closer.cgi/hadoop/common/ are
> returning 404s, even the backup site at:
> http://www.us.apache.org/dist/hadoop/common/. However, the eu site:
> http://www.eu.apache.org/dist/hadoop/common/ does work.
>
> -Steven Willis
>
>



--
Harsh J

This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Hadoop web interface authentication

2012-08-09 Thread Artem Ervits
Hello all,

I added the following setting  
hadoop.http.authentication.simple.anonymous.allowed. Now when I try to login to 
Web UI, I get 401 error unless I specify user.name=username. This is exactly 
what I want but I noticed that I can pass any user name and it will work as 
long as it is not null. What else am I missing from the core-site file to make 
it work? Also, is there a way I can add a configuration so that a password 
would be required to get access? I am aware of Kerberos authentication method 
but I have it configured with Active directory and until I find the correct 
location of the keytab, I can't use it.

Thanks all.

Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





Hadoop web interface authentication

2012-08-09 Thread Artem Ervits
Hello all,

I added the following setting  
hadoop.http.authentication.simple.anonymous.allowed. Now when I try to login to 
Web UI, I get 401 error unless I specify user.name=username. This is exactly 
what I want but I noticed that I can pass any user name and it will work as 
long as it is not null. What else am I missing from the core-site file to make 
it work? Also, is there a way I can add a configuration so that a password 
would be required to get access? I am aware of Kerberos authentication method 
but I have it configured with Active directory and until I find the correct 
location of the keytab, I can't use it.

Thanks all.

Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.





RE: Setting up HTTP authentication

2012-08-09 Thread Artem Ervits
This was a version dependency issue. The class is not in 0.20.203.0.

From: Artem Ervits [mailto:are9...@nyp.org]
Sent: Wednesday, August 08, 2012 2:34 PM
To: user@hadoop.apache.org
Subject: Setting up HTTP authentication

Hello all,

I followed the 1.0.3 docs to setup http simple authentication. I'd like to pass 
a username to open the web interface. My new settings for core-site.xml are 
below:


hadoop.http.filter.initializers
org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.security.AuthenticationFilterInitializer
HTTP Authentication document in hadoop tar file



hadoop.http.authentication.type
simple
authentication type for web UI



hadoop.http.authentication.token.validity
36000
how long authentication token is valid before it needs to be 
renewed



hadoop.http.authentication.signature.secret
test
signature secret for signing authentication tokens



hadoop.http.authentication.cookie.domain

domain to use for the http cookie that stores authentication 
token



hadoop.http.authentication.simple.anonymous.allowed
false
anonymous web UI requests enabled or disabled



Unfortunately, I don't have a 1.0.3 hadoop cluster handy and am testing this on 
0.20.203.0. When I start the cluster, jobtracker doesn't start. Looking at the 
logs, I see the following:

tasktracker log:

2012-08-08 13:40:26,178 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: 
Source name ugi already exists!

and

2012-08-08 13:40:26,445 ERROR org.apache.hadoop.mapred.TaskTracker: Can not 
start task tracker because java.lang.RuntimeException: 
java.lang.ClassNotFoundException: 
org.apache.hadoop.security.AuthenticationFilterInitializer

Namenode log shows:

2012-08-08 14:18:17,839 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: master/10.0.0.101:54310. Already tried 9 time(s).
2012-08-08 14:18:17,839 INFO org.apache.hadoop.ipc.RPC: Server at 
master/10.0.0.101:54310 not available yet, Z...
2012-08-08 14:18:19,841 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: master/10.0.0.101:54310. Already tried 0 time(s).

My question is, is the 
org.apache.hadoop.security.AuthenticationFilterInitiazlier only available in 
1.* releases or the problem is somewhere else?

Thank you.

Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.







This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.









This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.







Confidential Information subject to NYP's (and its affiliates') information 
management and security policies (http://infonet.nyp.org/QA/HospitalManual).




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may co

Setting up HTTP authentication

2012-08-08 Thread Artem Ervits
Hello all,

I followed the 1.0.3 docs to setup http simple authentication. I'd like to pass 
a username to open the web interface. My new settings for core-site.xml are 
below:


hadoop.http.filter.initializers
org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.security.AuthenticationFilterInitializer
HTTP Authentication document in hadoop tar file



hadoop.http.authentication.type
simple
authentication type for web UI



hadoop.http.authentication.token.validity
36000
how long authentication token is valid before it needs to be 
renewed



hadoop.http.authentication.signature.secret
test
signature secret for signing authentication tokens



hadoop.http.authentication.cookie.domain

domain to use for the http cookie that stores authentication 
token



hadoop.http.authentication.simple.anonymous.allowed
false
anonymous web UI requests enabled or disabled



Unfortunately, I don't have a 1.0.3 hadoop cluster handy and am testing this on 
0.20.203.0. When I start the cluster, jobtracker doesn't start. Looking at the 
logs, I see the following:

tasktracker log:

2012-08-08 13:40:26,178 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: 
Source name ugi already exists!

and

2012-08-08 13:40:26,445 ERROR org.apache.hadoop.mapred.TaskTracker: Can not 
start task tracker because java.lang.RuntimeException: 
java.lang.ClassNotFoundException: 
org.apache.hadoop.security.AuthenticationFilterInitializer

Namenode log shows:

2012-08-08 14:18:17,839 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: master/10.0.0.101:54310. Already tried 9 time(s).
2012-08-08 14:18:17,839 INFO org.apache.hadoop.ipc.RPC: Server at 
master/10.0.0.101:54310 not available yet, Z...
2012-08-08 14:18:19,841 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: master/10.0.0.101:54310. Already tried 0 time(s).

My question is, is the 
org.apache.hadoop.security.AuthenticationFilterInitiazlier only available in 
1.* releases or the problem is somewhere else?

Thank you.

Artem Ervits
Data Analyst
New York Presbyterian Hospital



This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged. If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited. If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message. Thank you.




This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.






This electronic message is intended to be for the use only of the named 
recipient, and may contain information that is confidential or privileged.  If 
you are not the intended recipient, you are hereby notified that any 
disclosure, copying, distribution or use of the contents of this message is 
strictly prohibited.  If you have received this message in error or are not the 
named recipient, please notify us immediately by contacting the sender at the 
electronic mail address noted above, and delete and destroy all copies of this 
message.  Thank you.