Re: What is the class that launches the reducers?

2016-08-26 Thread xeon Mailinglist
rify the map output? It's only partially > dumped to disk. None of the intermediate data goes into HDFS. > > Daniel > > On Aug 25, 2016 4:10 PM, "xeon Mailinglist" <xeonmailingl...@gmail.com> > wrote: > >> But then I need to set identity maps to run the reduc

Re: What is the class that launches the reducers?

2016-08-25 Thread xeon Mailinglist
ra.com> wrote: One thing you can try is to write a map-only job first and then verify the map out. On Thu, Aug 25, 2016 at 1:18 PM, xeon Mailinglist <xeonmailingl...@gmail.com > wrote: > I am using Mapreduce v2. > > On Aug 25, 2016 8:18 PM, "xeon Mailinglist" <xeonmaili

Re: What is the class that launches the reducers?

2016-08-25 Thread xeon Mailinglist
I am using Mapreduce v2. On Aug 25, 2016 8:18 PM, "xeon Mailinglist" <xeonmailingl...@gmail.com> wrote: > I am trying to implement a mechanism in MapReduce v2 that allows to > suspend and resume a job. I must suspend a job when all the mappers finish, > and resume the

What is the class that launches the reducers?

2016-08-25 Thread xeon Mailinglist
I am trying to implement a mechanism in MapReduce v2 that allows to suspend and resume a job. I must suspend a job when all the mappers finish, and resume the job from that point after some time. I do this, because I want to verify the integrity of the map output before executing the reducers. I

Fwd: Submit, suspend and resume a mapreduce job execution

2016-08-21 Thread xeon Mailinglist
I know that it is not possible to suspend and resume mapreduce job, but I really need to find a workaround. I have looked to the ChainedJobs and to the CapacityScheduler, but I am really clueless on what to do. The main goal was to suspend a job when the map tasks finish and the reduce tasks

Improve IdentityMapper code for wordcount

2016-08-21 Thread xeon Mailinglist
Hi, I have created a map method that reads the map output of the wordcount example [1]. This example is away from using the IdentityMapper.class that MapReduce offers, but this is the only way that I have found to make a working IdentityMapper for the Wordcount. The only problem is that this

Pause between tasks or jobs?

2016-08-11 Thread xeon Mailinglist
I am looking for a way to pause a chained job or a chained task. I want to do this because I want to validate the output of each map or reduce phase, or between each job execution. Is it possible to pause the execution of chained jobs or chained mappers or reducers in MapReduce V2. I was looking

Where is the temp output data of a map or reduce tasks

2016-08-11 Thread xeon Mailinglist
With MapReduce v2 (Yarn), the output data that comes out from a map or a reduce task is saved in the local disk or the HDFS when all the tasks finish. Since tasks end at different times, I was expecting that the data were written as a task finish. For example, task 0 finish and so the output is

Re: Copy data from local disc with WebHDFS?

2015-03-02 Thread xeon Mailinglist
1. I am using these 2 commands below to try to copy data from local disk to HDFS. Unfortunately these commands are not working, and I don't understand why they are not working. I have configured HDFS to use WebHDFS protocol. How I copy data from the local disk to HDFS using WebHDfS protocol?

Re: 1 job with Input data from 2 HDFS?

2015-02-27 Thread xeon Mailinglist
Hi, I don't understand this part of your answer: read the other as a side-input directly by creating a client.. If I consider both inputs through the InputFormat, this means that a job will contain both input path in its configuration, and this is enough to work. So, what is the other? Is is the

Copy data between clusters during the job execution.

2015-02-02 Thread xeon Mailinglist
Hi I want to have a job that copies the map output, or the reduce output to another hdfs. Is is possible? E.g., the job runs in cluster 1 and takes the input from this cluster. Then, before the job finishes, it copies the map output or the reduce output to the hdfs in the cluster 2. Thanks,

Set Hadoop MRv2 behind a NAT

2014-02-18 Thread xeon Mailinglist
I am trying to set Hadoop MapReduce (MRv2) behind the NAT, but when I try to connect the Datanode, I get the error below. The hosts have 2 interfaces, one with a private address and another with the NAT address. To access the host with SSH, I must use an external IP, that NAT server will

Unable to load native-hadoop library for your platform

2014-02-12 Thread xeon Mailinglist
I am trying to run an example and I get the following error: HadoopMaster-nh:~# /root/Programs/hadoop/bin/hdfs dfs -count /wiki OpenJDK 64-Bit Server VM warning: You have loaded library /root/Programs/hadoop-2.0.5-alpha/lib/native/libhadoop.so.1.0.0 which might have disabled stack guard. The VM

DisallowedDatanodeException: Datanode denied communication with namenode

2014-02-06 Thread xeon Mailinglist
I am trying to launch the datanodes in Hadoop MRv2, and I get the error below. I looked to Hadoop conf files and the /etc/hosts and everything looks ok. What is wrong in my configuration? org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode denied communication with

org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException

2014-01-08 Thread xeon Mailinglist
When I try to launch the namenode and the datanode in MRv2, the datanode can't connect to the namenode, giving me the error below. I also put the core-site file that I use below. The Firewall in the hosts is disabled. I don't have excluded nodes defined. Why the datanodes can't connect to the

is it possible to list jobs that are waiting to run?

2014-01-03 Thread xeon Mailinglist
Hi, Is it possible that jobs submitted stay waiting before starting to run? Is there a command that list the jobs that are submitted and are waiting to start to run? -- Thanks,

java.net.SocketTimeoutException in the Datanode

2014-01-03 Thread xeon Mailinglist
I am running an wordcount example it MRv2, but I get this error in a Datanode. It looks that it is a problem in the network between the Namenode and the Datanode, but I am not sure. What is this error? How can I fix this problem? 2014-01-03 16:46:29,319 INFO