Re: Hadoop 2.6.0 Error

2015-03-25 Thread Azuryy Yu
please also set correct JAVA_HOME in hadoop-env.sh. On Wed, Mar 25, 2015 at 1:53 PM, Anand Murali anand_vi...@yahoo.com wrote: Dear All: Request help/advise as I am unable to start Hadoop. Performed follow steps in Ubuntu 14.10 1. ssh localhost 2. Did following exports in user defined

Re: Passing an array of float type to mapper class

2015-03-18 Thread Azuryy Yu
please take a look at LongWritable.java, you need to implement a similiar class. On Wed, Mar 18, 2015 at 3:43 PM, Vandana kumari kvandana1...@gmail.com wrote: Hello all I want to pass an array of float type to mapper class. How to do that. Any help? -- Thanks and regards Vandana kumari

Re: Unable to see application in http://localhost:8088/cluster/apps

2015-03-17 Thread Azuryy Yu
Did you start JHS(job history server)? On Wed, Mar 18, 2015 at 1:51 AM, Ravi Prakash ravi...@ymail.com wrote: Perhaps yarn.resourcemanager.max-completed-applications ? On Tuesday, March 17, 2015 10:02 AM, hitarth trivedi t.hita...@gmail.com wrote: Hi, When I submit a job to yarn

Re: Snappy Configuration in Hadoop2.5.2

2015-03-16 Thread Azuryy Yu
Hi, please : run cp -a /opt/snappy/lib/libsnappy.* /opt/hadoop/hadoophome/lib/native on each datanode. also you need to install Snappy on each datanode firstly. On Sat, Mar 7, 2015 at 6:57 PM, donhoff_h 165612...@qq.com wrote: Hi, experts. I meet the following problem when configuring

Re: Snappy Configuration in Hadoop2.5.2

2015-03-16 Thread Azuryy Yu
tried JAVA_LIBRARY_PATH and LD_LIBRARY_PATH. But they didn't work. -- Original -- *From: * Azuryy Yu;azury...@gmail.com; *Send time:* Monday, Mar 16, 2015 2:43 PM *To:* user@hadoop.apache.orguser@hadoop.apache.org; *Subject: * Re: Snappy Configuration

Re: Prune out data to a specific reduce task

2015-03-16 Thread Azuryy Yu
Hi, Can you set only one reduce task? why did you want set up two reudce tasks and only one work? On Mon, Mar 16, 2015 at 9:04 AM, Drake민영근 drake@nexr.com wrote: Hi, If you write custom partitioner, just call them to confrim the key match with which partition. You can get the number

Re: File is not written on HDFS after running libhdfs C API

2015-03-05 Thread Azuryy Yu
, 2015 at 10:22 AM, Azuryy Yu azury...@gmail.com wrote: Yes, you should do it:) On Thu, Mar 5, 2015 at 4:17 PM, Alexandru Calin alexandrucali...@gmail.com wrote: Wow, you are so right! it's on the local filesystem! Do I have to manually specify hdfs-site.xml and core-site.xml

Re: File is not written on HDFS after running libhdfs C API

2015-03-05 Thread Azuryy Yu
restarted hadoop. On Thu, Mar 5, 2015 at 10:22 AM, Azuryy Yu azury...@gmail.com wrote: Yes, you should do it:) On Thu, Mar 5, 2015 at 4:17 PM, Alexandru Calin alexandrucali...@gmail.com wrote: Wow, you are so right! it's on the local filesystem! Do I have to manually specify hdfs

Re: File is not written on HDFS after running libhdfs C API

2015-03-05 Thread Azuryy Yu
/local/hadoop/etc/hadoop/core-site.xml ? On Thu, Mar 5, 2015 at 10:04 AM, Azuryy Yu azury...@gmail.com wrote: you need to include core-site.xml as well. and I think you can find '/tmp/testfile.txt' on your local disk, instead of HDFS. if so, My guess is right. because you don't include core

Re: File is not written on HDFS after running libhdfs C API

2015-03-05 Thread Azuryy Yu
-site.xml and yarn-site.xml are those on apache website. A *jps *command shows: 11257 NodeManager 11129 ResourceManager 11815 Jps 10620 NameNode 10966 SecondaryNameNode On Thu, Mar 5, 2015 at 10:48 AM, Azuryy Yu azury...@gmail.com wrote: Can you share your core-site.xml here? On Thu, Mar 5

Re: File is not written on HDFS after running libhdfs C API

2015-03-05 Thread Azuryy Yu
you need to include core-site.xml as well. and I think you can find '/tmp/testfile.txt' on your local disk, instead of HDFS. if so, My guess is right. because you don't include core-site.xml, then your Filesystem schema is file:// by default, not hdfs://. On Thu, Mar 5, 2015 at 3:52 PM,

Re: Need advice about OLAP on Hadoop

2015-03-04 Thread Azuryy Yu
Hi VK, I have a similar requirement. we need a real time data analysis platform. Actually, you don't pay more attention on the Spark or Apache Drill, because data for LOAP cubes was calculated before cube build. you just consider two questions: 1) how to calculate the data for cube quickly?

Re: Can't list files in a federation of HDFS

2015-03-04 Thread Azuryy Yu
For HDFS federation, data share all datanodes, but namespace is separate, so did you write some data on hadoop-coc-2 namespace? you don't need to login hadoop-coc-2 then write data, just config a new client, which connect to hadoop-coc-2 for write. On Tue, Mar 3, 2015 at 6:20 PM,

Re: The Activities of Apache Hadoop Community

2015-03-04 Thread Azuryy Yu
That's good to know, On Tue, Mar 3, 2015 at 8:12 PM, Akira AJISAKA ajisa...@oss.nttdata.co.jp wrote: Hi all, One year after the previous post, we collected and analyzed JIRA tickets again to investigate the activities of Apache Hadoop community in 2014.

Hadoop svn cannot connected.

2015-02-06 Thread Azuryy Yu
Hi, http://svn.apache.org/viewcvs.cgi/hadoop/common/trunk/ I cannot open this URL. does that anybody can access it? another, I cannot svn up the new release in branch-2. It always stay in Aug 2014.

Re: Hadoop svn cannot connected.

2015-02-06 Thread Azuryy Yu
Oh, thanks. On Fri, Feb 6, 2015 at 4:58 PM, Andre Kelpe ake...@concurrentinc.com wrote: Hadoop has moved to git: https://wiki.apache.org/hadoop/GitAndHadoop -- André On Fri, Feb 6, 2015 at 9:13 AM, Azuryy Yu azury...@gmail.com wrote: Hi, http://svn.apache.org/viewcvs.cgi/hadoop/common

Re: How to rolling upgrade??

2015-02-05 Thread Azuryy Yu
please refer to : http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html On Tue, Feb 3, 2015 at 5:38 PM, Mr.J tjst...@kgrid.co.kr wrote: my cluster A, and cluster B. To upgrade to version 2.6 In what order should I upgrade? Journalnode 1 Journalnode

Re: Compilation failing for AuthenticatedURL (no suitable constructor found) Hadoop 2.6.0

2015-02-05 Thread Azuryy Yu
Can you build Hadoop on Linux? or can you check out the whole code from branch-2 again, then build it? On Thu, Feb 5, 2015 at 10:21 AM, Venkat Ramakrishnan venkat.archit...@gmail.com wrote: I need to compile Hadoop, and I am not able to figure out why this error is occuring. Can someone

Re: Which [open-souce] SQL engine atop Hadoop?

2015-02-05 Thread Azuryy Yu
please look at: http://mail-archives.apache.org/mod_mbox/tajo-user/201502.mbox/browser On Tue, Jan 27, 2015 at 5:13 PM, Daniel Haviv danielru...@gmail.com wrote: Can you elaborate on why you prefer Tajo? Daniel On 27 בינו׳ 2015, at 10:35, Azuryy Yu azury...@gmail.com wrote: You almost

Re: How I list files in HDFS?

2015-02-05 Thread Azuryy Yu
Hi, You cannot use new File(...) as parameter, which should be new Path(/outputmp) On Fri, Feb 6, 2015 at 3:51 AM, Ravi Prakash ravi...@ymail.com wrote: Hi Xeon! Can you try using the FileContext or FileSystem API? HTH Ravi On Thursday, February 5, 2015 8:13 AM, xeonmailinglist

Re: MapReduce job is not picking up appended data.

2015-01-27 Thread Azuryy Yu
Are you sure you can 'cat' the lastest batch of the data on HDFS? for Flume, the data is available only after file rolled, because Flume only call FileSystem.close() during file rolling. On Mon, Jan 26, 2015 at 8:17 PM, Uthayan Suthakar uthayan.sutha...@gmail.com wrote: I have a Flume which

Re: Which [open-souce] SQL engine atop Hadoop?

2015-01-27 Thread Azuryy Yu
You almost list all open sourced MPP real time SQL-ON-Hadoop. I prefer Tajo, which was relased by 0.9.0 recently, and still working in progress for 1.0 On Mon, Jan 26, 2015 at 10:19 PM, Samuel Marks samuelma...@gmail.com wrote: Since Hadoop https://hive.apache.org came out, there have been

Re: Time until a datanode is marked as dead

2015-01-26 Thread Azuryy Yu
Hi Frank, can you file an issue to add this configuration to the hdfs-default.xml? On Mon, Jan 26, 2015 at 5:39 PM, Frank Lanitz frank.lan...@sql-ag.de wrote: Hi, Am 23.01.2015 um 19:23 schrieb Chris Nauroth: The time period for determining if a datanode is dead is calculated as a

Re: Multiple separate Hadoop clusters on same physical machines

2015-01-26 Thread Azuryy Yu
Hi, I think the best way is deploy HDFS federation with Hadoop 2.x. On Mon, Jan 26, 2015 at 5:18 PM, Harun Reşit Zafer harun.za...@tubitak.gov.tr wrote: Hi everyone, We have set up and been playing with Hadoop 1.2.x and its friends (Hbase, pig, hive etc.) on 7 physical servers. We want to

Re: Failure to write to HDFS in MapReduce job

2014-11-05 Thread Azuryy Yu
please check your network issue. generally this was casued by unstable network device. On Wed, Nov 5, 2014 at 5:55 PM, Hayden Marchant hayd...@amobee.com wrote: I have a MapReduce job running on Hadoop 2.0.0, and on some 'heavy' jobs, I am seeing the following errors in the reducer.

Re: issue about submit job to local ,not to cluster

2014-10-21 Thread Azuryy Yu
please check your mapred-site.xml is available under conf. On Tue, Oct 21, 2014 at 2:47 PM, ch huang justlo...@gmail.com wrote: hi,maillist: my cluster move from one IDC to another IDC ,when all done ,i run job ,and find the job run on local box not on cluster ,why? it is normal on

Re: HDFS - Consolidate 2 small volumes into 1 large volume

2014-10-21 Thread Azuryy Yu
yes, you can. stop the cluster, change your hdfs-site.xml on your datanode. (dfs.datanode.dir) to the large volume. copy two small data volumes to the large volumes, which was configured on above. start cluster. then you are done. On Tue, Oct 21, 2014 at 9:57 PM, Brian C. Huffman

Re: hadoop 2.4 using Protobuf - How does downgrade back to 2.3 works ?

2014-10-20 Thread Azuryy Yu
after upgrade, there are current an previous subdirs in the namenode metadata directory. so during rollback, just delete current directory and rename previous to current again. On Tue, Oct 21, 2014 at 9:27 AM, Manoj Samel manojsamelt...@gmail.com wrote: Is the pre-upgrade metadata also kept

Re: how to copy data between two hdfs cluster fastly?

2014-10-17 Thread Azuryy Yu
Did you specified how many map tasks? On Fri, Oct 17, 2014 at 4:58 PM, ch huang justlo...@gmail.com wrote: hi,maillist: i now use distcp to migrate data from CDH4.4 to CDH5.1 , i find when copy small file,it very good, but when transfer big data ,it very slow ,any good method

Re: hadoop 2.4 using Protobuf - How does downgrade back to 2.3 works ?

2014-10-17 Thread Azuryy Yu
just stop your cluster, then start your HDFS with '-rollback'. but it's only if you don't finalize HDFS upgrade using command line. On Fri, Oct 17, 2014 at 8:15 AM, Manoj Samel manojsamelt...@gmail.com wrote: Hadoop 2.4.0 mentions that FSImage is stored using protobuf. So upgrade from 2.3.0 to

Re: C++ development framework under Hadoop

2014-10-14 Thread Azuryy Yu
Hadoop streaming is the best option for you. It doesn't has high I/O overhead if you don't add a high I/O in your c++ code. hadoop streaming use buidin MapReduce, it just redirect input/out stream for your c++ application. On Tue, Oct 14, 2014 at 10:33 PM, Y. Z. zhaoyansw...@gmail.com wrote:

Re: C++ development framework under Hadoop

2014-10-14 Thread Azuryy Yu
yes. it always supports hadoop pipe in v2. On Wed, Oct 15, 2014 at 11:33 AM, Y Z zhaoyansw...@gmail.com wrote: Thanks, Azuryy! I found some examples about Pipes. Is Hadoop Pipes still support in Hadoop 2.2? Sincerely, Yongan On 10/14/2014 11:20 PM, Azuryy Yu wrote: Hadoop streaming

Re: ETL using Hadoop

2014-10-08 Thread Azuryy Yu
Hi Moin, Yes, you can replace your ETL using hadoop. but it would be a big change. such as data collection, pre-process, ETL tasks rewrite etc. I don't think there is .NET API in Hadoop. On Wed, Oct 8, 2014 at 2:31 PM, Dattatrya Moin dattatryam...@gmail.com wrote: Hi , We have our own ETL

Re: Datanode volume full, but not moving to free volume

2014-10-08 Thread Azuryy Yu
Hi Brian, Did you try set dfs.datanode.fsdataset.volume.choosing.policy to org.apache.hadoop.hdfs.server.datanode.fsdataset.AvailableSpaceVolumeChoosingPolicy then there are some other related options with this policy. you can google it. On Wed, Oct 8, 2014 at 4:44 PM, Aitor Cedres

Re: Datanode volume full, but not moving to free volume

2014-10-08 Thread Azuryy Yu
yes, all blockpool directory names are all same is expected if you don't configure HDFS federation. On Wed, Oct 8, 2014 at 7:39 PM, Brian C. Huffman bhuff...@etinternational.com wrote: Hmmm.. It seems that there's only one block pool per disk. So that won't help me. :-( Also, I see the

Re: Datanode disk considerations

2014-10-07 Thread Azuryy Yu
I think Brian gave the answer. On Tue, Oct 7, 2014 at 9:13 PM, Brian C. Huffman bhuff...@etinternational.com wrote: What about setting the dfs.datanode.fsdataset.volume.choosing.policy to org.apache.hadoop.hdfs.server. datanode.fsdataset. AvailableSpaceVolumeChoosingPolicy? Would that

Re: Why I cannot delete all the nameNode metadata?

2014-10-07 Thread Azuryy Yu
To make sure your dfs.namenode.name.dir is by default. then, how did you find /user exists? hdfs dfs -ls ? or you checked dfs.datanode.data.dir? if the latter, then don't worry. On Wed, Oct 8, 2014 at 11:56 AM, Tianyin Xu t...@cs.ucsd.edu wrote: Hi, I wanna run some experiments on Hadoop

Re: Realtime sensor's tcpip data to hadoop

2014-05-13 Thread Azuryy Yu
Hi Alex, you can try Apache Flume. On Wed, May 7, 2014 at 10:48 AM, Alex Lee eliy...@hotmail.com wrote: Sensors' may send tcpip data to server. Each sensor may send tcpip data like a stream to the server, the quatity of the sensors and the data rate of the data is high. Firstly, how the

Re: Change Yarn webUI time format

2014-04-29 Thread Azuryy Yu
resolved, I've filed a JIRA for this. https://issues.apache.org/jira/browse/YARN-1998 On Tue, Apr 29, 2014 at 6:46 PM, Azuryy Yu azury...@gmail.com wrote: Hi, how to change the time zone of startTime and finishTime on the yarn web ui? I cannot find the code, I just found render() returns

configure HBase

2014-04-24 Thread Azuryy Yu
Hi, what's mean of HBASE_REGIONSERVER_MLOCK? I cannot find the doucment for it. there is only Uncomment and adjust to keep all the Region Server pages mapped to be memory resident in the hbase-env.sh, can you explain in detail? Thanks for any inputs.

Re: configure HBase

2014-04-24 Thread Azuryy Yu
On Thu, Apr 24, 2014 at 5:46 PM, Ted Yu yuzhih...@gmail.com wrote: Please take a look at https://issues.apache.org/jira/browse/HBASE-6567 Cheers On Apr 24, 2014, at 3:33 AM, Azuryy Yu azury...@gmail.com wrote: Hi, what's mean of HBASE_REGIONSERVER_MLOCK? I cannot find the doucment

Re: Problem with jobtracker hadoop 1.2

2014-04-17 Thread Azuryy Yu
There is because your HDFS has no space left. please check your datanodes are all started. also please check dfs.datanode.du.reserved in hdfs-site.xml to make sure you don't reserve large capacity. On Fri, Apr 18, 2014 at 7:42 AM, Shengjun Xin s...@gopivotal.com wrote: Did you start datanode

Re: hadoop 2.4.0?

2014-04-17 Thread Azuryy Yu
Hadoop 2.4.0 doesn't has the known issue now. I think it's a stable release even if it's not in the stable download list. the only one issue I met is that you should upgrade Hive to Hive-0.12.0 after upgrade to 2.4.0 for the API compatible. On Fri, Apr 18, 2014 at 1:07 AM, MrAsanjar .

Re: JVM option

2014-04-17 Thread Azuryy Yu
Do you want add -Xmx4g to your MR tasks? if so, just add it as mapred.child.java.opts in the mapred-site.xml On Fri, Apr 18, 2014 at 9:35 AM, Andy Srine andy.sr...@gmail.com wrote: Quick question. How would I pass the following JVM option to the Hadoop command line? -Xmx4G hadoop jar jar

Re: Hadoop NoClassDefFoundError

2014-04-15 Thread Azuryy Yu
Please use: hadoop jar myjob.jar myjob.MyJob input output On Tue, Apr 15, 2014 at 3:06 PM, laozh...@sina.cn laozh...@sina.cn wrote: Hello EveryOne: I am new to hadoop,and i am reading Hadoop in action. When i tried to run a demo from this book,I got a problem and could not find answer

Re: Hadoop NoClassDefFoundError

2014-04-15 Thread Azuryy Yu
Please use: hadoop jar myjob.jar myjob.MyJob input output On Tue, Apr 15, 2014 at 3:06 PM, laozh...@sina.cn laozh...@sina.cn wrote: Hello EveryOne: I am new to hadoop,and i am reading Hadoop in action. When i tried to run a demo from this book,I got a problem and could not find answer

Re: Container states trantition questions

2014-04-03 Thread Azuryy Yu
. all mapper ends with 143. I've no idea what it means 2014-04-03 8:45 GMT+08:00 Azuryy Yu azury...@gmail.com: Hi, Does it normal for each container end with TERMINATED(143) ? The whole MR job is successful, but all containers in the map phase end with 143. There are no any useful logs

Re: How to open .gv file for Yarn event model

2014-04-03 Thread Azuryy Yu
Thanks Ravi, I am using Graphviz as Jeff said. it's enough to me. On Thu, Apr 3, 2014 at 5:12 AM, Ravi Prakash ravi...@ymail.com wrote: Hi Azuryy! You have to use dot to convert it to png On Tuesday, April 1, 2014 6:38 PM, Azuryy Yu azury...@gmail.com wrote: Hi, I compiled Yarn

Re: Value for yarn.nodemanager.address in configuration file

2014-04-03 Thread Azuryy Yu
Hi Krishna, Don't worrry about that, there is no ssh 0.0.0.0 during start NM, it's a local service. On Fri, Apr 4, 2014 at 9:12 AM, Shengjun Xin s...@gopivotal.com wrote: you can use yarn-daemon.sh to start nodemanager without ssh On Thu, Apr 3, 2014 at 10:36 PM, Krishna Kishore Bonagiri

Fwd: Container states trantition questions

2014-04-02 Thread Azuryy Yu
Hi, Does it normal for each container end with TERMINATED(143) ? The whole MR job is successful, but all containers in the map phase end with 143. There are no any useful logs in the NM, AM, Container logs. Another minor question: There are only WARN logs in the stderr: log4j:WARN No appenders

Re: patch format

2014-04-01 Thread Azuryy Yu
There is no special patch format, just using svn diff or git diff to generate the patch, which depeds on your code repository. On Tue, Apr 1, 2014 at 5:14 PM, Avinash Kujur avin...@gmail.com wrote: hello everyone, Is there any kind of specific format of the patch in which we need to write

Re: patch format

2014-04-01 Thread Azuryy Yu
There is no special patch format, just using svn diff or git diff to generate the patch, which depeds on your code repository. On Tue, Apr 1, 2014 at 5:14 PM, Avinash Kujur avin...@gmail.com wrote: hello everyone, Is there any kind of specific format of the patch in which we need to write

How to open .gv file for Yarn event model

2014-04-01 Thread Azuryy Yu
Hi, I compiled Yarn event model using maven, but how to open .gv file to view it? Thanks.

Re: job not shown in job history server

2014-03-31 Thread Azuryy Yu
Did you configure mapreduce.jobhistory.address and mapreduce.jobhistory.webapp.address in the mapred-site.xml? On Mon, Mar 31, 2014 at 3:10 PM, Rahul Singh smart.rahul.i...@gmail.comwrote: Hi, I have configured hadoop 2.3 single node setup and ran a sample job using hadoop command,

Re: HADOOP_MAPRED_HOME not found!

2014-03-28 Thread Azuryy Yu
it was defined at hadoop-config.sh On Fri, Mar 28, 2014 at 1:19 PM, divye sheth divs.sh...@gmail.com wrote: Which version of hadoop are u using? AFAIK the hadoop mapred home is the directory where hadoop is installed or in other words untarred. Thanks Divye Sheth On Mar 28, 2014 10:43

Re: how to be assignee ?

2014-03-28 Thread Azuryy Yu
Hi Avin, You should be added as an sub-project's contributor, then you can be an assignee. so you can find how to be an contributor on the Wiki. On Fri, Mar 28, 2014 at 6:50 PM, Avinash Kujur avin...@gmail.com wrote: hi, how can i be assignee fro a particular issue? i can't see any option

Re: namenode could not start

2014-03-26 Thread Azuryy Yu
start-dfs.. *Command for namenode format:* hdfs namenode -format Thanks Regards Brahma Reddy Battula -- *From:* Azuryy Yu [azury...@gmail.com] *Sent:* Tuesday, March 25, 2014 2:39 PM *To:* user@hadoop.apache.org *Subject:* Re: namenode could

Re: Decommissioning a node takes forever

2014-03-26 Thread Azuryy Yu
Hi, which version HDFS you used? On Wed, Mar 26, 2014 at 3:17 PM, Bharath Kumar bharath...@gmail.com wrote: Hi All, I am a novice hadoop user . I tried removing a node from my cluster of 2 nodes by adding the ip in excludes file and running dfsadmin -refreshNodes command . But

Re: namenode could not start

2014-03-25 Thread Azuryy Yu
Are you sure /home/hadoop/mydata/hdfs/namenode exists and has a right permission? On Tue, Mar 25, 2014 at 4:51 PM, haihong lu ung3...@gmail.com wrote: Dear all: I had a problem that the namenode could not start when i ran start-dfs.sh . Show the message as below: FATAL

Re: Hadoop versions

2014-03-24 Thread Azuryy Yu
Hi, I do think Hadoop-0.2x is not similar to 2.0. Hadoop version is composed of major-version.minor-version.released such as: Hadoop-2.0.5, major version is 2, 0 is minor versioin. 5 the the fifth release on 2.0 currently, Hadoop-1.2 and Hadoop-2.2 are all stable, but there are big difference

Re: hadoop source code

2014-03-24 Thread Azuryy Yu
Hi, you'd bettern check out from svn, then build with maven: mvn eclipse:eclipse, which will generate eclipse projects to you. On Mon, Mar 24, 2014 at 3:24 PM, Avinash Kujur avin...@gmail.com wrote: hi, i downloaded the hadoop source code from github. after importing those files in eclipse

Re: how does hdfs client write a file .

2014-03-24 Thread Azuryy Yu
Can you give the URL where you looked at? On Mon, Mar 24, 2014 at 4:49 PM, 张建伟 connerlyzh...@gmail.com wrote: hi, In the apache site,I find a document said : A client request to create a file does not reach the NameNode immediately. In fact, initially the HDFS client caches the file data

Re: how does hdfs client write a file .

2014-03-24 Thread Azuryy Yu
the HDFS client caches the file data into a temporary local file. Application writes are transparently redirected to this temporary local file. 2014-03-24 17:01 GMT+08:00 Azuryy Yu azury...@gmail.com: Can you give the URL where you looked at? On Mon, Mar 24, 2014 at 4:49 PM, 张建伟

Re: lot of attempt_local296445216_0001_m_000386_0 dir in NN dir

2014-03-24 Thread Azuryy Yu
This is a known issue, but not a bug. one safe way is remove files older than sevral hours under these dirs. On Tue, Mar 25, 2014 at 8:57 AM, ch huang justlo...@gmail.com wrote: hi,maillist: i find many dirs in

Re: Hadoop dfs upgrade fail when HA enabled

2014-03-21 Thread Azuryy Yu
It'll be supported in 2.4. please look at here: https://issues.apache.org/jira/browse/HDFS-5138 On Fri, Mar 21, 2014 at 3:46 PM, Meng QingPing mqingp...@gmail.com wrote: Hi, Hadoop dfs upgrade fail when HA enabled. Can Hadoop add feature to upgrade dfs based on HA configure automatically

Re: I am about to lose all my data please help

2014-03-18 Thread Azuryy Yu
property namehadoop.tmp.dir/name value/home/hadoop/project/hadoop-data/value /property On Tue, Mar 18, 2014 at 2:06 PM, Azuryy Yu azury...@gmail.com wrote: I don't think this is the case, because there is; property namehadoop.tmp.dir/name value/home/hadoop/project

Re: I am about to lose all my data please help

2014-03-18 Thread Azuryy Yu
I don't think this is the case, because there is; property namehadoop.tmp.dir/name value/home/hadoop/project/hadoop-data/value /property On Tue, Mar 18, 2014 at 1:55 PM, Stanley Shi s...@gopivotal.com wrote: one possible reason is that you didn't set the namenode working directory,

Re: Question on DFS Balancing

2014-03-05 Thread Azuryy Yu
should take while doing this? Data loss is a NO NO for me. Thanks Divye Sheth On Wed, Mar 5, 2014 at 1:28 PM, Azuryy Yu azury...@gmail.com wrote: Hi, That probably break something if you apply the patch from 2.x to 0.20.x, but it depends on. AFAIK, Balancer had a major refactor in HDFSv2

Re: Question on DFS Balancing

2014-03-05 Thread Azuryy Yu
precautions that I should take? Thanks Azuryy. On Wed, Mar 5, 2014 at 2:19 PM, Azuryy Yu azury...@gmail.com wrote: you can write a simple tool to move blocks peer to peer. I had such tool before, but I cannot find it now. background: our cluster is not balanced, load balancer is very slow, so i

Re: The best practice of migrating hadoop 1.0.1 to hadoop 2.2.3

2014-03-05 Thread Azuryy Yu
Hi, 1) Is it possible to do an in-place migration, while keeping all data in HDFS safely? yes. stop the HDFS firstly, then run start-dfs.sh -upgrade 2) If it is yes, is there any doc/guidance to do this? you just want a HDFS upgrade, so I don't think there are some useful doc. 3)

Re: Question on DFS Balancing

2014-03-04 Thread Azuryy Yu
Hi, That probably break something if you apply the patch from 2.x to 0.20.x, but it depends on. AFAIK, Balancer had a major refactor in HDFSv2, so you'd better fix it by yourself based on HDFS-1804. On Wed, Mar 5, 2014 at 3:47 PM, divye sheth divs.sh...@gmail.com wrote: Thanks Harsh. The

Re: Unable to export hadoop trunk into eclipse

2014-03-03 Thread Azuryy Yu
Have you installed protobuf on your computer? https://code.google.com/p/protobuf/downloads/list On Tue, Mar 4, 2014 at 3:08 PM, nagarjuna kanamarlapudi nagarjuna.kanamarlap...@gmail.com wrote: Hi Ted, I didn't do that earlier. Now , I did it mvn:eclipse eclipse and tried importing

Re: hadoop 2.2.0 cluster setup error : could only be replicated to 0 nodes instead of minReplication (=1)

2014-02-24 Thread Azuryy Yu
Generally, this is caused by insufficient space. please check the total capacity of your cluster and used, remaining ratio, and check dfs.datanode.du.reserved in the hdfs-site.xml if this value is larger than your remained capacity, then you got this Exception. On Tue, Feb 25, 2014 at 10:35

Re: Having trouble adding external JAR to MapReduce Program

2014-02-21 Thread Azuryy Yu
Hi, you cannot add jar like this way. please look at DistributeCache in the Hadoop Java Doc. please call DistributeCache.addArchive() in your main Class before submit the MR job. On Sat, Feb 22, 2014 at 9:30 AM, Gaurav Gupta gau...@datatorrent.comwrote: Jonathan, You have to make sure

Re: Reduce task hang[EMERGENCE]

2014-01-03 Thread Azuryy Yu
. On Fri, Jan 3, 2014 at 1:59 PM, Harsh J ha...@cloudera.com wrote: Hi again, I did not mention the TaskTracker log, but the actual Task JVM's syslog/stderr/stdout. Do those log *any* problems at all? On Fri, Jan 3, 2014 at 11:16 AM, Azuryy Yu azury...@gmail.com wrote: Hi Harsh, Thanks

Re: Reduce task hang[EMERGENCE]

2014-01-03 Thread Azuryy Yu
actuall these are all logs in the stderr, and stdout is empty On Fri, Jan 3, 2014 at 4:12 PM, Azuryy Yu azury...@gmail.com wrote: Hi Harsh, There is only three warnnings in stderr *stderr logs* log4j:WARN No appenders could be found for logger (org.apache.hadoop.mapred.Child

Reduce task hang[EMERGENCE]

2014-01-02 Thread Azuryy Yu
Hi, Our prod cluster met some issues recently, All map tasks finished successfully, but reduce task hanged. but It's not happened on all TaskTrackers, only sometimes. we used mapred-1.0.4 There is 0.0% reduce copy forever until kill task manually. reduce logs on the TaskTracker:

Re: Reduce task hang[EMERGENCE]

2014-01-02 Thread Azuryy Yu
Add addtional: Our MR version is 1.2.1, not 1.0.4 There is no useful information in the JT log. On Fri, Jan 3, 2014 at 12:20 PM, Azuryy Yu azury...@gmail.com wrote: Hi, Our prod cluster met some issues recently, All map tasks finished successfully, but reduce task hanged. but It's

Re: Reduce task hang[EMERGENCE]

2014-01-02 Thread Azuryy Yu
TaskTrackers in trying to obtain the data? On Fri, Jan 3, 2014 at 9:54 AM, Azuryy Yu azury...@gmail.com wrote: Add addtional: Our MR version is 1.2.1, not 1.0.4 There is no useful information in the JT log. On Fri, Jan 3, 2014 at 12:20 PM, Azuryy Yu azury...@gmail.com wrote

Re: Reduce task hang[EMERGENCE]

2014-01-02 Thread Azuryy Yu
In detail: 'and these people's job never hanged...' these people's map and reduce tasks never hanged. On Fri, Jan 3, 2014 at 1:46 PM, Azuryy Yu azury...@gmail.com wrote: Hi Harsh, Thanks. There is no any error logs for attempt_201312201200_34795_r_00_0 in the tasktracker log. only

Re: XML to TEXT

2014-01-02 Thread Azuryy Yu
Hi, you can use org.apache.hadoop.streaming.StreamInputFormat using map reduce to convert XML to text. such as your xml like this: xml namelll/name /xml you need to specify stream.recordreader.begin and stream.recordreader.end in the Configuration: Configuration conf = new Configuration();

Re: any suggestions on IIS log storage and analysis?

2013-12-30 Thread Azuryy Yu
You can run a mapreduce firstly, Join these data sets into one data set. then analyze the joined dataset. On Mon, Dec 30, 2013 at 3:58 PM, Fengyun RAO raofeng...@gmail.com wrote: Hi, HDFS splits files into blocks, and mapreduce runs a map task for each block. However, Fields could be

Re: get job data in command line in MRv2

2013-12-30 Thread Azuryy Yu
Generally, MRv2 indicates Yarn. you can try: yarn application then there are full help lists. On Tue, Dec 31, 2013 at 12:32 PM, 小网客 smallnetvisi...@foxmail.com wrote: ui or hadoop job command like:hadoop job -list -- -

Re: Estimating the time of my hadoop jobs

2013-12-17 Thread Azuryy Yu
Hi Kandoi, It depends on: how many cores on each VNode how complicated of your analysis application But I don't think it's normal spent 3hr to process 30GB data even on your *not good* hareware. On Tue, Dec 17, 2013 at 6:39 PM, Kandoi, Nikhil nikhil.kan...@emc.comwrote: Hello everyone,

Re: How to set hadoop.tmp.dir if I have multiple disks per node?

2013-12-16 Thread Azuryy Yu
Hi Tao, No, you need to set mapred.local.dir in the mapred-site.xml with comma separated list of path to spread I/O . On Mon, Dec 16, 2013 at 5:11 PM, Tao Xiao xiaotao.cs@gmail.com wrote: Thanks. In order to spread I/O among multiple disks, should I assign a comma-separated list of

Re: How does mapreduce job determine the compress codec

2013-12-15 Thread Azuryy Yu
Hi Jiayu, For the Sequence file as an input, CompressCodec class was serialized in the file header, then Sequence Filereader will know the compression algo. thanks. On Mon, Dec 16, 2013 at 8:28 AM, Jiayu Ji jiayu...@gmail.com wrote: Thanks Tao. I know I can tell it is a lzo file based on the

Re: Site-specific dfs.client.local.interfaces setting not respected for Yarn MR container

2013-12-15 Thread Azuryy Yu
Jeff, DFSClient don't use copied Configuration from RM. did you add hostname or IP addr in the conf/slaves? if hostname, Can you check /etc/hosts? does there have confilicts? and y On Mon, Dec 16, 2013 at 5:01 AM, Jeff Stuckman stuck...@umd.edu wrote: Thanks for the response. I have the

Re: how to create symbolic link in hdfs with c++ code or webhdfs interface?

2013-12-15 Thread Azuryy Yu
Why enable test flag under non-test environment? On Mon, Dec 16, 2013 at 11:10 AM, Xiaobin She xiaobin...@gmail.com wrote: To Vinod, thank you for your reply, I found that I got this error message because I did not add the following property in the hdfs-site.xml file 28 property 29

Re: hadoop 2.2 build

2013-12-13 Thread Azuryy Yu
Can you check the datanode log to find what error happended? On Sat, Dec 14, 2013 at 9:27 AM, Karim Awara karim.aw...@kaust.edu.sawrote: Hi, I was setting up hadoop 2.2 following thesehttp://raseshmori.wordpress.com/2012/09/23/install-hadoop-2-0-1-yarn-nextgen/instructions on a single

Re: Write Access Denied in hdfs

2013-12-03 Thread Azuryy Yu
I cannot open you shared URL. but if you write or create file by user A under a HDFS path , which is not created by user A, then it's permission failed. one solution is 'hdfs fs -chown -R A:A your_path' , another is 'hdfs fs -chmod -R 777 your_path' On Tue, Dec 3, 2013 at 5:09 PM, unmesha

Re: how to prevent JAVA HEAP OOM happen in shuffle process in a MR job?

2013-12-03 Thread Azuryy Yu
Hi Jay, Can you point out a detail about 1.x that job tracker having an ever-expanding hash map. I want fix it. On Tue, Dec 3, 2013 at 3:08 AM, Jay Vyas jayunit...@gmail.com wrote: version is rewally important here.. - If 1.x, then Where (NN , JT , TT ?) - if 2.x, then where? (AM, NM, ...

Re: Perfect configuration setting

2013-12-03 Thread Azuryy Yu
Hi Sandy, I think there are some permission limited on your shared URL. maybe only Cloudera employees allowed. On Tue, Dec 3, 2013 at 4:26 PM, Sandy Ryza sandy.r...@cloudera.com wrote: Hi Geelong, Check out Todd Lipcon's presentation on tuning MapReduce performance:

Re: Perfect configuration setting

2013-12-03 Thread Azuryy Yu
, Dec 3, 2013 at 5:46 PM, Azuryy Yu azury...@gmail.com wrote: Hi Sandy, I think there are some permission limited on your shared URL. maybe only Cloudera employees allowed. On Tue, Dec 3, 2013 at 4:26 PM, Sandy Ryza sandy.r...@cloudera.comwrote: Hi Geelong, Check out Todd Lipcon's

Re: issue about read file from HDFS

2013-12-03 Thread Azuryy Yu
One suggestion is change your hive partition, add a hive partition every five minutes, and your HDFS file also roller every five minutes. On Wed, Dec 4, 2013 at 11:56 AM, ch huang justlo...@gmail.com wrote: hi,mailist: when HDFS 's file is in appending ,no other reader can get

Re: issue about read file from HDFS

2013-12-03 Thread Azuryy Yu
Ok, so you go for yourself. On Wed, Dec 4, 2013 at 12:15 PM, ch huang justlo...@gmail.com wrote: it seems not a good suggestion,get lot of partition dir and data file will be a big compact to NN On Wed, Dec 4, 2013 at 12:08 PM, Azuryy Yu azury...@gmail.com wrote: One suggestion is change

Re: Can not auto-failover when unplug network interface

2013-12-02 Thread Azuryy Yu
This is still because your fence method configuraed improperly. plseae paste your fence configuration. and double check you can ssh on active NN to standby NN without password. On Tue, Dec 3, 2013 at 10:23 AM, YouPeng Yang yypvsxf19870...@gmail.comwrote: Hi Another auto-failover testing

Re: Error for larger jobs

2013-11-27 Thread Azuryy Yu
Your disk is full from the log. On 2013-11-28 5:27 AM, Siddharth Tiwari siddharth.tiw...@live.com wrote: Hi Team I am getting following strange error, can you point me to the possible reason. I have set heap size to 4GB but still getting it. please help *syslog logs* 2013-11-27

Re: Error for larger jobs

2013-11-27 Thread Azuryy Yu
Siddharth, please check 'mapred.local.dir', but I would like advice you check GC logs and OS logs. pay more attention on OS logs. I suspect you start too many threads concurrently, then consumed all OS avaliable resources. On Thu, Nov 28, 2013 at 9:08 AM, Vinayakumar B

Re: Error for larger jobs

2013-11-27 Thread Azuryy Yu
yes. you need to increase it, a simple way is put it in your /etc/profile On Thu, Nov 28, 2013 at 9:59 AM, Siddharth Tiwari siddharth.tiw...@live.com wrote: Hi Vinay and Azuryy Thanks for your responses. I get these error when I just run a teragen. Also, do you suggest me to increase

Re: Heterogeneous Cluster

2013-11-25 Thread Azuryy Yu
I don't think this is a normal way, and It's not suggested. we can deploy cluster cross IDC, cross different network, but don't cross OS. at least currently. On Tue, Nov 26, 2013 at 6:56 AM, Andrew Machtolff amachto...@askcts.comwrote: Yes, I set one up as a test. I had a windows cluster of

  1   2   3   >