Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Raj K Singh
in mainframe you can have 3 type of datasources --flat files --VSAM files --DB2/IMS DB2 or IMS supprt the export utilities to copy the data into flat file which you can get through ftp/sftp VSAM file can be exported using IDCAMS utility flat files can be get using the ftp/sftp :::

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Balamurali
Hi, I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . Created table in HBase.Inserted records.Processing the data using Hive. I have to show a graph with some points ( 7 - 7 days or 12 for one year).In one day records may include 1000 - lacks.I need to show average of these 1000 - lacks re

RE: Copy data from Mainframe to HDFS

2013-07-23 Thread Devaraj k
Hi Balamurali, As per my knowledge, there is nothing in the hadoop which does exactly as per your requirement. You can write mapreduce jobs according to your functionality and submit hourly/daily/weekly or monthly . And then you can aggregate the results. If you want some help regarding Hive,

Re: ERROR orm.ClassWriter: Cannot resolve SQL type 1111

2013-07-23 Thread Fatih Haltas
At those columns, I am using uint type. I tried to cast them via sqoop option still it gave the same error. For other columns having type int, text etc, I am able to import them but I have hundreds of data in uint type that I need. While looking at some solutions, I saw that sqoop does not suppo

Join Operation with Regular Expression

2013-07-23 Thread enes yĆ¼cer
Hi, I have 2 data set one of them contain string text, and other table contain string patern (which is searching in text), id. I have create volatile solution, to create two external hive table and full join of them and after full join,I use regex function in where case. but, it takes too long. b

RE: Join Operation with Regular Expression

2013-07-23 Thread Devaraj k
You can try writing the mapreduce job for this. In the Job, you can filter the records in Mapper based on the where condition regex and then perform the join in the Reducer. Please refer the classes present in hadoop-datajoin module to get an idea how to implement the join job. Thanks Devaraj

Re: setting mapred.task.timeout programmatically from client

2013-07-23 Thread Ted Yu
For scheduling mechanism please take a look at oozie. Cheers On Jul 22, 2013, at 10:37 PM, Balamurali wrote: > Hi, > > I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . > Created table in HBase.Inserted records.Processing the data using Hive. > I have to show a graph with some points

Re: setting mapred.task.timeout programmatically from client

2013-07-23 Thread Balamurali
Ok thanks On Tue, Jul 23, 2013 at 3:02 PM, Ted Yu wrote: > For scheduling mechanism please take a look at oozie. > > Cheers > > On Jul 22, 2013, at 10:37 PM, Balamurali wrote: > > Hi, > > I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . > Created table in HBase.Inserted records.Proces

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Sandeep Nemuri
Thanks for your reply guys , i am looking for open source do we have any ?? On Tue, Jul 23, 2013 at 12:53 PM, Devaraj k wrote: > Hi Balamurali, > > ** ** > > As per my knowledge, there is nothing in the hadoop which does exactly as > per your requirement. > > ** ** > > You can write ma

Re: ERROR orm.ClassWriter: Cannot resolve SQL type 1111

2013-07-23 Thread Shahab Yunus
I think you will have to write a custom code to handle this. Regards, Shahab On Tue, Jul 23, 2013 at 3:50 AM, Fatih Haltas wrote: > At those columns, I am using uint type. I tried to cast them via sqoop > option still it gave the same error. > > For other columns having type int, text etc, I

New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hi There, First of all, sorry if I am asking some stupid question. Myself being new to the Hadoop environment , am finding it a bit difficult to figure out why its failing I have installed hadoop 1.2, based on instructions given in the folllowing link http://www.michael-noll.com/tutorials/runnin

Re: copy files from ftp to hdfs in parallel, distcp failed

2013-07-23 Thread Hao Ren
Hi, I am just wondering whether I can move data from Ftp to Hdfs via Hadoop distcp. Can someone give me an example ? In my case, I always encounter the "can not access ftp" error. I am quite sure that the link, login et passwd are correct, actually, I have just copy and paste the ftp addres

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Yexi Jiang
Maybe the conf file is missing or no privilege to access or there is something wrong about the format of your conf files (hdfs-site, core-site, mapred-site). You can double check them. Also probably the typo of the tag or something like that. 2013/7/23 Ashish Umrani > Hi There, > > First of al

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Hi, You might have missed some configuration (XML tags ), Please check all the Conf files. Thanks On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani wrote: > Hi There, > > First of all, sorry if I am asking some stupid question. Myself being new > to the Hadoop environment , am finding it a bit dif

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Jun Ping Du
Hi Sandeep, I think Apache Oozie is something you are looking for. It provide workflow management on Hadoop (and Pig, Hive, etc.) jobs and support continuously run jobs in specific time period. Please refer: http://oozie.apache.org/docs/3.3.2/ for details. Thanks, Junping - Original M

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hey thanks for response. I have changed 4 files during installation core-site.xml mapred-site.xml hdfs-site.xml and hadoop-env.sh I could not find any issues except that all params in the hadoop-env.sh are commented out. Only java_home is un commented. If you have a quick minute can you ple

Re: Incrementally adding to existing output directory

2013-07-23 Thread Max Lebedev
Hi Devaraj, Thanks for the advice. That did the trick. Thanks, Max Lebedev On Wed, Jul 17, 2013 at 10:51 PM, Devaraj k wrote: > It seems, It is not taking the CutomOutputFormat for the Job. You need > to set the custom output format class using the > org.apache.hadoop.mapred.JobConf.setOutp

Only log.index

2013-07-23 Thread Ajay Srivastava
Hi, I see that most of the tasks have only log.index created in /opt/hadoop/logs/userlogs/jobId/task_attempt directory. When does this happen ? Is there a config setting for this OR this is a bug ? Regards, Ajay Srivastava

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread bejoy . hadoop
Hi Ashish In your hdfs-site.xml within tag you need to have the tag and inside a tag you can have , and tags. Regards Bejoy KS Sent from remote device, Please excuse typos -Original Message- From: Ashish Umrani Date: Tue, 23 Jul 2013 09:28:00 To: Reply-To: user@hadoop.apache

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Yexi Jiang
Seems *hdfs-site.xml has no property tag.* 2013/7/23 Ashish Umrani > Hey thanks for response. I have changed 4 files during installation > > core-site.xml > mapred-site.xml > hdfs-site.xml and > hadoop-env.sh > > > I could not find any issues except that all params in the hadoop-env.sh > are

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Hi Ashish, Please check in hdfs-site.xml. It is missing. Thanks. On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani wrote: > Hey thanks for response. I have changed 4 files during installation > > core-site.xml > mapred-site.xml > hdfs-site.xml and > hadoop-env.sh > > > I could not find any is

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks Jitendra, Bejoy and Yexi, I got past that. And now the ls command says it can not access the directory. I am sure this is a permissions issue. I am just wondering which directory and I missing permissions on. Any pointers? And once again, thanks a lot Regards ashish *hduser@ashish-HP

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
Its warning not error... Create a directory and then do ls ( In your case /user/hduser is not created untill and unless for the first time you create a directory or put some file) hadoop fs -mkdir sample hadoop fs -ls I would suggest if you are getting pemission problem, please check the foll

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Mohammad Tariq
Hello Ashish, Change the permissions of /app/hadoop/tmp to 755 and see if it helps. Warm Regards, Tariq cloudfront.blogspot.com On Tue, Jul 23, 2013 at 10:27 PM, Ashish Umrani wrote: > Thanks Jitendra, Bejoy and Yexi, > > I got past that. And now the ls command says it can not access the > di

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hey, Thanks Shekhar. That worked like a chimp. Appreciate help from you all. Now I will try to put files and try the word count or similar program. Regards ashish On Tue, Jul 23, 2013 at 10:07 AM, Shekhar Sharma wrote: > Its warning not error... > > Create a directory and then do ls ( In yo

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks, But the issue was that there was no directory and hence it was not showing anything. Adding a directory cleared the warning. I appreciate your help. Regards ashish On Tue, Jul 23, 2013 at 10:08 AM, Mohammad Tariq wrote: > Hello Ashish, > > Change the permissions of /app/hadoop/tmp t

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
After starting i would suggest always check whether your NameNode and job tracker UI are working or not and check the number of live nodes in both of the UI.. Regards, Som Shekhar Sharma +91-8197243810 On Tue, Jul 23, 2013 at 10:41 PM, Ashish Umrani wrote: > Thanks, > > But the issue was that t

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Try.. *hadoop fs -ls /* ** Thanks On Tue, Jul 23, 2013 at 10:27 PM, Ashish Umrani wrote: > Thanks Jitendra, Bejoy and Yexi, > > I got past that. And now the ls command says it can not access the > directory. I am sure this is a permissions issue. I am just wondering > which directory and I

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Jitendra, Som, Thanks. Issue was in not having any file there. Its working fine now. I am able to do -ls and could also do -mkdir and -put. Now is time to run the jar and apparently I am getting no main manifest attribute, in wc.jar But I believe its because of maven pom file does not have

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
hadoop jar wc.jar inputdata outputdestination Regards, Som Shekhar Sharma +91-8197243810 On Tue, Jul 23, 2013 at 10:58 PM, Ashish Umrani wrote: > Jitendra, Som, > > Thanks. Issue was in not having any file there. Its working fine now. > > I am able to do -ls and could also do -mkdir and -pu

Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Huy Pham
Hi All, Do any of you have or can refer me to some sample Java code that get the tree structure of a HDFS directory, similar to the file system? For example: I have a HDFS dir, called /data, inside data, there is /data/valid and /data/invalid, and so on, so I would need to be able to get th

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks Shekhar, The problem was not in my building of the jar. It was in fact in execution I was running command *hadoop -jar* input output The problem was with -jar. It should be *hadoop jar* input output Thanks for help once again regards ashish On Tue, Jul 23, 2013 at 10:31 AM, S

Re: Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Shahab Yunus
See this https://sites.google.com/site/hadoopandhive/home/how-to-read-all-files-in-a-directory-in-hdfs-using-hadoop-filesystem-api and http://hadoop.apache.org/docs/stable/api/org/apache/hadoop/fs/FileSystem.html#isDirectory(org.apache.hadoop.fs.Path) Basically you can write your own function, p

Saving counters in Mapfile

2013-07-23 Thread Elazar Leibovich
Hi, A common use case one want an ordered structure for, is for saving counters. Naturally, I wanted to save the counters in a Mapfile: for (long ix = 0; ix < MAXVALUE; ix++) { mapfile.append(new Text("counter key of val " + ix), new LongWritable(ix)); } This however looks a bit

Re: Saving counters in Mapfile

2013-07-23 Thread manishbh...@rocketmail.com
Hi, If you intend to use those counters in further functions then I think Hadoop will take care this by itself, you can explore combiner for the same. In Sequence file mapfile and index files has its own functionality and as per my understanding index file expect offset to move over map file and

Re: Saving counters in Mapfile

2013-07-23 Thread Michael Segel
Uhm... You want to save the counters as in counts per job run or something? (Remember HDFS == WORM) Then you could do a sequence file and then use something like HBase to manage the index. (Every time you add a set of counters, you have a new file and a new index.) Heck you could use HBase f

Re: Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Harsh J
The FileSystem interface provides a recursive option for this. See http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FileSystem.html#listFiles(org.apache.hadoop.fs.Path,%20boolean) On Tue, Jul 23, 2013 at 11:35 PM, Huy Pham wrote: > Hi All, >Do any of you have or can refer me to

Re: Only log.index

2013-07-23 Thread Vinod Kumar Vavilapalli
It could either mean that all those task-attempts are crashing before the process itself is getting spawned (check TT logs) or those logs are getting deleted after the fact. Suspect the earlier. Thanks, +Vinod On Jul 23, 2013, at 9:33 AM, Ajay Srivastava wrote: > Hi, > > I see that most of t

Re: Only log.index

2013-07-23 Thread Ajay Srivastava
Hi Vinod, Thanks. It seems that something else is going on - Here is the content of log.index - ajay-srivastava:userlogs ajay.srivastava$ cat job_201307222115_0188/attempt_201307222115_0188_r_00_0/log.index LOG_DIR:/opt/hadoop/bin/../logs/userlogs/job_201307222115_0188/attempt_201307222115_

Re: Only log.index

2013-07-23 Thread Vinod Kumar Vavilapalli
Ah, I should've guessed that. You seem to have JVM reuse enabled. Even if JVMs are reused, all the tasks write to the same files as they share the JVM. They only have different index files. The same thing happens for what we call the TaskCleanup tasks which are launched for failing/killed tasks

Re: Only log.index

2013-07-23 Thread Ajay Srivastava
Yes. That explains it and confirms my guess too :-) stderr:156 0 syslog:995 166247 What are these numbers ? Byte offset in corresponding files from where logs of this task starts. Regards, Ajay Srivastava On 24-Jul-2013, at 12:10 PM, Vinod Kumar Vavilapalli wrote: Ah, I should've guessed