Copy data from Mainframe to HDFS

2013-07-23 Thread Sandeep Nemuri
Hi , How to copy datasets from Mainframe to HDFS directly? I know that we can NDM files to Linux box and then we can use simple put command to copy data to HDFS. But, how to copy data directly from mainframe to HDFS? I have PS, PDS and VSAM datasets to copy to HDFS for analysis using

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Mohammad Tariq
Hello Sandeep, You don't have to convert the data in order to copy it into the HDFS. But you might have to think about the MR processing of these files because of the format of these files. You could probably make use of Sqoop http://sqoop.apache.org/. I also came across DMX-H a few days ago

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Raj K Singh
in mainframe you can have 3 type of datasources --flat files --VSAM files --DB2/IMS DB2 or IMS supprt the export utilities to copy the data into flat file which you can get through ftp/sftp VSAM file can be exported using IDCAMS utility flat files can be get using the ftp/sftp

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Balamurali
Hi, I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . Created table in HBase.Inserted records.Processing the data using Hive. I have to show a graph with some points ( 7 - 7 days or 12 for one year).In one day records may include 1000 - lacks.I need to show average of these 1000 - lacks

RE: Copy data from Mainframe to HDFS

2013-07-23 Thread Devaraj k
Hi Balamurali, As per my knowledge, there is nothing in the hadoop which does exactly as per your requirement. You can write mapreduce jobs according to your functionality and submit hourly/daily/weekly or monthly . And then you can aggregate the results. If you want some help regarding Hive,

Re: ERROR orm.ClassWriter: Cannot resolve SQL type 1111

2013-07-23 Thread Fatih Haltas
At those columns, I am using uint type. I tried to cast them via sqoop option still it gave the same error. For other columns having type int, text etc, I am able to import them but I have hundreds of data in uint type that I need. While looking at some solutions, I saw that sqoop does not

Join Operation with Regular Expression

2013-07-23 Thread enes yĆ¼cer
Hi, I have 2 data set one of them contain string text, and other table contain string patern (which is searching in text), id. I have create volatile solution, to create two external hive table and full join of them and after full join,I use regex function in where case. but, it takes too long.

RE: Join Operation with Regular Expression

2013-07-23 Thread Devaraj k
You can try writing the mapreduce job for this. In the Job, you can filter the records in Mapper based on the where condition regex and then perform the join in the Reducer. Please refer the classes present in hadoop-datajoin module to get an idea how to implement the join job. Thanks Devaraj

Re: setting mapred.task.timeout programmatically from client

2013-07-23 Thread Ted Yu
For scheduling mechanism please take a look at oozie. Cheers On Jul 22, 2013, at 10:37 PM, Balamurali balamurali...@gmail.com wrote: Hi, I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . Created table in HBase.Inserted records.Processing the data using Hive. I have to show a

Re: setting mapred.task.timeout programmatically from client

2013-07-23 Thread Balamurali
Ok thanks On Tue, Jul 23, 2013 at 3:02 PM, Ted Yu yuzhih...@gmail.com wrote: For scheduling mechanism please take a look at oozie. Cheers On Jul 22, 2013, at 10:37 PM, Balamurali balamurali...@gmail.com wrote: Hi, I configured hadoop-1.0.3, hbase-0.92.1 and hive-0.10.0 . Created table

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Sandeep Nemuri
Thanks for your reply guys , i am looking for open source do we have any ?? On Tue, Jul 23, 2013 at 12:53 PM, Devaraj k devara...@huawei.com wrote: Hi Balamurali, ** ** As per my knowledge, there is nothing in the hadoop which does exactly as per your requirement. ** ** You

Re: ERROR orm.ClassWriter: Cannot resolve SQL type 1111

2013-07-23 Thread Shahab Yunus
I think you will have to write a custom code to handle this. Regards, Shahab On Tue, Jul 23, 2013 at 3:50 AM, Fatih Haltas fatih.hal...@nyu.edu wrote: At those columns, I am using uint type. I tried to cast them via sqoop option still it gave the same error. For other columns having type

New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hi There, First of all, sorry if I am asking some stupid question. Myself being new to the Hadoop environment , am finding it a bit difficult to figure out why its failing I have installed hadoop 1.2, based on instructions given in the folllowing link

Re: copy files from ftp to hdfs in parallel, distcp failed

2013-07-23 Thread Hao Ren
Hi, I am just wondering whether I can move data from Ftp to Hdfs via Hadoop distcp. Can someone give me an example ? In my case, I always encounter the can not access ftp error. I am quite sure that the link, login et passwd are correct, actually, I have just copy and paste the ftp address

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Yexi Jiang
Maybe the conf file is missing or no privilege to access or there is something wrong about the format of your conf files (hdfs-site, core-site, mapred-site). You can double check them. Also probably the typo of the property/property tag or something like that. 2013/7/23 Ashish Umrani

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Hi, You might have missed some configuration (XML tags ), Please check all the Conf files. Thanks On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Hi There, First of all, sorry if I am asking some stupid question. Myself being new to the Hadoop environment , am

Re: Copy data from Mainframe to HDFS

2013-07-23 Thread Jun Ping Du
Hi Sandeep, I think Apache Oozie is something you are looking for. It provide workflow management on Hadoop (and Pig, Hive, etc.) jobs and support continuously run jobs in specific time period. Please refer: http://oozie.apache.org/docs/3.3.2/ for details. Thanks, Junping - Original

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hey thanks for response. I have changed 4 files during installation core-site.xml mapred-site.xml hdfs-site.xml and hadoop-env.sh I could not find any issues except that all params in the hadoop-env.sh are commented out. Only java_home is un commented. If you have a quick minute can you

Re: Incrementally adding to existing output directory

2013-07-23 Thread Max Lebedev
Hi Devaraj, Thanks for the advice. That did the trick. Thanks, Max Lebedev On Wed, Jul 17, 2013 at 10:51 PM, Devaraj k devara...@huawei.com wrote: It seems, It is not taking the CutomOutputFormat for the Job. You need to set the custom output format class using the

Only log.index

2013-07-23 Thread Ajay Srivastava
Hi, I see that most of the tasks have only log.index created in /opt/hadoop/logs/userlogs/jobId/task_attempt directory. When does this happen ? Is there a config setting for this OR this is a bug ? Regards, Ajay Srivastava

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread bejoy . hadoop
Hi Ashish In your hdfs-site.xml within configuration tag you need to have the property tag and inside a property tag you can have name,value and description tags. Regards Bejoy KS Sent from remote device, Please excuse typos -Original Message- From: Ashish Umrani

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Yexi Jiang
Seems *hdfs-site.xml has no property tag.* 2013/7/23 Ashish Umrani ashish.umr...@gmail.com Hey thanks for response. I have changed 4 files during installation core-site.xml mapred-site.xml hdfs-site.xml and hadoop-env.sh I could not find any issues except that all params in the

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Hi Ashish, Please check property/property in hdfs-site.xml. It is missing. Thanks. On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Hey thanks for response. I have changed 4 files during installation core-site.xml mapred-site.xml hdfs-site.xml and

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks Jitendra, Bejoy and Yexi, I got past that. And now the ls command says it can not access the directory. I am sure this is a permissions issue. I am just wondering which directory and I missing permissions on. Any pointers? And once again, thanks a lot Regards ashish

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
Its warning not error... Create a directory and then do ls ( In your case /user/hduser is not created untill and unless for the first time you create a directory or put some file) hadoop fs -mkdir sample hadoop fs -ls I would suggest if you are getting pemission problem, please check the

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Mohammad Tariq
Hello Ashish, Change the permissions of /app/hadoop/tmp to 755 and see if it helps. Warm Regards, Tariq cloudfront.blogspot.com On Tue, Jul 23, 2013 at 10:27 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Thanks Jitendra, Bejoy and Yexi, I got past that. And now the ls command says it

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Hey, Thanks Shekhar. That worked like a chimp. Appreciate help from you all. Now I will try to put files and try the word count or similar program. Regards ashish On Tue, Jul 23, 2013 at 10:07 AM, Shekhar Sharma shekhar2...@gmail.comwrote: Its warning not error... Create a directory and

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks, But the issue was that there was no directory and hence it was not showing anything. Adding a directory cleared the warning. I appreciate your help. Regards ashish On Tue, Jul 23, 2013 at 10:08 AM, Mohammad Tariq donta...@gmail.com wrote: Hello Ashish, Change the permissions of

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
After starting i would suggest always check whether your NameNode and job tracker UI are working or not and check the number of live nodes in both of the UI.. Regards, Som Shekhar Sharma +91-8197243810 On Tue, Jul 23, 2013 at 10:41 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Thanks, But

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Jitendra Yadav
Try.. *hadoop fs -ls /* ** Thanks On Tue, Jul 23, 2013 at 10:27 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Thanks Jitendra, Bejoy and Yexi, I got past that. And now the ls command says it can not access the directory. I am sure this is a permissions issue. I am just wondering

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Jitendra, Som, Thanks. Issue was in not having any file there. Its working fine now. I am able to do -ls and could also do -mkdir and -put. Now is time to run the jar and apparently I am getting no main manifest attribute, in wc.jar But I believe its because of maven pom file does not have

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Shekhar Sharma
hadoop jar wc.jar fully qualified driver name inputdata outputdestination Regards, Som Shekhar Sharma +91-8197243810 On Tue, Jul 23, 2013 at 10:58 PM, Ashish Umrani ashish.umr...@gmail.comwrote: Jitendra, Som, Thanks. Issue was in not having any file there. Its working fine now. I am

Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Huy Pham
Hi All, Do any of you have or can refer me to some sample Java code that get the tree structure of a HDFS directory, similar to the file system? For example: I have a HDFS dir, called /data, inside data, there is /data/valid and /data/invalid, and so on, so I would need to be able to get

Re: New hadoop 1.2 single node installation giving problems

2013-07-23 Thread Ashish Umrani
Thanks Shekhar, The problem was not in my building of the jar. It was in fact in execution I was running command *hadoop -jar* jar filename qualified class name input output The problem was with -jar. It should be *hadoop jar* jar filename qualified class name input output Thanks for help

Re: Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Shahab Yunus
See this https://sites.google.com/site/hadoopandhive/home/how-to-read-all-files-in-a-directory-in-hdfs-using-hadoop-filesystem-api and http://hadoop.apache.org/docs/stable/api/org/apache/hadoop/fs/FileSystem.html#isDirectory(org.apache.hadoop.fs.Path) Basically you can write your own function,

Saving counters in Mapfile

2013-07-23 Thread Elazar Leibovich
Hi, A common use case one want an ordered structure for, is for saving counters. Naturally, I wanted to save the counters in a Mapfile: for (long ix = 0; ix MAXVALUE; ix++) { mapfile.append(new Text(counter key of val + ix), new LongWritable(ix)); } This however looks a bit

Re: Saving counters in Mapfile

2013-07-23 Thread manishbh...@rocketmail.com
Hi, If you intend to use those counters in further functions then I think Hadoop will take care this by itself, you can explore combiner for the same. In Sequence file mapfile and index files has its own functionality and as per my understanding index file expect offset to move over map file

Re: Saving counters in Mapfile

2013-07-23 Thread Michael Segel
Uhm... You want to save the counters as in counts per job run or something? (Remember HDFS == WORM) Then you could do a sequence file and then use something like HBase to manage the index. (Every time you add a set of counters, you have a new file and a new index.) Heck you could use HBase

Re: Get the tree structure of a HDFS dir, similar to dir/files

2013-07-23 Thread Harsh J
The FileSystem interface provides a recursive option for this. See http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FileSystem.html#listFiles(org.apache.hadoop.fs.Path,%20boolean) On Tue, Jul 23, 2013 at 11:35 PM, Huy Pham pha...@yahoo-inc.com wrote: Hi All, Do any of you have

Re: Only log.index

2013-07-23 Thread Vinod Kumar Vavilapalli
It could either mean that all those task-attempts are crashing before the process itself is getting spawned (check TT logs) or those logs are getting deleted after the fact. Suspect the earlier. Thanks, +Vinod On Jul 23, 2013, at 9:33 AM, Ajay Srivastava wrote: Hi, I see that most of the