CfP 2013 Workshop on Middleware for HPC and Big Data Systems (MHPC'13)

2013-04-23 Thread MHPC 2013
we apologize if you receive multiple copies of this message === CALL FOR PAPERS 2013 Workshop on Middleware for HPC and Big Data Systems MHPC '13 as part of Euro-Par 2013, Aachen, Germany

CfP 2013 Workshop on Middleware for HPC and Big Data Systems (MHPC'13)

2013-04-23 Thread MHPC 2013
we apologize if you receive multiple copies of this message === CALL FOR PAPERS 2013 Workshop on Middleware for HPC and Big Data Systems MHPC '13 as part of Euro-Par 2013, Aachen, Germany

Re: Is disk use reported with replication?

2013-04-23 Thread burberry blues
Hi I am new to Hadoop world. Can you please let me know what is a hadoop stack? Thanks, Burberry On Mon, Apr 22, 2013 at 10:19 AM, Keith Wiley kwi...@keithwiley.com wrote: Simple question: When I issue a hadoop fs -du command and/or when I view the namenode web UI to see HDFS disk

Re: Is disk use reported with replication?

2013-04-23 Thread Harsh J
Hi Keith, The fs -du computes length of files, and would not report replicated on-disk size. HDFS disk utilization OTOH, is the current, simple report of used/free disk space, which would certainly include replicated data. On Mon, Apr 22, 2013 at 10:49 PM, Keith Wiley kwi...@keithwiley.com

Re: How to open DEBUG level for YARN application master ?

2013-04-23 Thread Harsh J
To change the MR AM's default log level from INFO, set the job config: yarn.app.mapreduce.am.log.level to DEBUG or whatever level you prefer. On Mon, Apr 22, 2013 at 6:21 PM, Nitzan Raanan raanan.nit...@comverse.com wrote: Hi How do I open the DEBUG level for YARN application master process

Error about MR task when running 2T data

2013-04-23 Thread 姚吉龙
Hi Everyone Today I am testing about 2T data on my cluster, there several failed map task and reduce task on same node Here is the log Map failed: org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for output/spill0.out at

Re: Error about MR task when running 2T data

2013-04-23 Thread Harsh J
Does your node5 have adequate free space and proper multi-disk mapred.local.dir configuration set in it? On Tue, Apr 23, 2013 at 12:41 PM, 姚吉龙 geelong...@gmail.com wrote: Hi Everyone Today I am testing about 2T data on my cluster, there several failed map task and reduce task on same node

RE: How to open DEBUG level for YARN application master ?

2013-04-23 Thread Nitzan Raanan
Thanks That worked ! BR Raanan Nitzan -Original Message- From: Harsh J [mailto:ha...@cloudera.com] Sent: Tuesday, April 23, 2013 9:42 AM To: user@hadoop.apache.org Subject: Re: How to open DEBUG level for YARN application master ? To change the MR AM's default log level from INFO, set

Re: Sorting Values sent to reducer NOT based on KEY (Depending on part of VALUE)

2013-04-23 Thread Sofia Georgiakaki
Hello, Sorting is done by the SortingComparator which performs sorting based on the value of key. A possible solution would be the following: You could write a custom Writable comparable class which extends WritableComparable (lets call it MyCompositeFieldWritableComparable), that will store

unsubscribe

2013-04-23 Thread Nitzan Raanan
This e-mail message may contain confidential, commercial or privileged information that constitutes proprietary information of Comverse Technology or its subsidiaries. If you are not the intended recipient of this message, you are hereby notified that any

Mapper in-memory buffering

2013-04-23 Thread Peter Marron
Hi, Just a question about the implementation of Map/Reduce. I've been thinking about the output of the map stage. Logically all of the records emitted by the mapper have to be partitioned and sorted before they go into the reducers. (We can ignore the partitioning for the moment and so I'm just

Re: Error about MR task when running 2T data

2013-04-23 Thread Geelong Yao
I have set two disk available for tem file, one is /usr another is /sda But I found the first /usr is full while /sda has not been used. Why would this hadppen ? especially when the first path is full [image: 内嵌图片 1] 2013/4/23 Harsh J ha...@cloudera.com Does your node5 have adequate free space

unsubscribe

2013-04-23 Thread Gustavo Ioschpe
Enviado por Samsung Mobile

Re: unsubscribe

2013-04-23 Thread Panshul Whisper
Lol On Apr 23, 2013 10:39 AM, Gustavo Ioschpe gustavo.iosc...@bigdata.inf.br wrote: Enviado por Samsung Mobile

Re: Hadoop sampler related query!

2013-04-23 Thread Rahul Bhattacharjee
+ mapred dev On Tue, Apr 16, 2013 at 2:19 PM, Rahul Bhattacharjee rahul.rec@gmail.com wrote: Hi, I have a question related to Hadoop's input sampler ,which is used for investigating the data set before hand using random selection , sampling etc .Mainly used for total sort , used in

Re: Uploading file to HDFS

2013-04-23 Thread 超级塞亚人
Asking for help! I'm facing the problem that no datanode to stop. Namenode has been started but datanode can't be started. What should I do on namenode and datanode? Thank you very much 2013/4/19 超级塞亚人 shel...@gmail.com I have a problem. Our cluster has 32 nodes. Each disk is 1TB. I wanna

Job launch from eclipse

2013-04-23 Thread Han JU
Hi, I'm getting my hands on hadoop. One thing I really want to know is how you launch MR jobs in a development environment. I'm currently using Eclipse 3.7 with hadoop plugin from hadoop 1.0.2. With this plugin I can manage HDFS and submit job to cluster. But the strange thing is, every job

Re: Job launch from eclipse

2013-04-23 Thread shashwat shriparv
You need to generate a jar file, pass all the parameters on run time if any is fixed and run at hadoop like hadoop -jar jarfilename.jar parameters *Thanks Regards* ∞ Shashwat Shriparv On Tue, Apr 23, 2013 at 6:51 PM, Han JU ju.han.fe...@gmail.com wrote: Hi, I'm getting my hands on

Unsubscribe

2013-04-23 Thread neeraj.maha...@absicorp.com
Regards, Neeraj Mahajan Disclaimer This communication (including the attached document(s), if a document is attached hereto) contains information that is proprietary and confidential to ABSi, which shall not be disclosed or disseminated outside of ABSi except in connection with ABSi business

Re: Job launch from eclipse

2013-04-23 Thread Mohammad Tariq
Hell Han, The reason behind this is that the jobs are running inside the Eclipse itself and not getting submitted to your cluster. Please see if this links helps : http://cloudfront.blogspot.in/2013/03/mapreduce-jobs-running-through-eclipse.html#.UXaQsDWH6IQ Warm Regards, Tariq

Re: Unsubscribe

2013-04-23 Thread Mohammad Tariq
You need to send the request to this address : user-unsubscr...@hadoop.apache.org Warm Regards, Tariq https://mtariq.jux.com/ cloudfront.blogspot.com On Tue, Apr 23, 2013 at 7:14 PM, neeraj.maha...@absicorp.com neeraj.maha...@absicorp.com wrote: Regards, Neeraj Mahajan Disclaimer This

Re: Uploading file to HDFS

2013-04-23 Thread Mohammad Tariq
Hi there, Could you plz show me your config files and DN error logs? Warm Regards, Tariq https://mtariq.jux.com/ cloudfront.blogspot.com On Tue, Apr 23, 2013 at 4:35 PM, 超级塞亚人 shel...@gmail.com wrote: Asking for help! I'm facing the problem that no datanode to stop. Namenode has been

Re: Uploading file to HDFS

2013-04-23 Thread shashwat shriparv
On Tue, Apr 23, 2013 at 9:23 PM, Mohammad Tariq donta...@gmail.com wrote: What should I do on namenode and datanode? Thank you very much As Tariq has ask, can you provide datanode logs snapshots?? *Thanks Regards* ∞ Shashwat Shriparv

Re: Sqoop installation?

2013-04-23 Thread Chris Nauroth
Transferring to user list (hdfs-dev bcc'd). Hi Kevin, The datanodes are definitely more disposable than the namenodes. If a Sqoop command unexpectedly consumes a lot of resources, then stealing resources from the namenode could impact performance of the whole cluster. Stealing resources from a

Re: Hadoop MapReduce

2013-04-23 Thread Daryn Sharp
MR has a local mode that does what you want. Pig has the ability to use this mode. I did a quick search but didn't immediately find a good link to documentation, but hopefully this gets you going in the right direction. Daryn On Apr 22, 2013, at 6:01 PM, David Gkogkritsiani wrote: Helllo,

CfP 2013 Workshop on Middleware for HPC and Big Data Systems (MHPC'13)

2013-04-23 Thread MHPC 2013
we apologize if you receive multiple copies of this message === CALL FOR PAPERS 2013 Workshop on Middleware for HPC and Big Data Systems MHPC '13 as part of Euro-Par 2013, Aachen, Germany

Import with Sqoop

2013-04-23 Thread Kevin Burton
I execute the line: sqoop import --connect 'jdbc:sqlserver://nbreports:1433;databaseName=productcatalog' --username USER --password PASSWORD --table CatalogProducts And I get the following output: Warning: /usr/lib/hbase does not exist! HBase imports will fail. Please set $HBASE_HOME

Re: Import with Sqoop

2013-04-23 Thread Kathleen Ting
Hi Kevin, 1. What's the output from hadoop fs -cat CatalogProducts/part-m-0 2. Can you re-run with the --verbose option - i.e. sqoop import --connect 'jdbc:sqlserver://nbreports:1433;databaseName=productcatalog' --username USER --password PASSWORD --table CatalogProducts --verbose If

How to make the setting changes works

2013-04-23 Thread Geelong Yao
Hi Sorry to interrupt you.But nobody answer my question in Hadoop maillist. I have met a issue after I change the content of hdfs-site.xml to add another dfs.data.dir in my cluster. /usr/hadoop/tmp/dfs/data is the default value, /sda is the new one property namedata.dfs.dir/name

namenode memory test

2013-04-23 Thread 自己
Hi, I would like to know how much memory our data take on the name-node per block, file and directory. For example, the metadata size of a file. When I store some files in HDFS,how can I get the memory size take on the name-node? Is there some tools or commands to test the memory size take on

Re: Sorting Values sent to reducer NOT based on KEY (Depending on part of VALUE)

2013-04-23 Thread Vikas Jadhav
Thanks for reply. Will try to implement. I think there is problem in my case where i have modified write function of mapper context.write and tried to write same key value pair multiple times.Also for this purpose i have modified partitioner class. my partitioner class doesnt return single value

Pipes Property Not Passed In

2013-04-23 Thread Xun TANG
I implemented my own InputFormat/RecordReader, and I try to run it with Hadoop Pipes. I understand I could pass in properties to Pipes program by either: property namehadoop.pipes.java.recordreader/name valuefalse/value /property or alterntively -D

Re: Hadoop sampler related query!

2013-04-23 Thread Mahesh Balija
Hi Rahul, The limitation to use InputSampler is, the K and OK (I mean Map INKEY and OUTKEY) both should be of same type. Technically because, while collecting the samples (ie., arraylist of keys) in writePartitionFile method it uses the INKEY as the key. And for writing

Fwd: Multiple ways to write Hadoop program driver - Which one to choose?

2013-04-23 Thread Chandrashekhar Kotekar
Hi, I have observed that there are multiple ways to write driver method of Hadoop program. Following method is given in Hadoop Tutorial by Yahoohttp://developer.yahoo.com/hadoop/tutorial/module4.html public void run(String inputPath, String outputPath) throws Exception { JobConf conf =