Hi Jonathan,

Which Hadoop version are you using? I'm actually wondering if
mapred.child.java.opts is still supported in Hadoop 2.x (I think it
has been replaced by mapreduce.map.java.opts and
mapreduce.reduce.java.opts).

The HADOOP_CLIENT_OPTS won't make a difference if you're running in
(pseudo) distributed mode, as separate JVMs will be started up for the
tasks.

- Gabriel


On Fri, Dec 18, 2015 at 7:33 PM, Cox, Jonathan A <ja...@sandia.gov> wrote:
> Gabriel,
>
> I am running the job on a single machine in pseudo distributed mode. I've set 
> the max Java heap size in two different ways (just to be sure):
>
> export HADOOP_CLIENT_OPTS="$HADOOP_CLIENT_OPTS -Xmx48g"
>
> and also in mapred-site.xml:
>   <property>
>     <name>mapred.child.java.opts</name>
>     <value>-Xmx48g</value>
>   </property>
>
> -----Original Message-----
> From: Gabriel Reid [mailto:gabriel.r...@gmail.com]
> Sent: Friday, December 18, 2015 8:17 AM
> To: user@phoenix.apache.org
> Subject: [EXTERNAL] Re: Java Out of Memory Errors with CsvBulkLoadTool
>
> Hi Jonathan,
>
> Sounds like something is very wrong here.
>
> Are you running the job on an actual cluster, or are you using the local job 
> tracker (i.e. running the import job on a single computer).
>
> Normally an import job, regardless of the size of the input, should run with 
> map and reduce tasks that have a standard (e.g. 2GB) heap size per task 
> (although there will typically be multiple tasks started on the cluster). 
> There shouldn't be any need to have anything like a 48GB heap.
>
> If you are running this on an actual cluster, could you elaborate on 
> where/how you're setting the 48GB heap size setting?
>
> - Gabriel
>
>
> On Fri, Dec 18, 2015 at 1:46 AM, Cox, Jonathan A <ja...@sandia.gov> wrote:
>> I am trying to ingest a 575MB CSV file with 192,444 lines using the
>> CsvBulkLoadTool MapReduce job. When running this job, I find that I
>> have to boost the max Java heap space to 48GB (24GB fails with Java
>> out of memory errors).
>>
>>
>>
>> I’m concerned about scaling issues. It seems like it shouldn’t require
>> between 24-48GB of memory to ingest a 575MB file. However, I am pretty
>> new to Hadoop/HBase/Phoenix, so maybe I am off base here.
>>
>>
>>
>> Can anybody comment on this observation?
>>
>>
>>
>> Thanks,
>>
>> Jonathan

Reply via email to