Yes agreed.It seems to be issue with mapping the text file contents to case
classes, not sure though.

On Thu, Aug 4, 2016 at 8:17 PM, $iddhe$h Divekar <siddhesh.dive...@gmail.com
> wrote:

> Hi Deepak,
>
> My files are always > 50MB.
> I would think there would be a small config to overcome this.
> Tried almost everything i could after searching online.
>
> Any help from the mailing list would be appreciated.
>
> On Thu, Aug 4, 2016 at 7:43 AM, Deepak Sharma <deepakmc...@gmail.com>
> wrote:
>
>> I am facing the same issue with spark 1.5.2
>> If the file size that's being processed by spark , is of size 10-12 MB ,
>> it throws out of memory .
>> But if the same file is within 5 MB limit , it runs fine.
>> I am using spark configuration with 7GB of memory and 3 cores for
>> executors in the cluster of 8 executor.
>>
>> Thanks
>> Deepak
>>
>> On 4 Aug 2016 8:04 pm, "$iddhe$h Divekar" <siddhesh.dive...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I am running spark jobs using apache oozie in yarn-client mode.
>>> My job.properties has sparkConf which gets used in workflow.xml.
>>>
>>> I have tried increasing MaxPermSize using sparkConf in job.properties
>>> but that is not resolving the issue.
>>>
>>> *sparkConf*=--verbose --driver-java-options '-XX:MaxPermSize=8192M'
>>> --conf spark.speculation=false --conf spark.hadoop.spark.sql.
>>> parquet.output.committer.class=\
>>> "org.apache.spark.sql.parquet.DirectParquetOutputCommitter" --conf
>>> spark.hadoop.mapred.output.committer.class="org.apache.hadoop.mapred.
>>> DirectFileOutputCommit\
>>> ter.class" --conf spark.hadoop.mapreduce.use.
>>> directfileoutputcommitter=true
>>>
>>> Am I missing anything ?
>>>
>>> I am seeing following errors.
>>>
>>> 2016-08-03 22:33:43,318  WARN SparkActionExecutor:523 -
>>> SERVER[ip-10-0-0-161.ec2.internal] USER[hadoop] GROUP[-] TOKEN[]
>>> APP[ApprouteOozie] JOB[0000031-160803180548580-oozie-oozi-W]
>>> ACTION[0000031-160803180548580-oozie-oozi-W@spark-approute] Launcher
>>> ERROR, reason: Main class [org.apache.oozie.action.hadoop.SparkMain],
>>> main() threw exception, PermGen space
>>> 2016-08-03 22:33:43,319  WARN SparkActionExecutor:523 -
>>> SERVER[ip-10-0-0-161.ec2.internal] USER[hadoop] GROUP[-] TOKEN[]
>>> APP[ApprouteOozie] JOB[0000031-160803180548580-oozie-oozi-W]
>>> ACTION[0000031-160803180548580-oozie-oozi-W@spark-approute] Launcher
>>> exception: PermGen space
>>> java.lang.OutOfMemoryError: PermGen space
>>>
>>> oozie-oozi-W@spark-approute] Launcher exception: PermGen space
>>> java.lang.OutOfMemoryError: PermGen space
>>>     at java.lang.Class.getDeclaredConstructors0(Native Method)
>>>     at java.lang.Class.privateGetDeclaredConstructors(Class.java:2595)
>>>     at java.lang.Class.getConstructor0(Class.java:2895)
>>>     at java.lang.Class.newInstance(Class.java:354)
>>>     at sun.reflect.MethodAccessorGenerator$1.run(
>>> MethodAccessorGenerator.java:399)
>>>     at sun.reflect.MethodAccessorGenerator$1.run(
>>> MethodAccessorGenerator.java:396)
>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>     at sun.reflect.MethodAccessorGenerator.generate(
>>> MethodAccessorGenerator.java:395)
>>>     at sun.reflect.MethodAccessorGenerator.
>>> generateSerializationConstructor(MethodAccessorGenerator.java:113)
>>>     at sun.reflect.ReflectionFactory.newConstructorForSerialization
>>> (ReflectionFactory.java:331)
>>>     at java.io.ObjectStreamClass.getSerializableConstructor(
>>> ObjectStreamClass.java:1420)
>>>     at java.io.ObjectStreamClass.access$1500(ObjectStreamClass.java:72)
>>>     at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:497)
>>>     at java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:472)
>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>     at java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:472)
>>>     at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:369)
>>>     at java.io.ObjectOutputStream.writeObject0(
>>> ObjectOutputStream.java:1133)
>>>     at java.io.ObjectOutputStream.defaultWriteFields(
>>> ObjectOutputStream.java:1547)
>>>     at java.io.ObjectOutputStream.writeSerialData(
>>> ObjectOutputStream.java:1508)
>>>     at java.io.ObjectOutputStream.writeOrdinaryObject(
>>> ObjectOutputStream.java:1431)
>>>     at java.io.ObjectOutputStream.writeObject0(
>>> ObjectOutputStream.java:1177)
>>>     at java.io.ObjectOutputStream.defaultWriteFields(
>>> ObjectOutputStream.java:1547)
>>>     at java.io.ObjectOutputStream.writeSerialData(
>>> ObjectOutputStream.java:1508)
>>>     at java.io.ObjectOutputStream.writeOrdinaryObject(
>>> ObjectOutputStream.java:1431)
>>>     at java.io.ObjectOutputStream.writeObject0(
>>> ObjectOutputStream.java:1177)
>>>     at java.io.ObjectOutputStream.defaultWriteFields(
>>> ObjectOutputStream.java:1547)
>>>     at java.io.ObjectOutputStream.writeSerialData(
>>> ObjectOutputStream.java:1508)
>>>     at java.io.ObjectOutputStream.writeOrdinaryObject(
>>> ObjectOutputStream.java:1431)
>>>     at java.io.ObjectOutputStream.writeObject0(
>>> ObjectOutputStream.java:1177)
>>>     at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.
>>> java:1377)
>>>     at java.io.ObjectOutputStream.writeObject0(
>>> ObjectOutputStream.java:1173)
>>>
>>> --
>>> -$iddhi.
>>>
>>
>
>
> --
> -$iddhi.
>



-- 
Thanks
Deepak
www.bigdatabig.com
www.keosha.net

Reply via email to