I think you need to define datatypes of variable. By default pig consider each variable as chararray.
*Ajay Dubey* On Tue, Sep 1, 2015 at 12:09 PM, rakesh sharma <[email protected]> wrote: > -- read airline datacontent = load '/home/19659/testData.csv' using > PigStorage(',');data = foreach content generate $1 as id , $2 as > type;group_data = group data by type;rec_count = foreach group_data > generate COUNT($0);limit_rec_count = limit rec_count 1;dump limit_rec_count; > > the above is my pig script, but I am getting error when I am trying to > dump data else there is no error. > > java.lang.Exception: org.apache.pig.backend.executionengine.ExecException: > ERROR 0: Exception while executing [POCast (Name: Cast[bag:{}] - scope-12 > Operator Key: scope-12) children: [[POProject (Name: Project[bytearray][0] > - scope-11 Operator Key: scope-11) children: null at []]] at []]: > java.lang.NullPointerException at > org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) > at > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529)Caused > by: org.apache.pig.backend.executionengine.ExecException: ERROR 0: > Exception while executing [POCast (Name: Cast[bag:{}] - scope-12 Operator > Key: scope-12) children: [[POProject (Name: Project[bytearray][0] - > scope-11 Operator Key: scope-11) children: null at []]] at []]: > java.lang.NullPointerException at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:366) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.processInput(POUserFunc.java:216) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:270) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNextLong(POUserFunc.java:407) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:351) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.processPlan(POForEach.java:383) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.getNextTuple(POForEach.java:303) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.runPipeline(PigGenericMapReduce.java:474) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.processOnePackageOutput(PigGenericMapReduce.java:442) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:422) > at > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:269) > at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:171) > at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:627) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389) > at > org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745)Caused by: > java.lang.NullPointerException at > org.apache.pig.builtin.Utf8StorageConverter.consumeBag(Utf8StorageConverter.java:80) > at > org.apache.pig.builtin.Utf8StorageConverter.bytesToBag(Utf8StorageConverter.java:335) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POCast.getNextDataBag(POCast.java:1861) > at > org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:337) > ... 19 more2015-09-01 12:01:45,920 [main] INFO > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - HadoopJobId: job_local1740385966_00022015-09-01 12:01:45,920 [main] INFO > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - Processing aliases rec_count2015-09-01 12:01:45,920 [main] INFO > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - detailed locations: M: C: R: rec_count[5,12]2015-09-01 12:01:45,922 > [main] WARN > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to > stop immediately on failure.2015-09-01 12:01:45,923 [main] INFO > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - job job_local1740385966_0002 has failed! Stop running all dependent > jobs2015-09-01 12:01:45,923 [main] INFO > org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher > - 100% complete2015-09-01 12:01:45,923 [main] INFO > org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics > with processName=JobTracker, sessionId= - already initialized2015-09-01 > 12:01:45,924 [main] INFO org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot > initialize JVM Metrics with processName=JobTracker, sessionId= - already > initialized2015-09-01 12:01:45,924 [main] ERROR > org.apache.pig.tools.pigstats.mapreduce.MRPigStatsUtil - 1 map reduce > job(s) failed!2015-09-01 12:01:45,925 [main] INFO > org.apache.pig.tools.pigstats.mapreduce.SimplePigStats - Script Statistics: > HadoopVersion PigVersion UserId StartedAt FinishedAt > Features2.6.0.2.2.4.2-2 0.14.0.2.2.4.2-2 19659 2015-09-01 > 12:01:43 2015-09-01 12:01:45 GROUP_BY,LIMIT > Some jobs have failed! Stop running all dependent jobs > Job Stats (time in seconds):JobId Maps Reduces MaxMapTime > MinMapTime AvgMapTime MedianMapTime MaxReduceTime > MinReduceTime AvgReduceTime MedianReducetime Alias Feature > Outputsjob_local95835876_0001 1 1 n/a n/a n/a n/a > n/a n/a n/a n/a content,data,group_data,limit_rec_count > GROUP_BY > Failed Jobs:JobId Alias Feature Message > Outputsjob_local1740385966_0002 rec_count Message: Job > failed! file:/tmp/temp1464312773/tmp-2093473355, > Input(s):Successfully read 46556 records from: "/home/19659/testData.csv" > Output(s):Failed to produce result in > "file:/tmp/temp1464312773/tmp-2093473355" > Counters:Total records written : 0Total bytes written : 0Spillable Memory > Manager spill count : 0Total bags proactively spilled: 0Total records > proactively spilled: 0 > Job DAG:job_local95835876_0001 -> > job_local1740385966_0002,job_local1740385966_0002 > > > Can somebody help with this. I am running pig locally. >
