I think you need to define datatypes of variable. By default pig consider
each variable as chararray.

*Ajay Dubey*

On Tue, Sep 1, 2015 at 12:09 PM, rakesh sharma <[email protected]>
wrote:

> -- read airline datacontent = load '/home/19659/testData.csv' using
> PigStorage(',');data = foreach content generate $1 as id , $2 as
> type;group_data = group data by type;rec_count = foreach group_data
> generate COUNT($0);limit_rec_count = limit rec_count 1;dump limit_rec_count;
>
> the above is my pig script, but I am getting error when I am trying to
> dump data else there is no error.
>
> java.lang.Exception: org.apache.pig.backend.executionengine.ExecException:
> ERROR 0: Exception while executing [POCast (Name: Cast[bag:{}] - scope-12
> Operator Key: scope-12) children: [[POProject (Name: Project[bytearray][0]
> - scope-11 Operator Key: scope-11) children: null at []]] at []]:
> java.lang.NullPointerException        at
> org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)
>       at
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529)Caused
> by: org.apache.pig.backend.executionengine.ExecException: ERROR 0:
> Exception while executing [POCast (Name: Cast[bag:{}] - scope-12 Operator
> Key: scope-12) children: [[POProject (Name: Project[bytearray][0] -
> scope-11 Operator Key: scope-11) children: null at []]] at []]:
> java.lang.NullPointerException        at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:366)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.processInput(POUserFunc.java:216)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNext(POUserFunc.java:270)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POUserFunc.getNextLong(POUserFunc.java:407)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:351)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.processPlan(POForEach.java:383)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.relationalOperators.POForEach.getNextTuple(POForEach.java:303)
>       at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.runPipeline(PigGenericMapReduce.java:474)
>       at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.processOnePackageOutput(PigGenericMapReduce.java:442)
>       at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:422)
>       at
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce$Reduce.reduce(PigGenericMapReduce.java:269)
>       at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:171)
> at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:627)
>       at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
>   at
> org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319)
>       at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>     at java.util.concurrent.FutureTask.run(FutureTask.java:262)        at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:745)Caused by:
> java.lang.NullPointerException        at
> org.apache.pig.builtin.Utf8StorageConverter.consumeBag(Utf8StorageConverter.java:80)
>       at
> org.apache.pig.builtin.Utf8StorageConverter.bytesToBag(Utf8StorageConverter.java:335)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.expressionOperators.POCast.getNextDataBag(POCast.java:1861)
>       at
> org.apache.pig.backend.hadoop.executionengine.physicalLayer.PhysicalOperator.getNext(PhysicalOperator.java:337)
>       ... 19 more2015-09-01 12:01:45,920 [main] INFO
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - HadoopJobId: job_local1740385966_00022015-09-01 12:01:45,920 [main] INFO
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - Processing aliases rec_count2015-09-01 12:01:45,920 [main] INFO
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - detailed locations: M:  C:  R: rec_count[5,12]2015-09-01 12:01:45,922
> [main] WARN
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - Ooops! Some job has failed! Specify -stop_on_failure if you want Pig to
> stop immediately on failure.2015-09-01 12:01:45,923 [main] INFO
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - job job_local1740385966_0002 has failed! Stop running all dependent
> jobs2015-09-01 12:01:45,923 [main] INFO
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.MapReduceLauncher
> - 100% complete2015-09-01 12:01:45,923 [main] INFO
> org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot initialize JVM Metrics
> with processName=JobTracker, sessionId= - already initialized2015-09-01
> 12:01:45,924 [main] INFO  org.apache.hadoop.metrics.jvm.JvmMetrics - Cannot
> initialize JVM Metrics with processName=JobTracker, sessionId= - already
> initialized2015-09-01 12:01:45,924 [main] ERROR
> org.apache.pig.tools.pigstats.mapreduce.MRPigStatsUtil - 1 map reduce
> job(s) failed!2015-09-01 12:01:45,925 [main] INFO
> org.apache.pig.tools.pigstats.mapreduce.SimplePigStats - Script Statistics:
> HadoopVersion   PigVersion      UserId  StartedAt       FinishedAt
> Features2.6.0.2.2.4.2-2 0.14.0.2.2.4.2-2        19659   2015-09-01
> 12:01:43     2015-09-01 12:01:45     GROUP_BY,LIMIT
> Some jobs have failed! Stop running all dependent jobs
> Job Stats (time in seconds):JobId   Maps    Reduces MaxMapTime
> MinMapTime      AvgMapTime      MedianMapTime   MaxReduceTime
>  MinReduceTime   AvgReduceTime   MedianReducetime        Alias   Feature
> Outputsjob_local95835876_0001  1       1       n/a     n/a     n/a     n/a
>    n/a     n/a     n/a     n/a     content,data,group_data,limit_rec_count
> GROUP_BY
> Failed Jobs:JobId   Alias   Feature Message
> Outputsjob_local1740385966_0002        rec_count               Message: Job
> failed!    file:/tmp/temp1464312773/tmp-2093473355,
> Input(s):Successfully read 46556 records from: "/home/19659/testData.csv"
> Output(s):Failed to produce result in
> "file:/tmp/temp1464312773/tmp-2093473355"
> Counters:Total records written : 0Total bytes written : 0Spillable Memory
> Manager spill count : 0Total bags proactively spilled: 0Total records
> proactively spilled: 0
> Job DAG:job_local95835876_0001  ->
> job_local1740385966_0002,job_local1740385966_0002
>
>
> Can somebody help with this. I am running pig locally.
>

Reply via email to