liyunzhang_intel created PIG-5051:
-------------------------------------
Summary: Initialize PigContants.TASK_INDEX in spark mode correctly
Key: PIG-5051
URL: https://issues.apache.org/jira/browse/PIG-5051
Project: Pig
Issue Type: Sub-task
Reporter: liyunzhang_intel
in MR, we initialize PigContants.TASK_INDEX in
org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigGenericMapReduce.Reduce#setup
{code}
protected void setup(Context context) throws IOException, InterruptedException {
...
context.getConfiguration().set(PigConstants.TASK_INDEX,
Integer.toString(context.getTaskAttemptID().getTaskID().getId()));
...
}
{code}
But spark does not provide funtion like PigGenericMapReduce.Reduce#setup to
initialize PigContants.TASK_INDEX when job starts. We need find a solution to
initialize PigContants.TASK_INDEX correctly.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)