[ 
https://issues.apache.org/jira/browse/HIVE-7347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Howell updated HIVE-7347:
-------------------------------
    Tags: webhcat, Pig, execute, schema, MicrosoftCSS  (was: webhcat, Pig, 
execute, schema, CSS)

> Pig Query with defined schema fails when submitted via WebHcat 'execute' 
> parameter
> ----------------------------------------------------------------------------------
>
>                 Key: HIVE-7347
>                 URL: https://issues.apache.org/jira/browse/HIVE-7347
>             Project: Hive
>          Issue Type: Bug
>          Components: WebHCat
>    Affects Versions: 0.12.0, 0.13.0
>         Environment: HDP 2.1 on Windows; HDInsight deploying HDP 2.1  
>            Reporter: Azim Uddin
>
> 1. Consider you are using HDP 2.1 on Windows, and you have a tsv file (named 
> rawInput.tsv) like this (just an example, you can use any) -
> http://a.com  http://b.com    1
> http://b.com  http://c.com    2
> http://d.com  http://e.com    3
> 2. With the tsv file uploaded to HDFS, run the following Pig job via WebHcat 
> using 'execute' parameter, something like this-
> curl.exe -d execute="rawInput = load '/test/data' using PigStorage as 
> (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); 
> readyInput = limit rawInput 10; store readyInput into '/test/output' using 
> PigStorage;" -d statusdir="/test/status" 
> "http://localhost:50111/templeton/v1/pig?user.name=hadoop"; --user hadoop:any
> The job fails with exit code 255 -
> "[main] org.apache.hive.hcatalog.templeton.tool.LaunchMapper: templeton: job 
> failed with exit code 255"
> From stderr, we see the following -"readyInput was unexpected at this time."
> 3. The same job works via Pig Grunt Shell and if we use the WebHcat 'file' 
> parameter, instead of 'execute' parameter - 
> a. Create a pig script called pig-script.txt with the query below and put it 
> HDFS /test/script
> rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, 
> DestinationUrl:chararray, InstanceCount:int);
> readyInput = limit rawInput 10;
> store readyInput into '/test/Output' using PigStorage;
> b. Run the job via webHcat:
> curl.exe -d file="/test/script/pig_script.txt" -d statusdir="/test/status" 
> "http://localhost:50111/templeton/v1/pig?user.name=hadoop"; --user hadoop:any
> 4. Also, WebHcat 'execute' option works if we don't define the schema in the 
> Pig query, something like this-
> curl.exe -d execute="rawInput = load '/test/data' using PigStorage; 
> readyInput = limit rawInput 10; store readyInput into '/test/output' using 
> PigStorage;" -d statusdir="/test/status" 
> "http://localhost:50111/templeton/v1/pig?user.name=hadoop"; --user hadoop:any
> Ask is-
> WebHcat 'execute' option should work for Pig query with schema defined - it 
> appears to be a parsing issue with WebHcat.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to