[jira] [Updated] (HIVE-7347) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter
[ https://issues.apache.org/jira/browse/HIVE-7347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Howell updated HIVE-7347: --- Tags: webhcat, Pig, execute, schema, CSS (was: webhcat, Pig, execute, schema) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter -- Key: HIVE-7347 URL: https://issues.apache.org/jira/browse/HIVE-7347 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0, 0.13.0 Environment: HDP 2.1 on Windows; HDInsight deploying HDP 2.1 Reporter: Azim Uddin 1. Consider you are using HDP 2.1 on Windows, and you have a tsv file (named rawInput.tsv) like this (just an example, you can use any) - http://a.com http://b.com1 http://b.com http://c.com2 http://d.com http://e.com3 2. With the tsv file uploaded to HDFS, run the following Pig job via WebHcat using 'execute' parameter, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any The job fails with exit code 255 - [main] org.apache.hive.hcatalog.templeton.tool.LaunchMapper: templeton: job failed with exit code 255 From stderr, we see the following -readyInput was unexpected at this time. 3. The same job works via Pig Grunt Shell and if we use the WebHcat 'file' parameter, instead of 'execute' parameter - a. Create a pig script called pig-script.txt with the query below and put it HDFS /test/script rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/Output' using PigStorage; b. Run the job via webHcat: curl.exe -d file=/test/script/pig_script.txt -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any 4. Also, WebHcat 'execute' option works if we don't define the schema in the Pig query, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage; readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any Ask is- WebHcat 'execute' option should work for Pig query with schema defined - it appears to be a parsing issue with WebHcat. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-7347) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter
[ https://issues.apache.org/jira/browse/HIVE-7347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Howell updated HIVE-7347: --- Tags: webhcat, Pig, execute, schema, MicrosoftCSS (was: webhcat, Pig, execute, schema, CSS) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter -- Key: HIVE-7347 URL: https://issues.apache.org/jira/browse/HIVE-7347 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0, 0.13.0 Environment: HDP 2.1 on Windows; HDInsight deploying HDP 2.1 Reporter: Azim Uddin 1. Consider you are using HDP 2.1 on Windows, and you have a tsv file (named rawInput.tsv) like this (just an example, you can use any) - http://a.com http://b.com1 http://b.com http://c.com2 http://d.com http://e.com3 2. With the tsv file uploaded to HDFS, run the following Pig job via WebHcat using 'execute' parameter, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any The job fails with exit code 255 - [main] org.apache.hive.hcatalog.templeton.tool.LaunchMapper: templeton: job failed with exit code 255 From stderr, we see the following -readyInput was unexpected at this time. 3. The same job works via Pig Grunt Shell and if we use the WebHcat 'file' parameter, instead of 'execute' parameter - a. Create a pig script called pig-script.txt with the query below and put it HDFS /test/script rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/Output' using PigStorage; b. Run the job via webHcat: curl.exe -d file=/test/script/pig_script.txt -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any 4. Also, WebHcat 'execute' option works if we don't define the schema in the Pig query, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage; readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any Ask is- WebHcat 'execute' option should work for Pig query with schema defined - it appears to be a parsing issue with WebHcat. -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Updated] (HIVE-7347) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter
[ https://issues.apache.org/jira/browse/HIVE-7347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Howell updated HIVE-7347: --- Tags: webhcat, Pig, execute, schema, MicrosoftSupport (was: webhcat, Pig, execute, schema, MicrosoftCSS) Pig Query with defined schema fails when submitted via WebHcat 'execute' parameter -- Key: HIVE-7347 URL: https://issues.apache.org/jira/browse/HIVE-7347 Project: Hive Issue Type: Bug Components: WebHCat Affects Versions: 0.12.0, 0.13.0 Environment: HDP 2.1 on Windows; HDInsight deploying HDP 2.1 Reporter: Azim Uddin 1. Consider you are using HDP 2.1 on Windows, and you have a tsv file (named rawInput.tsv) like this (just an example, you can use any) - http://a.com http://b.com1 http://b.com http://c.com2 http://d.com http://e.com3 2. With the tsv file uploaded to HDFS, run the following Pig job via WebHcat using 'execute' parameter, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any The job fails with exit code 255 - [main] org.apache.hive.hcatalog.templeton.tool.LaunchMapper: templeton: job failed with exit code 255 From stderr, we see the following -readyInput was unexpected at this time. 3. The same job works via Pig Grunt Shell and if we use the WebHcat 'file' parameter, instead of 'execute' parameter - a. Create a pig script called pig-script.txt with the query below and put it HDFS /test/script rawInput = load '/test/data' using PigStorage as (SourceUrl:chararray, DestinationUrl:chararray, InstanceCount:int); readyInput = limit rawInput 10; store readyInput into '/test/Output' using PigStorage; b. Run the job via webHcat: curl.exe -d file=/test/script/pig_script.txt -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any 4. Also, WebHcat 'execute' option works if we don't define the schema in the Pig query, something like this- curl.exe -d execute=rawInput = load '/test/data' using PigStorage; readyInput = limit rawInput 10; store readyInput into '/test/output' using PigStorage; -d statusdir=/test/status http://localhost:50111/templeton/v1/pig?user.name=hadoop; --user hadoop:any Ask is- WebHcat 'execute' option should work for Pig query with schema defined - it appears to be a parsing issue with WebHcat. -- This message was sent by Atlassian JIRA (v6.3.4#6332)