[ 
https://issues.apache.org/jira/browse/HIVE-5072?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13966158#comment-13966158
 ] 

Eugene Koifman commented on HIVE-5072:
--------------------------------------

[~shuainie]  Here are a few comments:
0. If I understand this correctly, optionsFile should contain the details of 
Sqoop command to execute.  But in the code it seems that the expectation is 
that this file is present in DFS.  Thus to submit a Sqoop job via WebHcat (and 
use optionsFile) the user has to first upload this file to the cluster.  This 
is an extra call for job submission and possibly extra config on the cluster 
side to enable the client of WebHCat to upload files.  Why not just let the 
client upload the file WebHCat as part of the REST POST request?  This seems a 
lot more user friendly/usable.  
1. -d 'user.name=foo' is deprecated (i.e. user.name as a Form parameter).  
user.name has to be part of the query string.  The test cases and examples in 
.pdf should be updated.
2. Formatting in ScoopDelegator doesn't follow Hive conventions
3. Server.scoop() - there is Server.checkEnableLogPrerequisite() to check 
'enableLog' parameter setting.
4. I see that new parameters for Scoop tests are added in 3 places in 
build.xml.  Only the 'test' target actually runs jobsubmission.conf.
5. For the tests you added, where does the JDBC driver come from for any 
particular DB?
6. Can for Form parameter for optionsFile (Server.sqoop()) be called 
"optionsFile" instead of just "file"?
7. it seems from 
http://sqoop.apache.org/docs/1.4.4/SqoopUserGuide.html#_using_options_files_to_pass_arguments
 that in a Sqoop command, either options-file (with command and args) or 
command name and all args inline can be specifed.  The tests you added seem to 
expect only command args to be in options-file.  In particular Server.sqoop() 
tests "command == null && optionsFile == null" but not if both options are 
specified.  Seems like this is not expected usage.
8. Is there anything that can be done to make the test self-contained, so that 
the DB table is automatically created, for example in the DB that contains the 
metastore data?


> [WebHCat]Enable directly invoke Sqoop job through Templeton
> -----------------------------------------------------------
>
>                 Key: HIVE-5072
>                 URL: https://issues.apache.org/jira/browse/HIVE-5072
>             Project: Hive
>          Issue Type: Improvement
>          Components: WebHCat
>    Affects Versions: 0.12.0
>            Reporter: Shuaishuai Nie
>            Assignee: Shuaishuai Nie
>         Attachments: HIVE-5072.1.patch, HIVE-5072.2.patch, HIVE-5072.3.patch, 
> Templeton-Sqoop-Action.pdf
>
>
> Now it is hard to invoke a Sqoop job through templeton. The only way is to 
> use the classpath jar generated by a sqoop job and use the jar delegator in 
> Templeton. We should implement Sqoop Delegator to enable directly invoke 
> Sqoop job through Templeton.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to