[ 
https://issues.apache.org/jira/browse/HADOOP-6988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12919181#action_12919181
 ] 

Devaraj Das commented on HADOOP-6988:
-------------------------------------

Although it is true that HADOOP_TOKEN_FILE_LOCATION can be used to make normal 
hdfs commands work, the intent for having this was to support security for 
Map/Reduce tasks, and, hadoop streaming apps that internally invoke 
command-line hdfs operations (as Owen had pointed out earlier). If you want to 
pass multiple tokens during job submission, the preferred approach would be to 
write the tokens into a file (using the Credentials class's utilities), and 
then point mapreduce.job.credentials.binary to that file. 
Thinking about it, wouldn't the option of defining mapreduce.job.hdfs-servers 
in the job configuration work for you. The JobClient will automatically get 
delegation tokens from those namenodes and all tasks of the job can use those 
tokens..

> Add support for reading multiple hadoop delegation token files
> --------------------------------------------------------------
>
>                 Key: HADOOP-6988
>                 URL: https://issues.apache.org/jira/browse/HADOOP-6988
>             Project: Hadoop Common
>          Issue Type: Improvement
>          Components: security
>    Affects Versions: 0.22.0
>            Reporter: Aaron T. Myers
>            Assignee: Aaron T. Myers
>         Attachments: hadoop-6988.0.txt, hadoop-6988.1.txt
>
>
> It would be nice if there were a way to specify multiple delegation token 
> files via the HADOOP_TOKEN_FILE_LOCATION environment variable and the 
> "mapreduce.job.credentials.binary" configuration value. I suggest a 
> colon-separated list of paths, each of which is read as a separate delegation 
> token file.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to