Note that this is home directory in HDFS not in Linux. Even if its a sudo user, you can still give him a directory in HDFS. Assuming the sudo user is called "etl":
hdfs dfs -mkdir /user/etl hdfs dfs -chown etl /user/etl Enjoy! Gwen On Mon, Jun 9, 2014 at 5:42 PM, Jarek Jarcec Cecho <[email protected]> wrote: > I'm pretty sure that any user that is running mapreduce job needs to have > valid home directory, right? > > Jarcec > > On Mon, Jun 09, 2014 at 04:27:34PM +0000, Vikash Talanki -X (vtalanki - > INFOSYS LIMITED at Cisco) wrote: > > Hi Jarcec, > > > > I am trying to import incremental data from Oracle to hive but not HDFS. > Though I have not specified --hive-import parameter in below command, I am > getting the same issue even when I use it. > > The reason I provided --target-dir is that currently the user with which > I am running sqoop command is a sudo user and it has no permissions to > write or create anything in its home directory(which sqoop uses as default > directory to import data). > > So, please let me know how does it work in case of loading incremental > data into hive. > > Do we need to provide --target-dir? If yes, what value it should be? > Hive warehouse location? > > > > Thanks, > > Vikash Talanki > > +1 (408)838-4078 > > > > -----Original Message----- > > From: Jarek Jarcec Cecho [mailto:[email protected]] > > Sent: Monday, June 09, 2014 7:38 AM > > To: [email protected] > > Subject: Re: How to specify --target-dir for sqoop incremental imports > into hive > > > > Did you actually tried to let Sqoop finish it's job? > > > > I believe that you are observing valid behaviour - MapReduce won't allow > you to import data into existing directory and hence Sqoop will firstly > import data into temporal directory and then move them to final destination > specified with --target-dir argument. > > > > Jarcec > > > > On Mon, Jun 09, 2014 at 05:50:55AM +0000, Vikash Talanki -X (vtalanki - > INFOSYS LIMITED at Cisco) wrote: > > > Hi All, > > > > > > > > > I want to use existing sqoop incremental parameters to load data from > oracle to hive. > > > > > > Here is the sqoop command : > > > > > > sqoop import -D > mapred.child.java.opts='\-Djava.security.egd=file:/dev/../dev/urandom' > --connect > 'jdbc:oracle:thin:@(DESCRIPTION=(ADDRESS_LIST=(FAILOVER=on)(LOAD_BALANCE=on)(ADDRESS=(PROTOCOL=TCP)(HOST=XXXXXX)(PORT=1530))(ADDRESS=(PROTOCOL=TCP)(HOST=XXXXX)(PORT=1530)))(CONNECT_DATA=(SERVER=dedicated)(SERVICE_NAME=XXXXXX)))' > --username XXXXX --password XXXXXX -m 1 --table XXCSS_KTN_REQ_LINE_DETAIL > --target-dir > /app/SmartAnalytics/Apps/frameworks_dataingestion.db/xxcss_ktn_req_line_detail_vtest > --hive-table frameworks_dataingestion.XXCSS_KTN_REQ_LINE_DETAIL_vtest > --map-column-hive > LINE_ITEM_ID=BIGINT,LIST_PRICE=BIGINT,SERVICE_VALUE=BIGINT --null-string > '\\N' --null-non-string '\\N' --hive-delims-replacement ' ' --check-column > LID_DATE --incremental append --last-value '2014-05-27 10:38:17.0' > > > > > > Even when I specify the target directory of my existing tables HDFS > file location it is still creating a different output directory - > > > 14/06/08 21:28:52 INFO mapred.JobClient: Creating job's output > directory at _sqoop/08212846713XXCSS_KTN_REQ_LINE_DETAIL > > > > > > Why is this happening? What needs to be provided for -target-dir? > > > Thanks in advance. > > > > > > [Description: > http://www.cisco.com/web/europe/images/email/signature/logo05.jpg] > > > > > > Vikash Talanki > > > Engineer - Software > > > [email protected] > > > Phone: +1 (408)838 4078 > > > > > > Cisco Systems Limited > > > SJ-J 3 > > > 255 W Tasman Dr > > > San Jose > > > CA - 95134 > > > United States > > > Cisco.com<http://www.cisco.com/> > > > > > > > > > > > > > > > > > > [Description: Think before you print.]Think before you print. > > > > > > This email may contain confidential and privileged material for the > sole use of the intended recipient. Any review, use, distribution or > disclosure by others is strictly prohibited. If you are not the intended > recipient (or authorized to receive for the recipient), please contact the > sender by reply email and delete all copies of this message. > > > For corporate legal information go to: > > > http://www.cisco.com/web/about/doing_business/legal/cri/index.html > > > > > > > > > > > > > > > >
