Hi, I have successfully imported an Oracle table with CLOB data type to Hive Table. After processing that table, I am supposed to export back the same table to Oracle.
Source Oracle table types are : ( DOCUMENT_ID NUMBER,XML_DATA CLOB, SUBMIT_DATE DATE ) Hive table types : (document_id int,xml_data string, submit_date timestamp ) Export Oracle table : DOCUMENT_ID NUMBER, XML_DATA VARCHAR2(1000 BYTE), SUBMIT_DATE DATE) I created this export table because I cannot export the CLOB directly from Hive String type, so first export it to this table and then in oracle convert to CLOB. Problem: While exporting it hangs for a while and fails . I found an error in JOB Tracker. ERROR org.apache.sqoop.mapreduce.AsyncSqlOutputFormat: Got exception in update thread: java.sql.SQLException: ORA-01461: can bind a LONG value only for insert into a LONG column Export Error: 14/08/26 18:53:58 INFO mapred.JobClient: map 0% reduce 0% 14/08/26 18:54:19 INFO mapred.JobClient: map 100% reduce 0% 14/08/26 19:04:20 INFO mapred.JobClient: map 0% reduce 0% 14/08/26 19:04:22 INFO mapred.JobClient: Task Id : attempt_201408041327_29224_m_000000_0, Status : FAILED Task attempt_201408041327_29224_m_000000_0 failed to report status for 600 seconds. Killing! 14/08/26 19:04:32 INFO mapred.JobClient: map 100% reduce 0% 14/08/26 19:14:32 INFO mapred.JobClient: map 0% reduce 0% 14/08/26 19:14:34 INFO mapred.JobClient: Task Id : attempt_201408041327_29224_m_000000_1, Status : FAILED Task attempt_201408041327_29224_m_000000_1 failed to report status for 600 seconds. Killing! 14/08/26 19:14:45 INFO mapred.JobClient: map 100% reduce 0% My steps: Hive>create table sqp_clob_data7 (DOCUMENT_ID INT, XML_DATA STRING, SUBMIT_DATE TIMESTAMP); sqoop import options: sqoop import --connect jdbc:oracle:thin:@<host> --username <username>-P --table <tablename> --hive-import --hive-table sqp_clob_data7 --hive-overwrite --hive-drop-import-delims --map-column-java XML_DATA=String --verbose -m 1 export options: sqoop export --connect jdbc:oracle:thin:@<host> --username <username> -P --table <tabllename> --export-dir /apps/hive/warehouse/sqp_clob_data7 --verbose -m 1 --input-fields-terminated-by '\001' Versions: Hadoop - HW 1.3.2 Hive - 0.11.0.1.3.3.0 Sqoop - 1.4.3.1.3.3.0 I am not clear with the error which I see in job tracker, please let me know if anyone has come across such issue and if there is any fox for the same. Thanks & Regards Rajasekhar D
