Hi,

I have successfully imported an Oracle table with CLOB data type to Hive Table. 
After processing that table, I am supposed to export back the same table to 
Oracle.

Source Oracle table types are  : ( DOCUMENT_ID  NUMBER,XML_DATA     CLOB, 
SUBMIT_DATE  DATE )
Hive table  types  : (document_id int,xml_data string, submit_date timestamp )
Export Oracle table : DOCUMENT_ID  NUMBER, XML_DATA     VARCHAR2(1000 BYTE), 
SUBMIT_DATE  DATE)
 I created this export table  because I cannot export the CLOB directly from 
Hive String type, so first export it to this table and then in oracle convert 
to CLOB.


Problem:
While exporting it hangs for a while and fails . I found an error in JOB 
Tracker.

ERROR org.apache.sqoop.mapreduce.AsyncSqlOutputFormat: Got exception in update 
thread: java.sql.SQLException: ORA-01461: can bind a LONG value only for insert 
into a LONG column

Export Error:
14/08/26 18:53:58 INFO mapred.JobClient:  map 0% reduce 0%
14/08/26 18:54:19 INFO mapred.JobClient:  map 100% reduce 0%
14/08/26 19:04:20 INFO mapred.JobClient:  map 0% reduce 0%
14/08/26 19:04:22 INFO mapred.JobClient: Task Id : 
attempt_201408041327_29224_m_000000_0, Status : FAILED
Task attempt_201408041327_29224_m_000000_0 failed to report status for 600 
seconds. Killing!
14/08/26 19:04:32 INFO mapred.JobClient:  map 100% reduce 0%
14/08/26 19:14:32 INFO mapred.JobClient:  map 0% reduce 0%
14/08/26 19:14:34 INFO mapred.JobClient: Task Id : 
attempt_201408041327_29224_m_000000_1, Status : FAILED
Task attempt_201408041327_29224_m_000000_1 failed to report status for 600 
seconds. Killing!
14/08/26 19:14:45 INFO mapred.JobClient:  map 100% reduce 0%


My steps:
Hive>create table sqp_clob_data7 (DOCUMENT_ID INT, XML_DATA STRING, SUBMIT_DATE 
TIMESTAMP);

sqoop import options:
sqoop import --connect jdbc:oracle:thin:@<host> --username <username>-P --table 
<tablename> --hive-import  --hive-table sqp_clob_data7 --hive-overwrite 
--hive-drop-import-delims  --map-column-java XML_DATA=String --verbose -m 1

export options:
sqoop export --connect jdbc:oracle:thin:@<host>  --username <username> -P 
--table <tabllename>  --export-dir /apps/hive/warehouse/sqp_clob_data7 
--verbose -m 1 --input-fields-terminated-by '\001'

Versions:
Hadoop - HW 1.3.2
Hive - 0.11.0.1.3.3.0
Sqoop - 1.4.3.1.3.3.0

I am not clear with the error which I see in job tracker, please let me know if 
anyone has come across such issue and if there is any fox for the same.

Thanks & Regards
Rajasekhar D

Reply via email to