[
https://issues.apache.org/jira/browse/SQOOP-934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14009643#comment-14009643
]
Andrey Dmitriev commented on SQOOP-934:
---------------------------------------
Hi,
I'm sorry if this is wrong place to write, but we are using Sqoop v1.4.4
{quote}
14/05/27 13:49:14 INFO sqoop.Sqoop: Running Sqoop version: 1.4.4-cdh5.0.0
Sqoop 1.4.4-cdh5.0.0
git commit id 8e266e052e423af592871e2dfe09d54c03f6a0e8
{quote}
And when I importing a table from Oracle which takes more than 1 hour to
extract, I'm getting following error message at the stage when Sqoop tries to
import data from temporary HDFS location to Hive:
{quote}
14/05/27 13:05:51 INFO mapreduce.ImportJobBase: Transferred 47.2606 GB in
6,389.4644 seconds (6.7206 MB/sec)
14/05/27 13:05:51 INFO mapreduce.ImportJobBase: Retrieved 98235461 records.
14/05/27 13:05:51 DEBUG util.ClassLoaderStack: Restoring classloader:
sun.misc.Launcher$AppClassLoader@566d0085
14/05/27 13:05:51 DEBUG hive.HiveImport: Hive.inputTable: WAREHOUSE.MY_BIG_TABLE
14/05/27 13:05:51 DEBUG hive.HiveImport: Hive.outputTable:
WAREHOUSE.MY_BIG_TABLE
14/05/27 13:05:51 DEBUG manager.OracleManager: Using column names query: SELECT
t.* FROM WAREHOUSE.MY_BIG_TABLE t WHERE 1=0
14/05/27 13:05:51 DEBUG manager.SqlManager: Execute getColumnTypesRawQuery :
SELECT t.* FROM WAREHOUSE.MY_BIG_TABLE t WHERE 1=0
14/05/27 13:05:51 ERROR manager.SqlManager: Error executing statement:
java.sql.SQLException: ORA-02396: exceeded maximum idle time, please connect
again
java.sql.SQLException: ORA-02396: exceeded maximum idle time, please connect
again
at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:447)
at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:389)
{quote}
With small tables (under 1 hour) everything is fine.
This problems looks exactly as described in this (SQOOP-934) issue.
> JDBC Connection can timeout after import but before hive import
> ---------------------------------------------------------------
>
> Key: SQOOP-934
> URL: https://issues.apache.org/jira/browse/SQOOP-934
> Project: Sqoop
> Issue Type: Improvement
> Affects Versions: 1.4.2
> Reporter: Jarek Jarcec Cecho
> Assignee: Raghav Kumar Gautam
> Fix For: 1.4.4
>
> Attachments: SQOOP-934-2.patch, SQOOP-934.patch
>
>
> Our current [import
> rutine|https://github.com/apache/sqoop/blob/trunk/src/java/org/apache/sqoop/tool/ImportTool.java#L385]
> imports data into HDFS and then tries to do Hive import. As the connection
> to the remote server is opened only once at the begging it might timeout
> during very long mapreduce job. I believe that we should ensure that the
> connection is still valid before performing the hive import.
--
This message was sent by Atlassian JIRA
(v6.2#6252)