What is the format of the data you are trying to export? The hive table -
is it in textt/parquet or something else?

On Fri, Jul 17, 2015 at 12:25 AM, Kumar Jayapal <kjayapa...@gmail.com>
wrote:

> Hi James,
>
> I tried with uppercase and it worked now I am getting different error. I
> don't see the .metadata dir in the path.
>
> 15/07/16 18:49:12 ERROR sqoop.Sqoop: Got exception running Sqoop:
> org.kitesdk.data.DatasetNotFoundException: Descriptor location does not
> exist:
> hdfs://name/user/hive/warehouse/testing_bi.db/testing_dimension/.metadata
> org.kitesdk.data.DatasetNotFoundException: Descriptor location does not
> exist:
> hdfs://name/user/hive/warehouse/testing_bi.db/testing_dimension/.metadata
>         at
> org.kitesdk.data.spi.filesystem.FileSystemMetadataProvider.checkExists(FileSystemMetadataProvider.java:527)
>         at
> org.kitesdk.data.spi.filesystem.FileSystemMetadataProvider.find(FileSystemMetadataProvider.java:570)
>         at
> org.kitesdk.data.spi.filesystem.FileSystemMetadataProvider.load(FileSystemMetadataProvider.java:112)
>         at
> org.kitesdk.data.spi.filesystem.FileSystemDatasetRepository.load(FileSystemDatasetRepository.java:228)
>         at org.kitesdk.data.Datasets.load(Datasets.java:69)
>         at org.kitesdk.data.Datasets.load(Datasets.java:113)
>         at
> org.kitesdk.data.mapreduce.DatasetKeyInputFormat.load(DatasetKeyInputFormat.java:274)
>         at
> org.kitesdk.data.mapreduce.DatasetKeyInputFormat.setConf(DatasetKeyInputFormat.java:214)
>         at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
>         at
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
>         at
> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:586)
>         at
> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:606)
>         at
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:490)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1295)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1292)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:415)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)
>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1292)
>         at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1313)
>         at
> org.apache.sqoop.mapreduce.ExportJobBase.doSubmitJob(ExportJobBase.java:302)
>         at
> org.apache.sqoop.mapreduce.ExportJobBase.runJob(ExportJobBase.java:279)
>         at
> org.apache.sqoop.mapreduce.ExportJobBase.runExport(ExportJobBase.java:420)
>         at
> org.apache.sqoop.manager.OracleManager.exportTable(OracleManager.java:455)
>         at org.apache.sqoop.tool.ExportTool.exportTable(ExportTool.java:81)
>         at org.apache.sqoop.tool.ExportTool.run(ExportTool.java:100)
>         at org.apache.sqoop.Sqoop.run(Sqoop.java:143)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179)
>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227)
>         at org.apache.sqoop.Sqoop.main(Sqoop.java:236)
>
>
>
>
>
>
>
>
>
> Thanks
> Jay
>
> On Wed, Jul 15, 2015 at 9:12 PM, James Bond <bond.b...@gmail.com> wrote:
>
>> This clearly says its an oracle error. Try checking for the following -
>> 1. If the user "TESTUSER" has Read/write privileges.
>> 2. If the table TEST_DIMESION is the same schema as TESTUSER, if not try
>> appending the schema name to the table - <schema/owner>.TEST_DIMENSION.
>> 3. Make sure you are connecting to the right server and service name.
>>
>> Thanks,
>> Ashwin
>>
>>
>> On Thu, Jul 16, 2015 at 5:05 AM, Kumar Jayapal <kjayapa...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I see the table in my database. When I execute a export command I get
>>> table or view does not exist. When I do list-tables I can see the table.
>>>
>>> Can some one tell me what wrong. I have DBA privileges
>>>
>>> sqoop export --connect "jdbc:oracle:thin:@lorsasa.ss.com:1521/pocd01us"
>>>  --username " TESTUSER" --P --table "TEST_DIMENSION" --export-dir
>>> "/user/hive/warehouse/test.db/test_dimension/"
>>> Warning:
>>> /opt/cloudera/parcels/CDH-5.3.2-1.cdh5.3.2.p654.326/bin/../lib/sqoop/../accumulo
>>> does not exist! Accumulo imports will fail.
>>> Please set $ACCUMULO_HOME to the root of your Accumulo installation.
>>> 15/07/15 23:20:03 INFO sqoop.Sqoop: Running Sqoop version: 1.4.5-cdh5.3.2
>>> Enter password:
>>> 15/07/15 23:20:20 INFO oracle.OraOopManagerFactory: Data Connector for
>>> Oracle and Hadoop is disabled.
>>> 15/07/15 23:20:20 INFO manager.SqlManager: Using default fetchSize of
>>> 1000
>>> 15/07/15 23:20:20 INFO tool.CodeGenTool: Beginning code generation
>>> 15/07/15 23:20:21 INFO manager.OracleManager: Time zone has been set to
>>> GMT
>>> 15/07/15 23:20:21 INFO manager.SqlManager: Executing SQL statement:
>>> SELECT t.* FROM TEST_DIMENSION t WHERE 1=0
>>> 15/07/15 23:20:21 ERROR manager.SqlManager: Error executing statement:
>>> java.sql.SQLSyntaxErrorException: ORA-00942: table or view does not exist
>>>
>>> java.sql.SQLSyntaxErrorException: ORA-00942: table or view does not exist
>>>
>>>         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:447)
>>>         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:396)
>>>         at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:951)
>>>         at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:513)
>>>         at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:227)
>>>         at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:531)
>>>         at
>>> oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:208)
>>>         at
>>> oracle.jdbc.driver.T4CPreparedStatement.executeForDescribe(T4CPreparedStatement.java:886)
>>>         at
>>> oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:1175)
>>>         at
>>> oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1296)
>>>         at
>>> oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3613)
>>>         at
>>> oracle.jdbc.driver.OraclePreparedStatement.executeQuery(OraclePreparedStatement.java:3657)
>>>         at
>>> oracle.jdbc.driver.OraclePreparedStatementWrapper.executeQuery(OraclePreparedStatementWrapper.java:1495)
>>>         at
>>> org.apache.sqoop.manager.SqlManager.execute(SqlManager.java:750)
>>>         at
>>> org.apache.sqoop.manager.SqlManager.execute(SqlManager.java:759)
>>>         at
>>> org.apache.sqoop.manager.SqlManager.getColumnInfoForRawQuery(SqlManager.java:269)
>>>         at
>>> org.apache.sqoop.manager.SqlManager.getColumnTypesForRawQuery(SqlManager.java:240)
>>>         at
>>> org.apache.sqoop.manager.SqlManager.getColumnTypes(SqlManager.java:226)
>>>         at
>>> org.apache.sqoop.manager.ConnManager.getColumnTypes(ConnManager.java:295)
>>>         at
>>> org.apache.sqoop.orm.ClassWriter.getColumnTypes(ClassWriter.java:1833)
>>>         at
>>> org.apache.sqoop.orm.ClassWriter.generate(ClassWriter.java:1645)
>>>         at
>>> org.apache.sqoop.tool.CodeGenTool.generateORM(CodeGenTool.java:96)
>>>         at
>>> org.apache.sqoop.tool.ExportTool.exportTable(ExportTool.java:64)
>>>         at org.apache.sqoop.tool.ExportTool.run(ExportTool.java:100)
>>>         at org.apache.sqoop.Sqoop.run(Sqoop.java:143)
>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>>>         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179)
>>>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
>>>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227)
>>>         at org.apache.sqoop.Sqoop.main(Sqoop.java:236)
>>> 15/07/15 23:20:21 ERROR tool.ExportTool: Encountered IOException running
>>> export job: java.io.IOException: No columns to generate for ClassWriter
>>>
>>>
>>>
>>>
>>>
>>> Thanks
>>> Jay
>>>
>>
>>
>

Reply via email to