Thanks Richa.

The issue was suppose to be fixed in Hive 0.12 version as per the jira
https://issues.apache.org/jira/browse/HIVE-7164.

Even raised a ticket in sqoop jira [SQOOP-2840] for this .

Thanks
Biswa





On Tue, Mar 1, 2016 at 9:56 AM, Richa Sharma <mailtorichasha...@gmail.com>
wrote:

> Hi,
>
> The values should still persist if partition column data type in Hive is a
> string.
>
> I am checking HCatalog documentation for support of int data type in
> partition column.
>
> Cheers
> Richa
>
> On Tue, Mar 1, 2016 at 3:06 PM, Biswajit Nayak <biswa...@altiscale.com>
> wrote:
>
>> Hi Richa,
>>
>> Thats a work around. But how to handle the columns with INT type.
>> Changing the type will be the last option for me.
>>
>> Regards
>> Biswa
>>
>>
>>
>> On Tue, Mar 1, 2016 at 9:31 AM, Richa Sharma <mailtorichasha...@gmail.com
>> > wrote:
>>
>>> Hi Biswajit
>>>
>>> The answer is in the last line of the error message. Change the data
>>> type of partition column to string in hive and try again.
>>>
>>> Hope it helps !
>>>
>>> Richa
>>>
>>> 16/02/12 08:04:12 ERROR tool.ExportTool: Encountered IOException running 
>>> export job: java.io.IOException: The table provided default.emp_details1 
>>> uses unsupported  partitioning key type  for column salary : int.  Only 
>>> string fields are allowed in partition columns in Catalog
>>>
>>>
>>> On Tue, Mar 1, 2016 at 2:19 PM, Biswajit Nayak <biswa...@altiscale.com>
>>> wrote:
>>>
>>>> Hi All,
>>>>
>>>> I am trying to do a SQOOP export from hive( integer type partition) to
>>>> mysql through HCAT and it fails with the following error.
>>>>
>>>> Versions:-
>>>>
>>>> Hadoop :-  2.7.1
>>>> Hive      :-  1.2.0
>>>> Sqoop   :-  1.4.5
>>>>
>>>> Table in Hive :-
>>>>
>>>>
>>>> hive> use default;
>>>> OK
>>>> Time taken: 0.028 seconds
>>>> hive> describe emp_details1;
>>>> OK
>>>> id                      int
>>>> name                    string
>>>> deg                     string
>>>> dept                    string
>>>> salary                  int
>>>>
>>>> # Partition Information
>>>> # col_name              data_type               comment
>>>>
>>>> salary                  int
>>>> Time taken: 0.125 seconds, Fetched: 10 row(s)
>>>> hive>
>>>>
>>>> hive> select * from emp_details1;
>>>> OK
>>>> 1201    gopal           50000
>>>> 1202    manisha         50000
>>>> 1203    kalil           50000
>>>> 1204    prasanth        50000
>>>> 1205    kranthi         50000
>>>> 1206    satish          50000
>>>> Time taken: 0.195 seconds, Fetched: 6 row(s)
>>>> hive>
>>>>
>>>>
>>>> Conf added to Hive metastore site.xml
>>>>
>>>>
>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$ grep -A5 -B2 -i 
>>>> "hive.metastore.integral.jdo.pushdown" /etc/hive-metastore/hive-site.xml
>>>>     </property>
>>>>     <property>
>>>>         <name>hive.metastore.integral.jdo.pushdown</name>
>>>>         <value>TRUE</value>
>>>>     </property>
>>>>
>>>> </configuration>
>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$
>>>>
>>>>
>>>> The issue remains same
>>>>
>>>>
>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$ /opt/sqoop-1.4.5/bin/sqoop 
>>>> export --connect jdbc:mysql://localhost:3306/test --username hive 
>>>> --password ********* --table employee --hcatalog-database default 
>>>> --hcatalog-table emp_details1
>>>> Warning: /opt/sqoop-1.4.5/bin/../../hbase does not exist! HBase imports 
>>>> will fail.
>>>> Please set $HBASE_HOME to the root of your HBase installation.
>>>> Warning: /opt/sqoop-1.4.5/bin/../../accumulo does not exist! Accumulo 
>>>> imports will fail.
>>>> Please set $ACCUMULO_HOME to the root of your Accumulo installation.
>>>> Warning: /opt/sqoop-1.4.5/bin/../../zookeeper does not exist! Accumulo 
>>>> imports will fail.
>>>> Please set $ZOOKEEPER_HOME to the root of your Zookeeper installation.
>>>> 16/02/12 08:04:00 INFO sqoop.Sqoop: Running Sqoop version: 1.4.5
>>>> 16/02/12 08:04:00 WARN tool.BaseSqoopTool: Setting your password on the 
>>>> command-line is insecure. Consider using -P instead.
>>>> 16/02/12 08:04:00 INFO manager.MySQLManager: Preparing to use a MySQL 
>>>> streaming resultset.
>>>> 16/02/12 08:04:00 INFO tool.CodeGenTool: Beginning code generation
>>>> 16/02/12 08:04:01 INFO manager.SqlManager: Executing SQL statement: SELECT 
>>>> t.* FROM `employee` AS t LIMIT 1
>>>> 16/02/12 08:04:01 INFO manager.SqlManager: Executing SQL statement: SELECT 
>>>> t.* FROM `employee` AS t LIMIT 1
>>>> 16/02/12 08:04:01 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is 
>>>> /opt/hadoop
>>>> Note: 
>>>> /tmp/sqoop-alti-test-01/compile/1b0d4b1c30f167eb57ef488232ab49c8/employee.java
>>>>  uses or overrides a deprecated API.
>>>> Note: Recompile with -Xlint:deprecation for details.
>>>> 16/02/12 08:04:07 INFO orm.CompilationManager: Writing jar file: 
>>>> /tmp/sqoop-alti-test-01/compile/1b0d4b1c30f167eb57ef488232ab49c8/employee.jar
>>>> 16/02/12 08:04:07 INFO mapreduce.ExportJobBase: Beginning export of 
>>>> employee
>>>> 16/02/12 08:04:08 INFO mapreduce.ExportJobBase: Configuring HCatalog for 
>>>> export job
>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Configuring HCatalog 
>>>> specific details for job
>>>> 16/02/12 08:04:08 INFO manager.SqlManager: Executing SQL statement: SELECT 
>>>> t.* FROM `employee` AS t LIMIT 1
>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Database column names 
>>>> projected : [id, name, deg, salary, dept]
>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Database column name - 
>>>> info map :
>>>>     id : [Type : 4,Precision : 11,Scale : 0]
>>>>     name : [Type : 12,Precision : 20,Scale : 0]
>>>>     deg : [Type : 12,Precision : 20,Scale : 0]
>>>>     salary : [Type : 4,Precision : 11,Scale : 0]
>>>>     dept : [Type : 12,Precision : 10,Scale : 0]
>>>>
>>>> 16/02/12 08:04:10 INFO hive.metastore: Trying to connect to metastore with 
>>>> URI thrift://hive-hdpnightly271-ci-91.test.altiscale.com:9083
>>>> 16/02/12 08:04:10 INFO hive.metastore: Connected to metastore.
>>>> 16/02/12 08:04:11 INFO hcat.SqoopHCatUtilities: HCatalog full table schema 
>>>> fields = [id, name, deg, dept, salary]
>>>> 16/02/12 08:04:12 ERROR tool.ExportTool: Encountered IOException running 
>>>> export job: java.io.IOException: The table provided default.emp_details1 
>>>> uses unsupported  partitioning key type  for column salary : int.  Only 
>>>> string fields are allowed in partition columns in Catalog
>>>>
>>>>
>>>> Stuck with this issue. Any one had conquered this before.
>>>>
>>>> Regards
>>>> Biswa
>>>>
>>>>
>>>
>>
>

Reply via email to