Is it by setting --map-column-java email=com.cloudera.sqoop.lib.BlobRef? On Wed, Jul 15, 2015 at 8:11 PM, Suraj Nayak <[email protected]> wrote:
> Hi Abe, > > I think the 1st option suits my use case best. But, what is the method/way > to tell Sqoop to use Large Object Loading? > > Thanks! > > On Wed, Jul 15, 2015 at 7:50 PM, Abraham Elmahrek <[email protected]> > wrote: > >> Hey man, >> >> I think Sqoop does this intentionally: >> https://github.com/apache/sqoop/blob/trunk/src/java/org/apache/sqoop/mapreduce/RawKeyTextOutputFormat.java#L65 >> . >> >> I see two things that may get you beyond this point: >> >> 1. You can try type casting to a BLOB and tell Sqoop to use Large >> Object Loading. >> 2. Try the --as-avrodatafile option and write as Avro. >> >> -Abe >> >> On Wed, Jul 15, 2015 at 7:23 PM, Suraj Nayak <[email protected]> wrote: >> >>> Hi Abe, >>> >>> Thanks for highlighting missing required info quickly. Below are the >>> details: >>> >>> - *Version:* Sqoop 1.4.5 >>> - *Sqoop Command: *sqoop import --connect >>> jdbc:teradata://aa.bb.cc.internal/DATABASE=someDB --username sqoop_usr >>> --password sqoop_usr --table ENCRYPTED_TBL --fields-terminated-by \\001 >>> -m >>> 1 --target-dir /tmp/ENC_TBL --connection-manager >>> "org.apache.sqoop.manager.GenericJdbcManager" --driver >>> com.teradata.jdbc.TeraDriver >>> - *Table structure:* id:varchar, count:int, first_name:binary, >>> email:binary, column5:varchar. >>> >>> Binary is used as the data is encrypted. >>> >>> Thanks! >>> >>> >>> >>> On Wed, Jul 15, 2015 at 6:44 PM, Abraham Elmahrek <[email protected]> >>> wrote: >>> >>>> Hey man, >>>> >>>> Need some details to help: >>>> >>>> - What version of Sqoop? >>>> - Sqoop command? >>>> - Database table structure (preferably a describe on the database) >>>> >>>> -Abe >>>> >>>> On Wed, Jul 15, 2015 at 6:42 PM, Suraj Nayak <[email protected]> wrote: >>>> >>>> > Hi Sqoop Users and Developers, >>>> > >>>> > How can i import a Binary data column in a table into HDFS without >>>> > converting it into String. >>>> > >>>> > I have encrypted data in RDBMS, I need to import this column as is >>>> without >>>> > converting it into string. As of now, Sqoop is typecasting the data >>>> into >>>> > String/text and decryption is failing in Hadoop. >>>> > >>>> > Can someone provide pointers to solve this? Any workaround? >>>> > >>>> > -- >>>> > Thanks >>>> > Suraj Nayak M >>>> > >>>> >>> >>> >>> >>> -- >>> Thanks >>> Suraj Nayak M >>> >> >> > > > -- > Thanks > Suraj Nayak M > -- Thanks Suraj Nayak M
