Found the resolution
Refer the link below

http://stackoverflow.com/a/34920051/4981746

On 21 January 2016 at 14:45, Divya Gehlot <divya.htco...@gmail.com> wrote:

>  am trying to store data to MySQL using DBStorage in pig script. when I
> run the the script ,I am getting error unable to read the input file.
> Whereas when I try to dump to store the same data file its working fine.
> Sample pigscript :
>
> %default DATABASE_HOST 'localhost';%default DATABASE_NAME 'Test';%default 
> DATABASE_USER 'username';%default DATABASE_PASS 'password';%default 
> DATABASE_DRIVER 'com.mysql.jdbc.Driver';%default DATABASE_TYPE 'mysql';
>
>
>
> A = LOAD '/tmp/TestDivya/Pig/PigSQLTest.txt' using PigStorage() as (name: 
> chararray);
> STORE A into 'test' using 
> org.apache.pig.piggybank.storage.DBStorage('$DATABASE_DRIVER', 
> 'jdbc:$DATABASE_TYPE://$DATABASE_HOST/$DATABASE_NAME',  '$DATABASE_USER', 
> '$DATABASE_PASS', 'INSERT INTO Test(name) VALUES (?)');
>
> HadoopVersion   PigVersion      UserId  StartedAt       FinishedAt      
> Features2.7.1.2.3.4.0-3485      0.15.0.2.3.4.0-3485     hdfs    2016-01-21 
> 01:34:552016-01-21 01:35:07      UNKNOWN
>
> Failed!
>
> Failed Jobs:
> JobId   Alias   Feature Message Outputs
> job_1453263223178_0042  A       MAP_ONLY        Message: Job failed!    
> hdfs://ip-172-31-29-201.ap-southeast-1.compute.internal:8020/user/hdfs/test,
>
> Input(s):
> Failed to read data from "/tmp/TestDivya/Pig/PigSQLTest.txt"
>
> Output(s):
> Failed to produce result in 
> "hdfs://ip-xxx-xx-xx-xxx.ap-xxxxxx-1.compute.internal:8020/user/hdfs/test"
>
> Counters:
> Total records written : 0
> Total bytes written : 0
> Spillable Memory Manager spill count : 0
> Total bags proactively spilled: 0
> Total records proactively spilled: 0
>
> Job DAG:
> job_1453263223178_0042
>
> My cluster is set up on EC2 having Hortonworks HDP 2.3.4 and using the HDP
> MYSQL database for storing the pig output.
>
> Would really appreciate the pointers.
>
> Thanks,
>

Reply via email to