Hi Mich, Apparently, the tables are case sensitive. Since you have enclosed a double quote when creating the table, please pass the same when running the bulk load job.
HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.3.jar:/usr/lib/hbase/conf hadoop jar phoenix-4.8.1-HBase-1.2-client.jar org.apache.phoenix.mapreduce.C svBulkLoadTool --table "dummy" --input /data/prices/2016-10-23/prices .1477228923115 Regards On Sun, Oct 23, 2016 at 8:39 AM, Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > Not sure whether phoenix-4.8.1-HBase-1.2-client.jar is the correct jar > file? > > Thanks > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > *Disclaimer:* Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. > The author will in no case be liable for any monetary damages arising from > such loss, damage or destruction. > > > > On 23 October 2016 at 15:39, Mich Talebzadeh <mich.talebza...@gmail.com> > wrote: > >> Hi, >> >> My stack >> >> Hbase: hbase-1.2.3 >> Phoenix: apache-phoenix-4.8.1-HBase-1.2-bin >> >> >> As a suggestion I tried to load an Hbase file via >> org.apache.phoenix.mapreduce.CsvBulkLoadTool >> >> So >> >> I created a dummy table in Hbase as below >> >> create 'dummy', 'price_info' >> >> Then in Phoenix I created a table on Hbase table >> >> >> create table "dummy" (PK VARCHAR PRIMARY KEY, "price_info"."ticker" >> VARCHAR,"price_info"."timecreated" VARCHAR, "price_info"."price" >> VARCHAR); >> >> And then used the following comman to load the csv file >> >> >> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.3.jar:/usr/lib/hbase/conf >> hadoop jar phoenix-4.8.1-HBase-1.2-client.jar >> org.apache.phoenix.mapreduce.CsvBulkLoadTool --table dummy --input >> /data/prices/2016-10-23/prices.1477228923115 >> >> However, it does not seem to find the table dummy ! >> >> 2016-10-23 14:38:39,442 INFO [main] metrics.Metrics: Initializing >> metrics system: phoenix >> 2016-10-23 14:38:39,479 INFO [main] impl.MetricsConfig: loaded >> properties from hadoop-metrics2.properties >> 2016-10-23 14:38:39,529 INFO [main] impl.MetricsSystemImpl: Scheduled >> snapshot period at 10 second(s). >> 2016-10-23 14:38:39,529 INFO [main] impl.MetricsSystemImpl: phoenix >> metrics system started >> Exception in thread "main" java.lang.IllegalArgumentException: Table >> DUMMY not found >> at org.apache.phoenix.util.SchemaUtil.generateColumnInfo(Schema >> Util.java:873) >> at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.buildImpor >> tColumns(AbstractBulkLoadTool.java:377) >> at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(A >> bstractBulkLoadTool.java:214) >> at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(Abstra >> ctBulkLoadTool.java:183) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) >> at org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoa >> dTool.java:101) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >> ssorImpl.java:62) >> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >> thodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) >> at org.apache.hadoop.util.RunJar.run(RunJar.java:221) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:136) >> >> I tried putting it inside "" etc but no joy I am afraid! >> >> Dr Mich Talebzadeh >> >> >> >> LinkedIn * >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >> >> >> >> http://talebzadehmich.wordpress.com >> >> >> *Disclaimer:* Use it at your own risk. Any and all responsibility for >> any loss, damage or destruction of data or any other property which may >> arise from relying on this email's technical content is explicitly >> disclaimed. The author will in no case be liable for any monetary damages >> arising from such loss, damage or destruction. >> >> >> > >