Hi Mich,

Its recommended to use upper case for table and column name so that you
dont to explicitly quote table and column names.

~Anil



On Sun, Oct 23, 2016 at 9:07 AM, Ravi Kiran <[email protected]>
wrote:

> Sorry, I meant to say table names are case sensitive.
>
> On Sun, Oct 23, 2016 at 9:06 AM, Ravi Kiran <[email protected]>
> wrote:
>
>> Hi Mich,
>>    Apparently, the tables are case sensitive. Since you have enclosed a
>> double quote when creating the table, please pass the same when running the
>> bulk load job.
>>
>> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.3.jar:/usr/lib/hbase/conf
>> hadoop jar phoenix-4.8.1-HBase-1.2-client.jar
>> org.apache.phoenix.mapreduce.CsvBulkLoadTool --table "dummy" --input
>> /data/prices/2016-10-23/prices.1477228923115
>>
>> Regards
>>
>>
>> On Sun, Oct 23, 2016 at 8:39 AM, Mich Talebzadeh <
>> [email protected]> wrote:
>>
>>> Not sure whether phoenix-4.8.1-HBase-1.2-client.jar is the correct jar
>>> file?
>>>
>>> Thanks
>>>
>>> Dr Mich Talebzadeh
>>>
>>>
>>>
>>> LinkedIn * 
>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>
>>>
>>>
>>> http://talebzadehmich.wordpress.com
>>>
>>>
>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>> any loss, damage or destruction of data or any other property which may
>>> arise from relying on this email's technical content is explicitly
>>> disclaimed. The author will in no case be liable for any monetary damages
>>> arising from such loss, damage or destruction.
>>>
>>>
>>>
>>> On 23 October 2016 at 15:39, Mich Talebzadeh <[email protected]>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> My stack
>>>>
>>>> Hbase: hbase-1.2.3
>>>> Phoenix: apache-phoenix-4.8.1-HBase-1.2-bin
>>>>
>>>>
>>>> As a suggestion I tried to load an Hbase file via
>>>> org.apache.phoenix.mapreduce.CsvBulkLoadTool
>>>>
>>>> So
>>>>
>>>> I created a dummy table in Hbase as below
>>>>
>>>> create 'dummy', 'price_info'
>>>>
>>>> Then in Phoenix I created a table on Hbase table
>>>>
>>>>
>>>> create table "dummy" (PK VARCHAR PRIMARY KEY, "price_info"."ticker"
>>>> VARCHAR,"price_info"."timecreated" VARCHAR, "price_info"."price"
>>>> VARCHAR);
>>>>
>>>> And then used the following comman to load the csv file
>>>>
>>>>  
>>>> HADOOP_CLASSPATH=/home/hduser/jars/hbase-protocol-1.2.3.jar:/usr/lib/hbase/conf
>>>> hadoop jar phoenix-4.8.1-HBase-1.2-client.jar
>>>> org.apache.phoenix.mapreduce.CsvBulkLoadTool --table dummy --input
>>>> /data/prices/2016-10-23/prices.1477228923115
>>>>
>>>> However, it does not seem to find the table dummy !
>>>>
>>>> 2016-10-23 14:38:39,442 INFO  [main] metrics.Metrics: Initializing
>>>> metrics system: phoenix
>>>> 2016-10-23 14:38:39,479 INFO  [main] impl.MetricsConfig: loaded
>>>> properties from hadoop-metrics2.properties
>>>> 2016-10-23 14:38:39,529 INFO  [main] impl.MetricsSystemImpl: Scheduled
>>>> snapshot period at 10 second(s).
>>>> 2016-10-23 14:38:39,529 INFO  [main] impl.MetricsSystemImpl: phoenix
>>>> metrics system started
>>>> Exception in thread "main" java.lang.IllegalArgumentException: Table
>>>> DUMMY not found
>>>>         at org.apache.phoenix.util.SchemaUtil.generateColumnInfo(Schema
>>>> Util.java:873)
>>>>         at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.buildImpor
>>>> tColumns(AbstractBulkLoadTool.java:377)
>>>>         at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.loadData(A
>>>> bstractBulkLoadTool.java:214)
>>>>         at org.apache.phoenix.mapreduce.AbstractBulkLoadTool.run(Abstra
>>>> ctBulkLoadTool.java:183)
>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>>>>         at org.apache.phoenix.mapreduce.CsvBulkLoadTool.main(CsvBulkLoa
>>>> dTool.java:101)
>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce
>>>> ssorImpl.java:62)
>>>>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe
>>>> thodAccessorImpl.java:43)
>>>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>>>         at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
>>>>         at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
>>>>
>>>> I tried putting it inside "" etc but no joy I am afraid!
>>>>
>>>> Dr Mich Talebzadeh
>>>>
>>>>
>>>>
>>>> LinkedIn * 
>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>>>
>>>>
>>>>
>>>> http://talebzadehmich.wordpress.com
>>>>
>>>>
>>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>>>> any loss, damage or destruction of data or any other property which may
>>>> arise from relying on this email's technical content is explicitly
>>>> disclaimed. The author will in no case be liable for any monetary damages
>>>> arising from such loss, damage or destruction.
>>>>
>>>>
>>>>
>>>
>>>
>>
>


-- 
Thanks & Regards,
Anil Gupta

Reply via email to