[jira] [Updated] (PHOENIX-4895) NoClassDefFound when use IndexTool create async index

2018-09-11 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

张延召 updated PHOENIX-4895:
-
Description: 
First I created a table

^CREATE TABLE TMP_TEST(^
 ^ID VARCHAR NOT NULL PRIMARY KEY,^
 ^NAME VARCHAR,^
 ^ADDR VARCHAR,^
 ^AGE BIGINT DEFAULT 10^
 ^);^

Then I created the asynchronous index table

^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^

Finally, perform the MapReduce Task

^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
--data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^

But I received an error message

^Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/commons/cli/DefaultParser^
 ^at 
org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
 ^Caused by: java.lang.ClassNotFoundException: 
org.apache.commons.cli.DefaultParser^
 ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
 ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
 ^... 5 more^

Please give me some advice,Thanks

 

  was:
First I created a table

^CREATE TABLE TMP_TEST(^
 ^ID VARCHAR NOT NULL PRIMARY KEY,^
 ^NAME VARCHAR,^
 ^ADDR VARCHAR,^
 ^AGE BIGINT DEFAULT 10^
 ^);^

Then I created the asynchronous index table

^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^

Finally, perform the MapReduce Task

^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
--data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^

But I received an error message

^Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/commons/cli/DefaultParser^
 ^at 
org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
^Caused by: java.lang.ClassNotFoundException: 
org.apache.commons.cli.DefaultParser^
 ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
 ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
 ^... 5 more^

Please give me some advice,Thanks

 


> NoClassDefFound when use IndexTool create async index
> -
>
> Key: PHOENIX-4895
> URL: https://issues.apache.org/jira/browse/PHOENIX-4895
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 5.0.0
> Environment: HDP :3.0.0
> HBase :2.0.0
> Phoenix : 5.0.0
> Hadoop : 3.1.0
>Reporter: 张延召
>Priority: Major
>  Labels: 5.0.0, ASYNC, INDEX
> Fix For: 5.1.0
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> First I created a table
> ^CREATE TABLE TMP_TEST(^
>  ^ID VARCHAR NOT NULL PRIMARY KEY,^
>  ^NAME VARCHAR,^
>  ^ADDR VARCHAR,^
>  ^AGE BIGINT DEFAULT 10^
>  ^);^
> Then I created the asynchronous index table
> ^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^
> Finally, perform the MapReduce Task
> ^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
> --data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^
> But I received an error message
> ^Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/commons/cli/DefaultParser^
>  ^at 
> org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
>  ^Caused by: java.lang.ClassNotFoundException: 
> org.apache.commons.cli.DefaultParser^
>  ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
>  ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
>  ^... 5 more^
> Please give me some advice,Thanks
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (PHOENIX-4895) NoClassDefFound when use IndexTool create async index

2018-09-11 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

张延召 updated PHOENIX-4895:
-
Description: 
*First I created a table:*

^CREATE TABLE TMP_TEST(^
 ^ID VARCHAR NOT NULL PRIMARY KEY,^
 ^NAME VARCHAR,^
 ^ADDR VARCHAR,^
 ^AGE BIGINT DEFAULT 10^
 ^);^

*Then I created the asynchronous index table:*

^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^

*Finally, perform the MapReduce Task:*

^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
--data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^

*But I received an error message:*

^Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/commons/cli/DefaultParser^
 ^at 
org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
 ^Caused by: java.lang.ClassNotFoundException: 
org.apache.commons.cli.DefaultParser^
 ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
 ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
 ^... 5 more^

*Please give me some advice,Thanks!*

 

  was:
First I created a table

^CREATE TABLE TMP_TEST(^
 ^ID VARCHAR NOT NULL PRIMARY KEY,^
 ^NAME VARCHAR,^
 ^ADDR VARCHAR,^
 ^AGE BIGINT DEFAULT 10^
 ^);^

Then I created the asynchronous index table

^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^

Finally, perform the MapReduce Task

^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
--data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^

But I received an error message

^Exception in thread "main" java.lang.NoClassDefFoundError: 
org/apache/commons/cli/DefaultParser^
 ^at 
org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
 ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
 ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
 ^Caused by: java.lang.ClassNotFoundException: 
org.apache.commons.cli.DefaultParser^
 ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
 ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
 ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
 ^... 5 more^

Please give me some advice,Thanks

 


> NoClassDefFound when use IndexTool create async index
> -
>
> Key: PHOENIX-4895
> URL: https://issues.apache.org/jira/browse/PHOENIX-4895
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 5.0.0
> Environment: HDP :3.0.0
> HBase :2.0.0
> Phoenix : 5.0.0
> Hadoop : 3.1.0
>Reporter: 张延召
>Priority: Major
>  Labels: 5.0.0, ASYNC, INDEX
> Fix For: 5.1.0
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> *First I created a table:*
> ^CREATE TABLE TMP_TEST(^
>  ^ID VARCHAR NOT NULL PRIMARY KEY,^
>  ^NAME VARCHAR,^
>  ^ADDR VARCHAR,^
>  ^AGE BIGINT DEFAULT 10^
>  ^);^
> *Then I created the asynchronous index table:*
> ^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^
> *Finally, perform the MapReduce Task:*
> ^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
> --data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^
> *But I received an error message:*
> ^Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/commons/cli/DefaultParser^
>  ^at 
> org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
>  ^Caused by: java.lang.ClassNotFoundException: 
> org.apache.commons.cli.DefaultParser^
>  ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
>  ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
>  ^... 5 more^
> *Please give me some advice,Thanks!*
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (PHOENIX-4897) Phoenix Kafka Plugin support specifying message offset

2018-09-11 Thread XiaoMifan (JIRA)
XiaoMifan created PHOENIX-4897:
--

 Summary: Phoenix Kafka Plugin support specifying message offset
 Key: PHOENIX-4897
 URL: https://issues.apache.org/jira/browse/PHOENIX-4897
 Project: Phoenix
  Issue Type: New Feature
Affects Versions: 4.13.0
Reporter: XiaoMifan


I found Kafka Plugin only consume the new messages after this plugin started, 
could not consume the message from beginning (at least I didn't found 
introduction on Phonenix docs).

And if the plugin app crashed, util it got restarted, the message happened 
during this period are not got ingested.

If I'm missing some thing that this plugin has already supported it, please 
give some advise, thanks.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (PHOENIX-4895) NoClassDefFound when use IndexTool create async index

2018-09-11 Thread Ankit Singhal (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ankit Singhal resolved PHOENIX-4895.

Resolution: Duplicate

> NoClassDefFound when use IndexTool create async index
> -
>
> Key: PHOENIX-4895
> URL: https://issues.apache.org/jira/browse/PHOENIX-4895
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 5.0.0
> Environment: HDP :3.0.0
> HBase :2.0.0
> Phoenix : 5.0.0
> Hadoop : 3.1.0
>Reporter: 张延召
>Priority: Major
>  Labels: 5.0.0, ASYNC, INDEX
> Fix For: 5.1.0
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> *First I created a table:*
> ^CREATE TABLE TMP_TEST(^
>  ^ID VARCHAR NOT NULL PRIMARY KEY,^
>  ^NAME VARCHAR,^
>  ^ADDR VARCHAR,^
>  ^AGE BIGINT DEFAULT 10^
>  ^);^
> *Then I created the asynchronous index table:*
> ^CREATE INDEX ASYNC_IDX ON TMP_TEST (NAME, ADDR) INCLUDE (AGE) ASYNC;^
> *Finally, perform the MapReduce Task:*
> ^./hbase org.apache.phoenix.mapreduce.index.IndexTool --schema default 
> --data-table TMP_TEST --index-table ASYNC_IDX --output-path ASYNC_IDX_HFILES^
> *But I received an error message:*
> ^Exception in thread "main" java.lang.NoClassDefFoundError: 
> org/apache/commons/cli/DefaultParser^
>  ^at 
> org.apache.phoenix.mapreduce.index.IndexTool.parseOptions(IndexTool.java:183)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.run(IndexTool.java:522)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)^
>  ^at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)^
>  ^at org.apache.phoenix.mapreduce.index.IndexTool.main(IndexTool.java:769)^
>  ^Caused by: java.lang.ClassNotFoundException: 
> org.apache.commons.cli.DefaultParser^
>  ^at java.net.URLClassLoader.findClass(URLClassLoader.java:381)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:424)^
>  ^at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:335)^
>  ^at java.lang.ClassLoader.loadClass(ClassLoader.java:357)^
>  ^... 5 more^
> *Please give me some advice,Thanks!*
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (PHOENIX-4898) LAST_VALUE aggregation fails when column contain NULLs

2018-09-11 Thread Abhishek Gupta (JIRA)
Abhishek Gupta created PHOENIX-4898:
---

 Summary: LAST_VALUE aggregation fails when column contain NULLs
 Key: PHOENIX-4898
 URL: https://issues.apache.org/jira/browse/PHOENIX-4898
 Project: Phoenix
  Issue Type: Bug
Affects Versions: 4.7.0
 Environment: HBase version: 1.1.2

Phoenix version: 4.7.0

 
Reporter: Abhishek Gupta


When I use last_value aggregation on a column that is sparse with NULLs it 
fails with the exception 
{code:java}
Error: Error -1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) (state=0,code=-1) 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaSqlException: Error 
-1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:54)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:41)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.execute(AvaticaStatement.java:208)
 at sqlline.Commands.execute(Commands.java:822) at 
sqlline.Commands.sql(Commands.java:732) at 
sqlline.SqlLine.dispatch(SqlLine.java:808) at 
sqlline.SqlLine.begin(SqlLine.java:681) at 
sqlline.SqlLine.start(SqlLine.java:398) at 
sqlline.SqlLine.main(SqlLine.java:292) at 
org.apache.phoenix.queryserver.client.SqllineWrapper.main(SqllineWrapper.java:93)
 java.lang.ArrayIndexOutOfBoundsException
{code}
Please find below the DDL and SQL queries to reproduce the issue

 
{code:java}
create table if not exists test_last_value (pk1 varchar not null, m1 bigint,m2 
bigint,m3 bigint, constraint test_last_value_pk primary key (pk1)); 
upsert into test_last_value(pk1,m1,m2) values('20180901',10,20); 
upsert into test_last_value(pk1,m1,m2) values('201809010030',10,20); 
upsert into test_last_value(pk1,m1,m2,m3) values('201809010100',10,20,11); 
upsert into test_last_value(pk1,m1,m2) values('201809010130',10,20); 
upsert into test_last_value(pk1,m1,m2) values('201809010200',10,20); 
upsert into test_last_value(pk1,m1,m2) values('201809010230',10,20); 
upsert into test_last_value(pk1,m1,m3) values('201809010300',10,22);
{code}
 
{noformat}
select * from test_last_value; 
+---+-+---+---+
|      PK1      | M1  |  M2   |  M3   |
+---+-+---+---+
| 20180901  | 10  | 20    | null  |
| 201809010030  | 10  | 20    | null  |
| 201809010100  | 10  | 20    | 11    |
| 201809010130  | 10  | 20    | null  |
| 201809010200  | 10  | 20    | null  |
| 201809010230  | 10  | 20    | null  |
| 201809010300  | 10  | null  | 22    |
+---+-+---+---+
Last value aggregation succeeds when column has no nulls 
 
select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m1) 
within group (order by pk1 asc) as m1_last, sum(m2) as m2_sum from 
test_last_value group by t;
+-+--+-+
|      T      | M1_LAST  | M2_SUM  |
+-+--+-+
| 2018090100  | 10       | 40      |
| 2018090101  | 10       | 40      |
| 2018090102  | 10       | 40      |
| 2018090103  | 10       | null    |
+-+--+-+
Last value aggregation fails with java.lang.ArrayIndexOutOfBoundsException when 
column has nulls 
==
select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) 
within group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from 
test_last_value group by t; 
java.lang.ArrayIndexOutOfBoundsException
{noformat}
 

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (PHOENIX-4898) LAST_VALUE aggregation fails when column contain NULLs

2018-09-11 Thread Abhishek Gupta (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Gupta updated PHOENIX-4898:

Labels: Aggregation UDF udfs  (was: UDF udfs)

> LAST_VALUE aggregation fails when column contain NULLs
> --
>
> Key: PHOENIX-4898
> URL: https://issues.apache.org/jira/browse/PHOENIX-4898
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.7.0
> Environment: HBase version: 1.1.2
> Phoenix version: 4.7.0
>  
>Reporter: Abhishek Gupta
>Priority: Major
>  Labels: Aggregation, UDF, udfs
>
> When I use last_value aggregation on a column that is sparse with NULLs it 
> fails with the exception 
> {code:java}
> Error: Error -1 (0) : Error while executing SQL "select 
> TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
> group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
> group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
> exception message) (state=0,code=-1) 
> org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaSqlException: 
> Error -1 (0) : Error while executing SQL "select 
> TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
> group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
> group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
> exception message) at 
> org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:54)
>  at 
> org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:41)
>  at 
> org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
>  at 
> org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.execute(AvaticaStatement.java:208)
>  at sqlline.Commands.execute(Commands.java:822) at 
> sqlline.Commands.sql(Commands.java:732) at 
> sqlline.SqlLine.dispatch(SqlLine.java:808) at 
> sqlline.SqlLine.begin(SqlLine.java:681) at 
> sqlline.SqlLine.start(SqlLine.java:398) at 
> sqlline.SqlLine.main(SqlLine.java:292) at 
> org.apache.phoenix.queryserver.client.SqllineWrapper.main(SqllineWrapper.java:93)
>  java.lang.ArrayIndexOutOfBoundsException
> {code}
> Please find below the DDL and SQL queries to reproduce the issue
>  
> {code:java}
> create table if not exists test_last_value (pk1 varchar not null, m1 
> bigint,m2 bigint,m3 bigint, constraint test_last_value_pk primary key (pk1)); 
> upsert into test_last_value(pk1,m1,m2) values('20180901',10,20); 
> upsert into test_last_value(pk1,m1,m2) values('201809010030',10,20); 
> upsert into test_last_value(pk1,m1,m2,m3) values('201809010100',10,20,11); 
> upsert into test_last_value(pk1,m1,m2) values('201809010130',10,20); 
> upsert into test_last_value(pk1,m1,m2) values('201809010200',10,20); 
> upsert into test_last_value(pk1,m1,m2) values('201809010230',10,20); 
> upsert into test_last_value(pk1,m1,m3) values('201809010300',10,22);
> {code}
>  
> {noformat}
> select * from test_last_value; 
> +---+-+---+---+
> |      PK1      | M1  |  M2   |  M3   |
> +---+-+---+---+
> | 20180901  | 10  | 20    | null  |
> | 201809010030  | 10  | 20    | null  |
> | 201809010100  | 10  | 20    | 11    |
> | 201809010130  | 10  | 20    | null  |
> | 201809010200  | 10  | 20    | null  |
> | 201809010230  | 10  | 20    | null  |
> | 201809010300  | 10  | null  | 22    |
> +---+-+---+---+
> Last value aggregation succeeds when column has no nulls 
>  
> select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m1) 
> within group (order by pk1 asc) as m1_last, sum(m2) as m2_sum from 
> test_last_value group by t;
> +-+--+-+
> |      T      | M1_LAST  | M2_SUM  |
> +-+--+-+
> | 2018090100  | 10       | 40      |
> | 2018090101  | 10       | 40      |
> | 2018090102  | 10       | 40      |
> | 2018090103  | 10       | null    |
> +-+--+-+
> Last value aggregation fails with java.lang.ArrayIndexOutOfBoundsException 
> when column has nulls 
> ==
> select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) 
> within group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from 
> test_last_value group by t; 
> java.lang.ArrayIndexOutOfBoundsException
> {noformat}
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (PHOENIX-4899) Incorrect behaviour of LAST_VALUE aggregation UDF

2018-09-11 Thread Abhishek Gupta (JIRA)
Abhishek Gupta created PHOENIX-4899:
---

 Summary: Incorrect behaviour of LAST_VALUE aggregation UDF
 Key: PHOENIX-4899
 URL: https://issues.apache.org/jira/browse/PHOENIX-4899
 Project: Phoenix
  Issue Type: Bug
Affects Versions: 4.7.0
 Environment: Phoenix version: 4.7.0

HBase version: 1.1.2
Reporter: Abhishek Gupta


 
The LAST_VALUE aggregation UDF doesn't give correct result when group by 
columns does not include all primary key columns. 
The function works in a way that it ignores all the group by columns except the 
column in the order by clause in finding the last value. For example
 
Following is a dataset (pk1,pk2,pk3,pk4 and pk5 are Primary key columns and 
m1-m3 are metric columns, I intend to do a SUM on m2 and m3 and a last value on 
m1 
 
{quote}pk1 .   | pk2                       | pk3 .    | pk4            | pk5 .  
 | m1         | m2 | m3
--
| test    | 20180801    | app1    | plat1          | lang1  | 1            
| 10  | 100  |
| test    | 20180801    | app1    | plat1          | lang2  | 2            
| 10  | 100  |
| test    | 20180801    | app1    | plat2          | lang1  | 3            
| 10  | 100  |
| test    | 20180801    | app2    | plat1          | lang1  | 4            
| 10  | 100  |
| test    | 201808010030    | app1    | plat1          | lang1  | 10           
| 10  | 100  |
| test    | 201808010030    | app1    | plat1          | lang2  | 20           
| 10  | 100  |
| test    | 201808010030    | app1    | plat2          | lang1  | 30           
| 10  | 100  |
| test    | 201808010030    | app2    | plat1          | lang1  | 40           
| 10  | 100  |
| test    | 201808010100    | app1    | plat1          | lang1  | 100          
| 10  | 100  |
| test    | 201808010100    | app1    | plat1          | lang2  | 200          
| 10  | 100  |
| test    | 201808010100    | app1    | plat2          | lang1  | 300          
| 10  | 100  |
| test    | 201808010100    | app2    | plat1          | lang1  | 400          
| 10  | 100  |
| test    | 201808010130    | app1    | plat1          | lang1  | 1000         
| 10  | 100  |
| test    | 201808010130    | app1    | plat1          | lang2  | 2000         
| 10  | 100  |
| test    | 201808010130    | app1    | plat2          | lang1  | 3000         
| 10  | 100  |
| test    | 201808010130    | app2    | plat1          | lang1  | 4000         
| 10  | 100  |
| test    | 201808010200    | app1    | plat1          | lang1  | 1        
| 10  | 100  |
| test    | 201808010200    | app1    | plat1          | lang2  | 2        
| 10  | 100  |
| test    | 201808010200    | app1    | plat2          | lang1  | 3        
| 10  | 100  |
| test    | 201808010200    | app2    | plat1          | lang1  | 4        
| 10  | 100  |{quote}
 
If I run the following query (using all primary key columns in group by)
 
{quote}select TO_CHAR(TO_DATE(pk2,'MMddHHmm'),'MMddHH') as 
t,pk3,pk4,pk5, last_value(m1) within group (order by pk2 asc) as oi, sum(m2), 
sum(m3) from test_table group by pk1,t,pk3,pk4,pk5;{quote}
 
I get the correct result for last value as for each pk2 value which is used in 
order by there is only 1 row
{quote}+-+-++++--+--+
|      T      | pk3  | pk4  |  pk5  |   OI   | SUM(M2)  | SUM(M3)  |
+-+-++++--+--+
| 2018080100  | app1    | plat1          | lang1  | 10     | 20       | 200     
 |
| 2018080100  | app1    | plat1          | lang2  | 20     | 20       | 200     
 |
| 2018080100  | app1    | plat2          | lang1  | 30     | 20       | 200     
 |
| 2018080100  | app2    | plat1          | lang1  | 40     | 20       | 200     
 |
| 2018080101  | app1    | plat1          | lang1  | 1000   | 20       | 200     
 |
| 2018080101  | app1    | plat1          | lang2  | 2000   | 20       | 200     
 |
| 2018080101  | app1    | plat2          | lang1  | 3000   | 20       | 200     
 |
| 2018080101  | app2    | plat1          | lang1  | 4000   | 20       | 200     
 |
| 2018080102  | app1    | plat1          | lang1  | 1  | 20       | 100     
 |
| 2018080102  | app1    | plat1          | lang2  | 2  | 10       | 100     
 |
| 2018080102  | app1    | plat2          | lang1  | 3  | 10       | 100     
 |
| 2018080102  | app2    | plat1          | lang1  | 4  | 10       | 100     
 |
+-+-++++--+--+{quote}
 
However if I do I group by on less than all the primary columns the LAST_VALUE 
function ignores the rest of the group by columns in sorting and returns 
incorrect last_value
 
{quote}select TO_CHAR(TO_DATE(pk2,'MMddHHmm'),'MMddHH') as t, pk3, 
last

[jira] [Updated] (PHOENIX-4899) Incorrect behaviour of LAST_VALUE aggregation UDF

2018-09-11 Thread Abhishek Gupta (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Gupta updated PHOENIX-4899:

Labels: Aggregation UDF aggregator udf udfs  (was: )

> Incorrect behaviour of LAST_VALUE aggregation UDF
> -
>
> Key: PHOENIX-4899
> URL: https://issues.apache.org/jira/browse/PHOENIX-4899
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.7.0
> Environment: Phoenix version: 4.7.0
> HBase version: 1.1.2
>Reporter: Abhishek Gupta
>Priority: Critical
>  Labels: Aggregation, UDF, aggregator, udf, udfs
>
>  
> The LAST_VALUE aggregation UDF doesn't give correct result when group by 
> columns does not include all primary key columns. 
> The function works in a way that it ignores all the group by columns except 
> the column in the order by clause in finding the last value. For example
>  
> Following is a dataset (pk1,pk2,pk3,pk4 and pk5 are Primary key columns and 
> m1-m3 are metric columns, I intend to do a SUM on m2 and m3 and a last value 
> on m1 
>  
> {quote}pk1 .   | pk2                       | pk3 .    | pk4            | pk5 
> .   | m1         | m2 | m3
> --
> | test    | 20180801    | app1    | plat1          | lang1  | 1           
>  | 10  | 100  |
> | test    | 20180801    | app1    | plat1          | lang2  | 2           
>  | 10  | 100  |
> | test    | 20180801    | app1    | plat2          | lang1  | 3           
>  | 10  | 100  |
> | test    | 20180801    | app2    | plat1          | lang1  | 4           
>  | 10  | 100  |
> | test    | 201808010030    | app1    | plat1          | lang1  | 10          
>  | 10  | 100  |
> | test    | 201808010030    | app1    | plat1          | lang2  | 20          
>  | 10  | 100  |
> | test    | 201808010030    | app1    | plat2          | lang1  | 30          
>  | 10  | 100  |
> | test    | 201808010030    | app2    | plat1          | lang1  | 40          
>  | 10  | 100  |
> | test    | 201808010100    | app1    | plat1          | lang1  | 100         
>  | 10  | 100  |
> | test    | 201808010100    | app1    | plat1          | lang2  | 200         
>  | 10  | 100  |
> | test    | 201808010100    | app1    | plat2          | lang1  | 300         
>  | 10  | 100  |
> | test    | 201808010100    | app2    | plat1          | lang1  | 400         
>  | 10  | 100  |
> | test    | 201808010130    | app1    | plat1          | lang1  | 1000        
>  | 10  | 100  |
> | test    | 201808010130    | app1    | plat1          | lang2  | 2000        
>  | 10  | 100  |
> | test    | 201808010130    | app1    | plat2          | lang1  | 3000        
>  | 10  | 100  |
> | test    | 201808010130    | app2    | plat1          | lang1  | 4000        
>  | 10  | 100  |
> | test    | 201808010200    | app1    | plat1          | lang1  | 1       
>  | 10  | 100  |
> | test    | 201808010200    | app1    | plat1          | lang2  | 2       
>  | 10  | 100  |
> | test    | 201808010200    | app1    | plat2          | lang1  | 3       
>  | 10  | 100  |
> | test    | 201808010200    | app2    | plat1          | lang1  | 4       
>  | 10  | 100  |{quote}
>  
> If I run the following query (using all primary key columns in group by)
>  
> {quote}select TO_CHAR(TO_DATE(pk2,'MMddHHmm'),'MMddHH') as 
> t,pk3,pk4,pk5, last_value(m1) within group (order by pk2 asc) as oi, sum(m2), 
> sum(m3) from test_table group by pk1,t,pk3,pk4,pk5;{quote}
>  
> I get the correct result for last value as for each pk2 value which is used 
> in order by there is only 1 row
> {quote}+-+-++++--+--+
> |      T      | pk3  | pk4  |  pk5  |   OI   | SUM(M2)  | SUM(M3)  |
> +-+-++++--+--+
> | 2018080100  | app1    | plat1          | lang1  | 10     | 20       | 200   
>    |
> | 2018080100  | app1    | plat1          | lang2  | 20     | 20       | 200   
>    |
> | 2018080100  | app1    | plat2          | lang1  | 30     | 20       | 200   
>    |
> | 2018080100  | app2    | plat1          | lang1  | 40     | 20       | 200   
>    |
> | 2018080101  | app1    | plat1          | lang1  | 1000   | 20       | 200   
>    |
> | 2018080101  | app1    | plat1          | lang2  | 2000   | 20       | 200   
>    |
> | 2018080101  | app1    | plat2          | lang1  | 3000   | 20       | 200   
>    |
> | 2018080101  | app2    | plat1          | lang1  | 4000   | 20       | 200   
>    |
> | 2018080102  | app1    | plat1          | lang1  | 1  | 20       | 100   
>    |
> | 2018080102  | app1    | plat1          | lang2  | 2  | 10       | 100   
>    |
> | 2018080102  | app1   

[jira] [Updated] (PHOENIX-4898) LAST_VALUE aggregation fails when column contain NULLs

2018-09-11 Thread Abhishek Gupta (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Abhishek Gupta updated PHOENIX-4898:

Description: 
When I use last_value aggregation on a column that is sparse with NULLs it 
fails with the exception 
{code:java}
Error: Error -1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) (state=0,code=-1) 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaSqlException: Error 
-1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:54)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(Helper.java:41)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.executeInternal(AvaticaStatement.java:156)
 at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaStatement.execute(AvaticaStatement.java:208)
 at sqlline.Commands.execute(Commands.java:822) at 
sqlline.Commands.sql(Commands.java:732) at 
sqlline.SqlLine.dispatch(SqlLine.java:808) at 
sqlline.SqlLine.begin(SqlLine.java:681) at 
sqlline.SqlLine.start(SqlLine.java:398) at 
sqlline.SqlLine.main(SqlLine.java:292) at 
org.apache.phoenix.queryserver.client.SqllineWrapper.main(SqllineWrapper.java:93)
 java.lang.ArrayIndexOutOfBoundsException
{code}
Please find below the DDL and SQL queries to reproduce the issue

 
{code:java}
create table if not exists test_last_value (pk1 varchar not null, m1 bigint,m2 
bigint,m3 bigint, constraint test_last_value_pk primary key (pk1)); 

upsert into test_last_value(pk1,m1,m2) values('20180901',10,20);
upsert into test_last_value(pk1,m1,m2) values('201809010030',10,20);
upsert into test_last_value(pk1,m1,m2,m3) values('201809010100',10,20,11);
upsert into test_last_value(pk1,m1,m2) values('201809010130',10,20); 
upsert into test_last_value(pk1,m1,m2) values('201809010200',10,20); 
upsert into test_last_value(pk1,m1,m2) values('201809010230',10,20); 
upsert into test_last_value(pk1,m1,m3) values('201809010300',10,22);
{code}
 
{noformat}
select * from test_last_value; 
+---+-+---+---+
|      PK1      | M1  |  M2   |  M3   |
+---+-+---+---+
| 20180901  | 10  | 20    | null  |
| 201809010030  | 10  | 20    | null  |
| 201809010100  | 10  | 20    | 11    |
| 201809010130  | 10  | 20    | null  |
| 201809010200  | 10  | 20    | null  |
| 201809010230  | 10  | 20    | null  |
| 201809010300  | 10  | null  | 22    |
+---+-+---+---+
Last value aggregation succeeds when column has no nulls 
 
select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m1) 
within group (order by pk1 asc) as m1_last, sum(m2) as m2_sum from 
test_last_value group by t;
+-+--+-+
|      T      | M1_LAST  | M2_SUM  |
+-+--+-+
| 2018090100  | 10       | 40      |
| 2018090101  | 10       | 40      |
| 2018090102  | 10       | 40      |
| 2018090103  | 10       | null    |
+-+--+-+
Last value aggregation fails with java.lang.ArrayIndexOutOfBoundsException when 
column has nulls 
==
select TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) 
within group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from 
test_last_value group by t; 
java.lang.ArrayIndexOutOfBoundsException
{noformat}
 

 

 

  was:
When I use last_value aggregation on a column that is sparse with NULLs it 
fails with the exception 
{code:java}
Error: Error -1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) (state=0,code=-1) 
org.apache.phoenix.shaded.org.apache.calcite.avatica.AvaticaSqlException: Error 
-1 (0) : Error while executing SQL "select 
TO_CHAR(TO_DATE(pk1,'MMddHHmm'),'MMddHH') as t, last_value(m3) within 
group (order by pk1 asc) as m3_last, sum(m2) as m2_sum from test_last_value 
group by t": Remote driver error: ArrayIndexOutOfBoundsException: (null 
exception message) at 
org.apache.phoenix.shaded.org.apache.calcite.avatica.Helper.createException(H

Speakers needed for Apache DC Roadshow

2018-09-11 Thread Rich Bowen
We need your help to make the Apache Washington DC Roadshow on Dec 4th a 
success.


What do we need most? Speakers!

We're bringing a unique DC flavor to this event by mixing Open Source 
Software with talks about Apache projects as well as OSS CyberSecurity, 
OSS in Government and and OSS Career advice.


Please take a look at: http://www.apachecon.com/usroadshow18/

(Note: You are receiving this message because you are subscribed to one 
or more mailing lists at The Apache Software Foundation.)


Rich, for the ApacheCon Planners

--
rbo...@apache.org
http://apachecon.com
@ApacheCon


[jira] [Created] (PHOENIX-4900) Modify MAX_MUTATION_SIZE_EXCEEDED and MAX_MUTATION_SIZE_BYTES_EXCEEDED exception message to recommend turning autocommit off for deletes

2018-09-11 Thread Thomas D'Silva (JIRA)
Thomas D'Silva created PHOENIX-4900:
---

 Summary: Modify MAX_MUTATION_SIZE_EXCEEDED and 
MAX_MUTATION_SIZE_BYTES_EXCEEDED exception message to recommend turning 
autocommit off for deletes
 Key: PHOENIX-4900
 URL: https://issues.apache.org/jira/browse/PHOENIX-4900
 Project: Phoenix
  Issue Type: Improvement
Reporter: Thomas D'Silva






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (PHOENIX-4874) psql doesn't support date/time with values smaller than milliseconds

2018-09-11 Thread Rajeshbabu Chintaguntla (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajeshbabu Chintaguntla updated PHOENIX-4874:
-
Attachment: PHOENIX-4874.patch

> psql doesn't support date/time with values smaller than milliseconds
> 
>
> Key: PHOENIX-4874
> URL: https://issues.apache.org/jira/browse/PHOENIX-4874
> Project: Phoenix
>  Issue Type: Task
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Attachments: PHOENIX-4874.patch
>
>
> [https://phoenix.apache.org/tuning.html] lacks entries for 
> phoenix.query.timeFormat, phoenix.query.timestampFormat which are used by 
> psql to parse out TIME and TIMESTAMP data types.
> Add them.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (PHOENIX-4874) psql doesn't support date/time with values smaller than milliseconds

2018-09-11 Thread Rajeshbabu Chintaguntla (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4874?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajeshbabu Chintaguntla reassigned PHOENIX-4874:


Assignee: Rajeshbabu Chintaguntla  (was: Josh Elser)

[~elserj] [~sergey.soldatov] please review.

> psql doesn't support date/time with values smaller than milliseconds
> 
>
> Key: PHOENIX-4874
> URL: https://issues.apache.org/jira/browse/PHOENIX-4874
> Project: Phoenix
>  Issue Type: Task
>Reporter: Josh Elser
>Assignee: Rajeshbabu Chintaguntla
>Priority: Major
> Attachments: PHOENIX-4874.patch
>
>
> [https://phoenix.apache.org/tuning.html] lacks entries for 
> phoenix.query.timeFormat, phoenix.query.timestampFormat which are used by 
> psql to parse out TIME and TIMESTAMP data types.
> Add them.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [DISCUSS] Suggestions for Phoenix from HBaseCon Asia notes

2018-09-11 Thread la...@apache.org
 Sorry for coming a bit late to this. I've been thinking about some of lines 
for a bit.
It seems Phoenix serves 4 distinct purposes:
1. Query parsing and compiling.2. A type system3. Query execution4. Efficient 
HBase interface
Each of these is useful by itself, but we do not expose these as stable 
interfaces.We have seen a lot of need to tie HBase into "higher level" service, 
such as Spark (and Presto, etc).
I think we can get a long way if we separate at least #1 (SQL) from the rest 
#2, #3, and #4 (Typed HBase Interface - THI).
Phoenix is used via SQL (#1), other tools such as Presto, Impala, Drill, Spark, 
etc, can interface efficiently with HBase via THI (#2, #3, and #4).
Thoughts?
-- Lars
On Monday, August 27, 2018, 11:03:33 AM PDT, Josh Elser  
wrote:  
 
 (bcc: dev@hbase, in case folks there have been waiting for me to send 
this email to dev@phoenix)

Hi,

In case you missed it, there was an HBaseCon event held in Asia 
recently. Stack took some great notes and shared them with the HBase 
community. A few of them touched on Phoenix, directly or in a related 
manner. I think they are good "criticisms" that are beneficial for us to 
hear.

1. The phoenix-$version-client.jar size is prohibitively large

In this day and age, I'm surprised that this is a big issue for people. 
I know have a lot of cruft, most of which coming from hadoop. We have 
gotten better here over recent releases, but I would guess that there is 
more we can do.

2. Can Phoenix be the de-facto schema for SQL on HBase?

We've long asserted "if you have to ask how Phoenix serializes data, you 
shouldn't be do it" (a nod that you have to write lots of code). What if 
we turn that on its head? Could we extract our PDataType serialization, 
composite row-key, column encoding, etc into a minimal API that folks 
with their own itches can use?

With the growing integrations into Phoenix, we could embrace them by 
providing an API to make what they're doing easier. In the same vein, we 
cement ourselves as a cornerstone of doing it "correctly".

3. Better recommendations to users to not attempt certain queries.

We definitively know that there are certain types of queries that 
Phoenix cannot support well (compared to optimal Phoenix use-cases). 
Users very commonly fall into such pitfalls on their own and this leaves 
a bad taste in their mouth (thinking that the product "stinks").

Can we do a better job of telling the user when and why it happened? 
What would such a user-interaction model look like? Can we supplement 
the "why" with instructions of what to do differently (even if in the 
abstract)?

4. Phoenix-Calcite

This was mentioned as a "nice to have". From what I understand, there 
was nothing explicitly from with the implementation or approach, just 
that it was a massive undertaking to continue with little immediate 
gain. Would this be a boon for us to try to continue in some form? Are 
there steps we can take that would help push us along the right path?

Anyways, I'd love to hear everyone's thoughts. While the concerns were 
raised at HBaseCon Asia, the suggestions that accompany them here are 
largely mine ;). Feel free to break them out into their own threads if 
you think that would be better (or say that you disagree with me -- 
that's cool too)!

- Josh
  

[jira] [Resolved] (PHOENIX-4897) Phoenix Kafka Plugin support specifying message offset

2018-09-11 Thread XiaoMifan (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4897?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

XiaoMifan resolved PHOENIX-4897.

Resolution: Fixed

let me mark this ticket as resolved.

> Phoenix Kafka Plugin support specifying message offset
> --
>
> Key: PHOENIX-4897
> URL: https://issues.apache.org/jira/browse/PHOENIX-4897
> Project: Phoenix
>  Issue Type: New Feature
>Affects Versions: 4.13.0
>Reporter: XiaoMifan
>Priority: Major
>
> I found Kafka Plugin only consume the new messages after this plugin started, 
> could not consume the message from beginning (at least I didn't found 
> introduction on Phonenix docs).
> And if the plugin app crashed, util it got restarted, the message happened 
> during this period are not got ingested.
> If I'm missing some thing that this plugin has already supported it, please 
> give some advise, thanks.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Assigned] (PHOENIX-4093) org.apache.phoenix.exception.PhoenixIOException: java.net.SocketTimeoutException: callTimeout=60000, callDuration=60304:

2018-09-11 Thread Jaanai Zhang (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaanai Zhang reassigned PHOENIX-4093:
-

Assignee: Jaanai Zhang

> org.apache.phoenix.exception.PhoenixIOException: 
> java.net.SocketTimeoutException: callTimeout=6, callDuration=60304:
> 
>
> Key: PHOENIX-4093
> URL: https://issues.apache.org/jira/browse/PHOENIX-4093
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.10.0
> Environment: Phoenix4.10
> HBase 1.2  CDH5.12
>Reporter: Jepson
>Assignee: Jaanai Zhang
>Priority: Major
>  Labels: performance
>
> SQL Error [101] [08000]: org.apache.phoenix.exception.PhoenixIOException: 
> Failed after attempts=36, exceptions:
> Thu Aug 17 10:51:48 UTC 2017, null, *java.net.SocketTimeoutException: 
> callTimeout=6, callDuration=60304*: row '' on table 'DW:OMS_TIO_IDX' at 
> region=DW:OMS_TIO_IDX,,1502808904791.06aa2e941810212e9c8733e5f6bdb9ec., 
> hostname=hadoop44,60020,1502954074181, seqNum=8



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (PHOENIX-4093) org.apache.phoenix.exception.PhoenixIOException: java.net.SocketTimeoutException: callTimeout=60000, callDuration=60304:

2018-09-11 Thread Jaanai (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4093?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaanai resolved PHOENIX-4093.
-
Resolution: Not A Bug

> org.apache.phoenix.exception.PhoenixIOException: 
> java.net.SocketTimeoutException: callTimeout=6, callDuration=60304:
> 
>
> Key: PHOENIX-4093
> URL: https://issues.apache.org/jira/browse/PHOENIX-4093
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.10.0
> Environment: Phoenix4.10
> HBase 1.2  CDH5.12
>Reporter: Jepson
>Assignee: Jaanai
>Priority: Major
>  Labels: performance
>
> SQL Error [101] [08000]: org.apache.phoenix.exception.PhoenixIOException: 
> Failed after attempts=36, exceptions:
> Thu Aug 17 10:51:48 UTC 2017, null, *java.net.SocketTimeoutException: 
> callTimeout=6, callDuration=60304*: row '' on table 'DW:OMS_TIO_IDX' at 
> region=DW:OMS_TIO_IDX,,1502808904791.06aa2e941810212e9c8733e5f6bdb9ec., 
> hostname=hadoop44,60020,1502954074181, seqNum=8



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Resolved] (PHOENIX-4364) java.sql.SQLException: ERROR 2008 (INT10): Unable to find cached index metadata.

2018-09-11 Thread Jaanai (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jaanai resolved PHOENIX-4364.
-
Resolution: Not A Problem

> java.sql.SQLException: ERROR 2008 (INT10): Unable to find cached index 
> metadata. 
> -
>
> Key: PHOENIX-4364
> URL: https://issues.apache.org/jira/browse/PHOENIX-4364
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.10.0
> Environment: phoenix4.10.0
>Reporter: Jepson
>Priority: Critical
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> use phoenix jdbc: 
> {code:java}
> java.sql.SQLException: ERROR 2008 (INT10): Unable to find cached index 
> metadata.  ERROR 2008 (INT10): ERROR 2008 (INT10): Unable to find cached 
> index metadata.  key=-1442130476102410039 
> region=JYDW:OMS_ORDERINFO,,1509703165591.421fdfea168d20112be0d74b27cdf23a.host=hadoop52,60020,1510212373872
>  Index update failed
> {code}
>



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (PHOENIX-4880) Phoenix IndexTool doesn't work on HBase2 per documentation

2018-09-11 Thread Josh Elser (JIRA)


 [ 
https://issues.apache.org/jira/browse/PHOENIX-4880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated PHOENIX-4880:

Attachment: PHOENIX-4880.001.patch

> Phoenix IndexTool doesn't work on HBase2 per documentation
> --
>
> Key: PHOENIX-4880
> URL: https://issues.apache.org/jira/browse/PHOENIX-4880
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Major
> Fix For: 4.15.0, 5.1.0
>
> Attachments: PHOENIX-4880.001.patch
>
>
> The website documentation states that to run {{IndexTool}}, you should do:
> {code}
> $ hbase org.apache.phoenix.mapreduce.index.IndexTool
> {code}
> This ends up running the class using the phoenix-server jar which fails 
> because we have conflicting versions of commons-cli, as described by 
> HBASE-20201. Phoenix-client.jar does not have this problem as we did the 
> workaround there as well.
> {code}
> $ hadoop jar $PHOENIX_HOME/phoenix-*client.jar 
> org.apache.phoenix.mapreduce.index.IndexTool
> {code}
> Does work, however. I suppose we still want to fix phoenix-server.jar? (no 
> reason not to?)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (PHOENIX-4901) support insert grammar for spark-phoenix module

2018-09-11 Thread Jaanai (JIRA)
Jaanai created PHOENIX-4901:
---

 Summary: support insert grammar for spark-phoenix module
 Key: PHOENIX-4901
 URL: https://issues.apache.org/jira/browse/PHOENIX-4901
 Project: Phoenix
  Issue Type: New Feature
Affects Versions: 4.14.0, 4.13.0, 4.12.0
Reporter: Jaanai
Assignee: Jaanai


Support insert in Phoenix Spark module, so directly writing Phoenix table by 
Spark SQL



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)