[jira] [Commented] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread Kalyan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425944#comment-15425944
 ] 

Kalyan commented on PHOENIX-3185:
-

Thanks William Yang , I am not aware of this issue then please make this issue 
as duplicate.

James Taylor, please ignore my previous patch.

sorry about my mistake.

> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Fix For: 4.8.1
>
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread William Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425796#comment-15425796
 ] 

William Yang commented on PHOENIX-3185:
---

I'm afraid we already has an issue and a applicable patch for this bug: 
(PHOENIX-930)[https://issues.apache.org/jira/browse/PHOENIX-930]

> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Fix For: 4.8.1
>
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (PHOENIX-3194) Document Hive integration

2016-08-17 Thread Sergey Soldatov (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Soldatov reassigned PHOENIX-3194:


Assignee: Sergey Soldatov

> Document Hive integration
> -
>
> Key: PHOENIX-3194
> URL: https://issues.apache.org/jira/browse/PHOENIX-3194
> Project: Phoenix
>  Issue Type: Bug
>Reporter: James Taylor
>Assignee: Sergey Soldatov
>
> We should document our Hive integration similar to the way we've documented 
> our Spark integration[1] and Pig integration[2]. This would focus on how to 
> use it (as opposed to how it was implemented), limitations, version 
> requirements, and include examples and any required/optional config 
> parameters or other setup required.
> [1] http://phoenix.apache.org/phoenix_spark.html
> [2] http://phoenix.apache.org/pig_integration.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3194) Document Hive integration

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425763#comment-15425763
 ] 

James Taylor commented on PHOENIX-3194:
---

[~sergey.soldatov] - could you take a stab at this. We author our website 
typically using markdown. See http://phoenix.apache.org/building_website.html 
and look at phoenix_spark.md as a guide. The menus are driven by site.xml.

> Document Hive integration
> -
>
> Key: PHOENIX-3194
> URL: https://issues.apache.org/jira/browse/PHOENIX-3194
> Project: Phoenix
>  Issue Type: Bug
>Reporter: James Taylor
>
> We should document our Hive integration similar to the way we've documented 
> our Spark integration[1] and Pig integration[2]. This would focus on how to 
> use it (as opposed to how it was implemented), limitations, version 
> requirements, and include examples and any required/optional config 
> parameters or other setup required.
> [1] http://phoenix.apache.org/phoenix_spark.html
> [2] http://phoenix.apache.org/pig_integration.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (PHOENIX-3194) Document Hive integration

2016-08-17 Thread James Taylor (JIRA)
James Taylor created PHOENIX-3194:
-

 Summary: Document Hive integration
 Key: PHOENIX-3194
 URL: https://issues.apache.org/jira/browse/PHOENIX-3194
 Project: Phoenix
  Issue Type: Bug
Reporter: James Taylor


We should document our Hive integration similar to the way we've documented our 
Spark integration[1] and Pig integration[2]. This would focus on how to use it 
(as opposed to how it was implemented), limitations, version requirements, and 
include examples and any required/optional config parameters or other setup 
required.

[1] http://phoenix.apache.org/phoenix_spark.html
[2] http://phoenix.apache.org/pig_integration.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-2807) Update logging and explain plan to show hbase tables instead of phoenix names

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-2807?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor updated PHOENIX-2807:
--
Issue Type: Improvement  (was: Sub-task)
Parent: (was: PHOENIX-1311)

> Update logging and explain plan to show hbase tables instead of phoenix names
> -
>
> Key: PHOENIX-2807
> URL: https://issues.apache.org/jira/browse/PHOENIX-2807
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
>Priority: Minor
> Fix For: 4.9.0
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (PHOENIX-2835) Add documentation for schema constructs and upgrade utility

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-2835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor resolved PHOENIX-2835.
---
Resolution: Fixed

> Add documentation for schema constructs and upgrade utility
> ---
>
> Key: PHOENIX-2835
> URL: https://issues.apache.org/jira/browse/PHOENIX-2835
> Project: Phoenix
>  Issue Type: Sub-task
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
>Priority: Minor
> Fix For: 4.8.0
>
> Attachments: PHOENIX-2835.patch, PHOENIX-2835_v1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-2835) Add documentation for schema constructs and upgrade utility

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-2835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor updated PHOENIX-2835:
--
Fix Version/s: (was: 4.9.0)
   4.8.0

> Add documentation for schema constructs and upgrade utility
> ---
>
> Key: PHOENIX-2835
> URL: https://issues.apache.org/jira/browse/PHOENIX-2835
> Project: Phoenix
>  Issue Type: Sub-task
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
>Priority: Minor
> Fix For: 4.8.0
>
> Attachments: PHOENIX-2835.patch, PHOENIX-2835_v1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-2835) Add documentation for schema constructs and upgrade utility

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-2835?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425616#comment-15425616
 ] 

James Taylor commented on PHOENIX-2835:
---

Looks good. I checked it in on your behalf.

> Add documentation for schema constructs and upgrade utility
> ---
>
> Key: PHOENIX-2835
> URL: https://issues.apache.org/jira/browse/PHOENIX-2835
> Project: Phoenix
>  Issue Type: Sub-task
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
>Priority: Minor
> Fix For: 4.9.0
>
> Attachments: PHOENIX-2835.patch, PHOENIX-2835_v1.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3188) Making renew leases of scanners more resilient

2016-08-17 Thread Andrew Purtell (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3188?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425601#comment-15425601
 ] 

Andrew Purtell commented on PHOENIX-3188:
-

That sounds like a better approach than using protected member variables. 
Please let us know if you still need a blessed way to get a scanner ID. 

> Making renew leases of scanners more resilient
> --
>
> Key: PHOENIX-3188
> URL: https://issues.apache.org/jira/browse/PHOENIX-3188
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Samarth Jain
> Attachments: PHOENIX-3188_hack.patch
>
>
> For renewing scanner leases, we need to perform an RPC per lease renewal. It 
> could happen that under load, these calls won't be picked up by region 
> handler threads soon enough which could cause queries to fail. There are a 
> couple of ways to solve this problem:
> 1) Issue renew lease calls at a higher priority. This essentially causes a 
> different thread pool to be used on the region servers. One such example is 
> the metadata and index updates calls we make. This would scale well unless 
> these special thread pools themselves get saturated with requests.
> 2) Batch up the the renewLease rpc calls. If HBase is able to provide us 
> scanner names and ids, then we can potentially renew leases for multiple 
> scanners by batching them up in one rpc. This would entail changes in both 
> client and server side of HBase. Client side - to expose scanner name. Server 
> side - to expose renewLease call on a scanner given a scanner name/id. 
> We still need to fix renewing leases for non-aggregate queries though. See 
> PHOENIX-1751



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread Samarth Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425417#comment-15425417
 ] 

Samarth Jain commented on PHOENIX-808:
--

We can always attempt to restore. Like you said, if we can't restore 
SYSTEM.STATS, then we can always regenerate the stats. Either way, in case of 
restore failures, we won't be deleting the snapshot that we failed to restore. 

> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>Assignee: Samarth Jain
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3009) Estimated Size for PTable is ~20% greater than actual size

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425419#comment-15425419
 ] 

James Taylor commented on PHOENIX-3009:
---

[~tdsilva] - would you mind taking this one after PHOENIX-3148, as you 
mentioned seeing the real size being more like 2500 bytes instead of 6000?

> Estimated Size for PTable is ~20% greater than actual size
> --
>
> Key: PHOENIX-3009
> URL: https://issues.apache.org/jira/browse/PHOENIX-3009
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Mujtaba Chohan
>Assignee: Thomas D'Silva
> Attachments: PHOENIX-3009.patch, ptable_retained.png
>
>
> {{PTable.estimatedSize}} returns size that is around 20% higher than actual 
> PTable size.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3009) Estimated Size for PTable is ~20% greater than actual size

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor updated PHOENIX-3009:
--
Assignee: Thomas D'Silva

> Estimated Size for PTable is ~20% greater than actual size
> --
>
> Key: PHOENIX-3009
> URL: https://issues.apache.org/jira/browse/PHOENIX-3009
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Mujtaba Chohan
>Assignee: Thomas D'Silva
> Attachments: PHOENIX-3009.patch, ptable_retained.png
>
>
> {{PTable.estimatedSize}} returns size that is around 20% higher than actual 
> PTable size.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425414#comment-15425414
 ] 

James Taylor commented on PHOENIX-808:
--

We could snapshot system.stats, but the data is more or less transient. Losing 
it just means you have to generate stats again. If we snapshot more than one 
table, it's a bit tricky to restore. What if we successfully restore one but 
not the other?



> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>Assignee: Samarth Jain
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread Samarth Jain (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425406#comment-15425406
 ] 

Samarth Jain commented on PHOENIX-808:
--

Agreed on not snapshotting the system.sequence table as we could end up losing 
increments on sequence values. Any reason why we shouldn't SYSTEM.STATS table? 
What if the upgrade for SYSTEM.STATS fails? For ex - when we changed the schema 
back in 4.3 for the SYSTEM.STATS table:

{code}
metaConnection = addColumnsIfNotExists(
metaConnection,
SYSTEM_STATS_NAME,

MetaDataProtocol.MIN_SYSTEM_TABLE_TIMESTAMP,

PhoenixDatabaseMetaData.GUIDE_POSTS_ROW_COUNT + " "
+ 
PLong.INSTANCE.getSqlTypeName());
{code}


> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>Assignee: Samarth Jain
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425402#comment-15425402
 ] 

James Taylor commented on PHOENIX-808:
--

IMHO, we should create only a snapshot of the system.catalog. We definitely 
don't want to snapshot the system.stats and system.tracing. Snapshoting and 
restoring the system.sequence table sounds risky. Making a snapshot of the 
system.function table might be fine too.

> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>Assignee: Samarth Jain
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread Samarth Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Samarth Jain reassigned PHOENIX-808:


Assignee: Samarth Jain

> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>Assignee: Samarth Jain
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [DISCUSS] RM for next release

2016-08-17 Thread Thomas D'Silva
Done.

On Wed, Aug 17, 2016 at 1:38 PM,  wrote:

> Great. Can somebody make me an administrator for Phoenix in Jira?I am for
> HBase but not for Phoenix.
> -- Lars
>   From: James Taylor 
>  To: "dev@phoenix.apache.org" ; lars hofhansl <
> la...@apache.org>
>  Sent: Tuesday, August 16, 2016 10:45 AM
>  Subject: Re: [DISCUSS] RM for next release
>
> Thanks, Lars. I don't think we need a vote - you have everyone's
> overwhelming support. If you're up for being both the 4.8.1 and 4.9 RM,
> that's be awesome. I'm hoping we can do 4.8.1 sooner rather than later as
> there are already a number of crucial fixes checked in. Can we document a
> target date for the releases? Other comments inline below.
>
> On Mon, Aug 15, 2016 at 2:33 PM,  wrote:
>
> > Should we put this up for a vote?
> > There are a few "conditions" that need to be met, IMHO, to be successful
> > with a monthly release cadence.
> > 1. Ability to skip releases. When we do monthly releases we can not
> expect
> > user to follow along. So it has to be possible to upgrade straight from
> > (say) 4.8.0 to 4.9.2. The upgrade paths would need to be defensively
> > implemented and tested.
>
>
> Yes, our upgrades support this and we have tests around this. They're run
> internally at SFDC currently, but Mujtaba is working on open sourcing them
> and having them run on ASF Jenkins jobs.
>
> 2. 100% stable test suite. Haven't looked at the Phoenix tests lately, but
> > with many releases in fairly quick succession, we need to be able to
> derive
> > confidence from the test suite.
>
>
> Yes, our tests are in good shape. We spent a lot of time on this over the
> last couple of releases. The failures we get are typically hiccups in ASF
> systems.
>
>
> > 2a. wide test coverage. Corollary to #2. Automated confidence needs to be
> > high. Even more so as we do many releases. I'd expect some of the
> > committers to spend some time to add tests.
> >
>
> We've improved this as well, but there's always more to do. One area we're
> working on for 4.9 is improving the test suite run speed so that we can
> potentially run more combinations of features. Also, there's been some good
> new stress-related tests developed by the folks over at Hortonworks that
> we're hoping will become open sources.
>
>
> > 3. The community and the PMC needs to be behind monthly release. Every
> > month there will be a release to test and to vote. At least 3 PMCs need
> to
> > +1.
> >
>
> Agreed. I think we can handle it.
>
>
> > 3a. committers are diligent about backporting fixes to earlier branches.
> > Everybody wants to work on the newest and coolest. At the same time, the
> RM
> > will not have enough B/W to decide what should be backported and to do
> the
> > actual backport.
> >
>
> Agreed.
>
>
> > 4. A compatibility story. Similar to the HBase one (
> > http://hbase.apache.org/book.html#hbase.versioning). I am happy to help
> > drive this. We may have a good compatibility story already, it's even
> > better if it's written down.
>
>
> Our compatibility story is documented here:
> https://phoenix.apache.org/upgrading.html. We actually support upgrade all
> the way back to 4.2. We can update as needed.
>
>
> > 5. Code quality. What follows is my opinion. No "beta" features. Those
> > should go into a newer version. No 1/2 baked code committed. Committers
> > should also look out for good code quality (like good comments, etc). Not
> > saying that is not the case now, btw.
> >
>
> Agreed.
>
>
> >
> > That's all I can think of offhand. If we can in principle agree on these,
> > I am volunteering to be RM for 4.9 and/or 4.8.1.
> > -- Lars
> >  From: Enis Söztutar 
> >  To: "dev@phoenix.apache.org" 
> >  Sent: Wednesday, July 13, 2016 6:49 PM
> >  Subject: Re: [DISCUSS] RM for next release
> >
> > Great, I think there is some consensus on:
> >
> >  1. Doing 4.8.1 release.
> >  2. Dropping HBase-1.0 for 4.9+
> >  3. Lars being RM for 4.9 (and 4.8.1?)
> >
> > I just created 4.8.1 as a release version. Let's use that to track issues
> > to backport. Committers, once 4.8.0 RC is cut, please start committing
> > issues to 4.8 branches as well, and don't forget to set the fixVersion
> with
> > 4.8.1 as well. Only bug fixes should be committed unless we need to do an
> > exception.
> >
> > Enis
> >
> > On Sun, Jul 10, 2016 at 9:39 AM, Andrew Purtell <
> andrew.purt...@gmail.com>
> > wrote:
> >
> > > It's the monthly sweep of commits on later branches, backporting, and
> > > testing each change. So yeah it's the other end of the spectrum and
> good
> > as
> > > another a data point. Where a Phoenix RM would fall on that spectrum
> will
> > > correlate with how long the long term support for a branch has been in
> > > effect.
> > >
> > > I am also planning to run ITBLL with active chaos agents for 24 hours
> for
> > > every HBase RC once I figure out Dima's clusterdock stuff. Speaking of
> > > which: We should have something like this for Phoenix. For the most
> part
> > we
> > > can reu

[jira] [Updated] (PHOENIX-808) Create snapshot of system tables prior to upgrade and restore on any failure

2016-08-17 Thread Samarth Jain (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-808?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Samarth Jain updated PHOENIX-808:
-
Summary: Create snapshot of system tables prior to upgrade and restore on 
any failure  (was: Create snapshot of SYSTEM.CATALOG prior to upgrade and 
restore on any failure)

> Create snapshot of system tables prior to upgrade and restore on any failure
> 
>
> Key: PHOENIX-808
> URL: https://issues.apache.org/jira/browse/PHOENIX-808
> Project: Phoenix
>  Issue Type: Task
>Reporter: James Taylor
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [DISCUSS] RM for next release

2016-08-17 Thread larsh
Great. Can somebody make me an administrator for Phoenix in Jira?I am for HBase 
but not for Phoenix.
-- Lars
  From: James Taylor 
 To: "dev@phoenix.apache.org" ; lars hofhansl 
 
 Sent: Tuesday, August 16, 2016 10:45 AM
 Subject: Re: [DISCUSS] RM for next release
   
Thanks, Lars. I don't think we need a vote - you have everyone's
overwhelming support. If you're up for being both the 4.8.1 and 4.9 RM,
that's be awesome. I'm hoping we can do 4.8.1 sooner rather than later as
there are already a number of crucial fixes checked in. Can we document a
target date for the releases? Other comments inline below.

On Mon, Aug 15, 2016 at 2:33 PM,  wrote:

> Should we put this up for a vote?
> There are a few "conditions" that need to be met, IMHO, to be successful
> with a monthly release cadence.
> 1. Ability to skip releases. When we do monthly releases we can not expect
> user to follow along. So it has to be possible to upgrade straight from
> (say) 4.8.0 to 4.9.2. The upgrade paths would need to be defensively
> implemented and tested.


Yes, our upgrades support this and we have tests around this. They're run
internally at SFDC currently, but Mujtaba is working on open sourcing them
and having them run on ASF Jenkins jobs.

2. 100% stable test suite. Haven't looked at the Phoenix tests lately, but
> with many releases in fairly quick succession, we need to be able to derive
> confidence from the test suite.


Yes, our tests are in good shape. We spent a lot of time on this over the
last couple of releases. The failures we get are typically hiccups in ASF
systems.


> 2a. wide test coverage. Corollary to #2. Automated confidence needs to be
> high. Even more so as we do many releases. I'd expect some of the
> committers to spend some time to add tests.
>

We've improved this as well, but there's always more to do. One area we're
working on for 4.9 is improving the test suite run speed so that we can
potentially run more combinations of features. Also, there's been some good
new stress-related tests developed by the folks over at Hortonworks that
we're hoping will become open sources.


> 3. The community and the PMC needs to be behind monthly release. Every
> month there will be a release to test and to vote. At least 3 PMCs need to
> +1.
>

Agreed. I think we can handle it.


> 3a. committers are diligent about backporting fixes to earlier branches.
> Everybody wants to work on the newest and coolest. At the same time, the RM
> will not have enough B/W to decide what should be backported and to do the
> actual backport.
>

Agreed.


> 4. A compatibility story. Similar to the HBase one (
> http://hbase.apache.org/book.html#hbase.versioning). I am happy to help
> drive this. We may have a good compatibility story already, it's even
> better if it's written down.


Our compatibility story is documented here:
https://phoenix.apache.org/upgrading.html. We actually support upgrade all
the way back to 4.2. We can update as needed.


> 5. Code quality. What follows is my opinion. No "beta" features. Those
> should go into a newer version. No 1/2 baked code committed. Committers
> should also look out for good code quality (like good comments, etc). Not
> saying that is not the case now, btw.
>

Agreed.


>
> That's all I can think of offhand. If we can in principle agree on these,
> I am volunteering to be RM for 4.9 and/or 4.8.1.
> -- Lars
>      From: Enis Söztutar 
>  To: "dev@phoenix.apache.org" 
>  Sent: Wednesday, July 13, 2016 6:49 PM
>  Subject: Re: [DISCUSS] RM for next release
>
> Great, I think there is some consensus on:
>
>  1. Doing 4.8.1 release.
>  2. Dropping HBase-1.0 for 4.9+
>  3. Lars being RM for 4.9 (and 4.8.1?)
>
> I just created 4.8.1 as a release version. Let's use that to track issues
> to backport. Committers, once 4.8.0 RC is cut, please start committing
> issues to 4.8 branches as well, and don't forget to set the fixVersion with
> 4.8.1 as well. Only bug fixes should be committed unless we need to do an
> exception.
>
> Enis
>
> On Sun, Jul 10, 2016 at 9:39 AM, Andrew Purtell 
> wrote:
>
> > It's the monthly sweep of commits on later branches, backporting, and
> > testing each change. So yeah it's the other end of the spectrum and good
> as
> > another a data point. Where a Phoenix RM would fall on that spectrum will
> > correlate with how long the long term support for a branch has been in
> > effect.
> >
> > I am also planning to run ITBLL with active chaos agents for 24 hours for
> > every HBase RC once I figure out Dima's clusterdock stuff. Speaking of
> > which: We should have something like this for Phoenix. For the most part
> we
> > can reuse HBase's chaos stuff. We'd want to write and verify the same
> type
> > of linked data, with all interesting features enabled like transactions
> and
> > local indexes. If we had this then doing it for Phoenix RCs would likely
> > fall to the RM.
> >
> >
> > > On Jul 10, 2016, at 2:42 AM,  
> > wrote:
> > >
> > > 0.94 never took that m

[jira] [Commented] (PHOENIX-2995) Write performance severely degrades with large number of views

2016-08-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-2995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425316#comment-15425316
 ] 

Hudson commented on PHOENIX-2995:
-

SUCCESS: Integrated in Jenkins build Phoenix-master #1364 (See 
[https://builds.apache.org/job/Phoenix-master/1364/])
PHOENIX-2995 Write performance severely degrades with large number of (tdsilva: 
rev 386cbbbf7a5dd736888e8e0bfe16e513d54e215c)
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixConnection.java


> Write performance severely degrades with large number of views 
> ---
>
> Key: PHOENIX-2995
> URL: https://issues.apache.org/jira/browse/PHOENIX-2995
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Mujtaba Chohan
>Assignee: Thomas D'Silva
>  Labels: Argus
> Fix For: 4.8.1
>
> Attachments: PHOENIX-2995-v2.patch, PHOENIX-2995-v3.patch, 
> PHOENIX-2995-v4.patch, PHOENIX-2995-v5.patch, PHOENIX-2995-v6.patch, 
> PHOENIX-2995.patch, create_view_and_upsert.png, image.png, image2.png, 
> image3.png, upsert_rate.png
>
>
> Write performance for each 1K batch degrades significantly when there are 
> *10K* views being written in random with default 
> {{phoenix.client.maxMetaDataCacheSize}}. With all views created, upsert rate 
> remains around 25 seconds per 1K batch i.e. ~2K rows/min upsert rate. 
> When {{phoenix.client.maxMetaDataCacheSize}} is increased to 100MB+ then view 
> does not need to get re-resolved and upsert rate gets back to normal ~60K 
> rows/min.
> With *100K* views and {{phoenix.client.maxMetaDataCacheSize}} set to 1GB, I 
> wasn't able create all 100K views as upsert time for each 1K batch keeps on 
> steadily increasing. 
> Following graph shows 1K batch upsert rate over time with variation of number 
> of views. Rows are upserted to random views {{CREATE VIEW IF NOT EXISTS ... 
> APPEND_ONLY_SCHEMA = true, UPDATE_CACHE_FREQUENCY=90}} is executed before 
> upsert statement.
> !upsert_rate.png!
> Base table is also created with {{APPEND_ONLY_SCHEMA = true, 
> UPDATE_CACHE_FREQUENCY = 90, AUTO_PARTITION_SEQ}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3193) Tracing UI cleanup - final tasks before GSoC pull request

2016-08-17 Thread Mujtaba Chohan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3193?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425155#comment-15425155
 ] 

Mujtaba Chohan commented on PHOENIX-3193:
-

FYI [~nishani]

> Tracing UI cleanup - final tasks before GSoC pull request
> -
>
> Key: PHOENIX-3193
> URL: https://issues.apache.org/jira/browse/PHOENIX-3193
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Mujtaba Chohan
>Assignee: Nishani 
>
> Points from GSoC presentation on tracing imporvements:
> *Tracing UI*
> * Remove line chart 
> * In list page, run query with description, start_time, (end_time-start_time) 
> duration from T where trace_id = ?
> * More space for descriptions on bar chart. Wrap if necessary
> * Label for X axis on timeline sometime again start from 0, if X axis is in 
> seconds then it should not rollover after 60 seconds unless minutes are also 
> shown
> * X-axis labeled as Node on various charts, but should be Percentage
> *Zipkin*
> * Flip zipkin chart on vertical axis with arrows going other way. So start 
> from the top level root on the leftmost side and work toward children on the 
> right.
> * Ask zipkin community if there's a way to tell it that date/time is in 
> milliseconds.
> *Overall*
> * Please put together a pull request to the phoenix project to add the 
> zipkiin work you've done to the OS project. Ideally, include the zipkin work 
> in the phoenix-tracing module and call it phoeix-tracing. Only if there is 
> some major hurdle, create a new module.
> * Test with trace_ids that have multiple spans with duration (end_time - 
> start_time) > 5 ms and verify that UI and Zipkin output shows the correct 
> corresponding timeline
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3193) Tracing UI cleanup - final tasks before GSoC pull request

2016-08-17 Thread Mujtaba Chohan (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mujtaba Chohan updated PHOENIX-3193:

Assignee: Nishani 

> Tracing UI cleanup - final tasks before GSoC pull request
> -
>
> Key: PHOENIX-3193
> URL: https://issues.apache.org/jira/browse/PHOENIX-3193
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Mujtaba Chohan
>Assignee: Nishani 
>
> Points from GSoC presentation on tracing imporvements:
> *Tracing UI*
> * Remove line chart 
> * In list page, run query with description, start_time, (end_time-start_time) 
> duration from T where trace_id = ?
> * More space for descriptions on bar chart. Wrap if necessary
> * Label for X axis on timeline sometime again start from 0, if X axis is in 
> seconds then it should not rollover after 60 seconds unless minutes are also 
> shown
> * X-axis labeled as Node on various charts, but should be Percentage
> *Zipkin*
> * Flip zipkin chart on vertical axis with arrows going other way. So start 
> from the top level root on the leftmost side and work toward children on the 
> right.
> * Ask zipkin community if there's a way to tell it that date/time is in 
> milliseconds.
> *Overall*
> * Please put together a pull request to the phoenix project to add the 
> zipkiin work you've done to the OS project. Ideally, include the zipkin work 
> in the phoenix-tracing module and call it phoeix-tracing. Only if there is 
> some major hurdle, create a new module.
> * Test with trace_ids that have multiple spans with duration (end_time - 
> start_time) > 5 ms and verify that UI and Zipkin output shows the correct 
> corresponding timeline
>   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (PHOENIX-3193) Tracing UI cleanup - final tasks before GSoC pull request

2016-08-17 Thread Mujtaba Chohan (JIRA)
Mujtaba Chohan created PHOENIX-3193:
---

 Summary: Tracing UI cleanup - final tasks before GSoC pull request
 Key: PHOENIX-3193
 URL: https://issues.apache.org/jira/browse/PHOENIX-3193
 Project: Phoenix
  Issue Type: Improvement
Reporter: Mujtaba Chohan


Points from GSoC presentation on tracing imporvements:

*Tracing UI*
* Remove line chart 
* In list page, run query with description, start_time, (end_time-start_time) 
duration from T where trace_id = ?
* More space for descriptions on bar chart. Wrap if necessary
* Label for X axis on timeline sometime again start from 0, if X axis is in 
seconds then it should not rollover after 60 seconds unless minutes are also 
shown
* X-axis labeled as Node on various charts, but should be Percentage

*Zipkin*
* Flip zipkin chart on vertical axis with arrows going other way. So start from 
the top level root on the leftmost side and work toward children on the right.
* Ask zipkin community if there's a way to tell it that date/time is in 
milliseconds.

*Overall*
* Please put together a pull request to the phoenix project to add the zipkiin 
work you've done to the OS project. Ideally, include the zipkin work in the 
phoenix-tracing module and call it phoeix-tracing. Only if there is some major 
hurdle, create a new module.
* Test with trace_ids that have multiple spans with duration (end_time - 
start_time) > 5 ms and verify that UI and Zipkin output shows the correct 
corresponding timeline





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-2995) Write performance severely degrades with large number of views

2016-08-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-2995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425139#comment-15425139
 ] 

Hudson commented on PHOENIX-2995:
-

FAILURE: Integrated in Jenkins build Phoenix-master #1363 (See 
[https://builds.apache.org/job/Phoenix-master/1363/])
PHOENIX-2995 Write performance severely degrades with large number of (tdsilva: 
rev 3130fa9919ba937d98853f9be8068a0a8a8a96f5)
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/query/MetaDataMutated.java
* (edit) phoenix-core/src/main/java/org/apache/phoenix/util/TransactionUtil.java
* (edit) 
phoenix-core/src/test/java/org/apache/phoenix/schema/PMetaDataImplTest.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/jdbc/PhoenixConnection.java
* (add) 
phoenix-core/src/main/java/org/apache/phoenix/schema/PSynchronizedMetaData.java
* (edit) phoenix-core/src/it/java/org/apache/phoenix/end2end/UpsertSelectIT.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionQueryServicesImpl.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/query/DelegateConnectionQueryServices.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java
* (edit) phoenix-core/src/main/java/org/apache/phoenix/schema/PMetaData.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/execute/MutationState.java
* (edit) phoenix-core/src/main/java/org/apache/phoenix/schema/PMetaDataImpl.java
* (edit) 
phoenix-core/src/main/java/org/apache/phoenix/compile/CreateTableCompiler.java


> Write performance severely degrades with large number of views 
> ---
>
> Key: PHOENIX-2995
> URL: https://issues.apache.org/jira/browse/PHOENIX-2995
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Mujtaba Chohan
>Assignee: Thomas D'Silva
>  Labels: Argus
> Fix For: 4.8.1
>
> Attachments: PHOENIX-2995-v2.patch, PHOENIX-2995-v3.patch, 
> PHOENIX-2995-v4.patch, PHOENIX-2995-v5.patch, PHOENIX-2995-v6.patch, 
> PHOENIX-2995.patch, create_view_and_upsert.png, image.png, image2.png, 
> image3.png, upsert_rate.png
>
>
> Write performance for each 1K batch degrades significantly when there are 
> *10K* views being written in random with default 
> {{phoenix.client.maxMetaDataCacheSize}}. With all views created, upsert rate 
> remains around 25 seconds per 1K batch i.e. ~2K rows/min upsert rate. 
> When {{phoenix.client.maxMetaDataCacheSize}} is increased to 100MB+ then view 
> does not need to get re-resolved and upsert rate gets back to normal ~60K 
> rows/min.
> With *100K* views and {{phoenix.client.maxMetaDataCacheSize}} set to 1GB, I 
> wasn't able create all 100K views as upsert time for each 1K batch keeps on 
> steadily increasing. 
> Following graph shows 1K batch upsert rate over time with variation of number 
> of views. Rows are upserted to random views {{CREATE VIEW IF NOT EXISTS ... 
> APPEND_ONLY_SCHEMA = true, UPDATE_CACHE_FREQUENCY=90}} is executed before 
> upsert statement.
> !upsert_rate.png!
> Base table is also created with {{APPEND_ONLY_SCHEMA = true, 
> UPDATE_CACHE_FREQUENCY = 90, AUTO_PARTITION_SEQ}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3148) Reduce size of PTable so that more tables can be cached in the metada cache.

2016-08-17 Thread Thomas D'Silva (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas D'Silva updated PHOENIX-3148:

Fix Version/s: (was: 4.9.0)
   4.8.1

> Reduce size of PTable so that more tables can be cached in the metada cache.
> 
>
> Key: PHOENIX-3148
> URL: https://issues.apache.org/jira/browse/PHOENIX-3148
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Thomas D'Silva
>Assignee: Thomas D'Silva
>  Labels: argus
> Fix For: 4.8.1
>
>
> According to PHOENIX-2995, the current size is 7KB per PTable which works out 
> to enabling 140K PTables per 1GB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (PHOENIX-2995) Write performance severely degrades with large number of views

2016-08-17 Thread Thomas D'Silva (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-2995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas D'Silva resolved PHOENIX-2995.
-
Resolution: Fixed

> Write performance severely degrades with large number of views 
> ---
>
> Key: PHOENIX-2995
> URL: https://issues.apache.org/jira/browse/PHOENIX-2995
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Mujtaba Chohan
>Assignee: Thomas D'Silva
>  Labels: Argus
> Fix For: 4.8.1
>
> Attachments: PHOENIX-2995-v2.patch, PHOENIX-2995-v3.patch, 
> PHOENIX-2995-v4.patch, PHOENIX-2995-v5.patch, PHOENIX-2995-v6.patch, 
> PHOENIX-2995.patch, create_view_and_upsert.png, image.png, image2.png, 
> image3.png, upsert_rate.png
>
>
> Write performance for each 1K batch degrades significantly when there are 
> *10K* views being written in random with default 
> {{phoenix.client.maxMetaDataCacheSize}}. With all views created, upsert rate 
> remains around 25 seconds per 1K batch i.e. ~2K rows/min upsert rate. 
> When {{phoenix.client.maxMetaDataCacheSize}} is increased to 100MB+ then view 
> does not need to get re-resolved and upsert rate gets back to normal ~60K 
> rows/min.
> With *100K* views and {{phoenix.client.maxMetaDataCacheSize}} set to 1GB, I 
> wasn't able create all 100K views as upsert time for each 1K batch keeps on 
> steadily increasing. 
> Following graph shows 1K batch upsert rate over time with variation of number 
> of views. Rows are upserted to random views {{CREATE VIEW IF NOT EXISTS ... 
> APPEND_ONLY_SCHEMA = true, UPDATE_CACHE_FREQUENCY=90}} is executed before 
> upsert statement.
> !upsert_rate.png!
> Base table is also created with {{APPEND_ONLY_SCHEMA = true, 
> UPDATE_CACHE_FREQUENCY = 90, AUTO_PARTITION_SEQ}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (PHOENIX-3148) Reduce size of PTable so that more tables can be cached in the metada cache.

2016-08-17 Thread Thomas D'Silva (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas D'Silva reassigned PHOENIX-3148:
---

Assignee: Thomas D'Silva

> Reduce size of PTable so that more tables can be cached in the metada cache.
> 
>
> Key: PHOENIX-3148
> URL: https://issues.apache.org/jira/browse/PHOENIX-3148
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Thomas D'Silva
>Assignee: Thomas D'Silva
>  Labels: argus
> Fix For: 4.8.1
>
>
> According to PHOENIX-2995, the current size is 7KB per PTable which works out 
> to enabling 140K PTables per 1GB.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3190) Creating Tracing Visualization Ducumentation

2016-08-17 Thread Josh Elser (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3190?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated PHOENIX-3190:

Fix Version/s: (was: 4.8.0)

> Creating Tracing Visualization Ducumentation
> 
>
> Key: PHOENIX-3190
> URL: https://issues.apache.org/jira/browse/PHOENIX-3190
> Project: Phoenix
>  Issue Type: Sub-task
>Affects Versions: 4.8.0
>Reporter: Nishani 
>Assignee: Nishani 
>
> Documentation will contain
> -Improved Search Features
> -How to start the Phoenix Zipkin Web App
> -How to change the zookeeper host and port 
> -Features of Phoenix Zipkin Web App



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3189) HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url

2016-08-17 Thread Josh Elser (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated PHOENIX-3189:

Affects Version/s: 4.8.0

> HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url
> 
>
> Key: PHOENIX-3189
> URL: https://issues.apache.org/jira/browse/PHOENIX-3189
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.8.0
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Blocker
> Fix For: 4.9.0, 4.8.1
>
>
> We've been doing some more testing after PHOENIX-3126 and, with the help of 
> [~arpitgupta] and [~harsha_ch], we've found an issue in a test between Storm 
> and Phoenix.
> Storm was configured to create a JDBC Bolt, specifying the principal and 
> keytab in the JDBC URL, relying on PhoenixDriver to do the Kerberos login for 
> them. After PHOENIX-3126, a ZK server blacklisted the host running the bolt, 
> and we observed that there were over 140 active ZK threads in the JVM.
> This results in a subtle change where every time the client tries to get a 
> new Connection, we end up getting a new UGI instance (because the 
> {{ConnectionQueryServicesImpl#openConnection()}} always does a new login).
> If users are correctly caching Connections, there isn't an issue (best as I 
> can presently tell). However, if users rely on the getting the same 
> connection every time (the pre-PHOENIX-3126), they will saturate their local 
> JVM with connections and crash.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3189) HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url

2016-08-17 Thread Josh Elser (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated PHOENIX-3189:

Fix Version/s: 4.9.0

> HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url
> 
>
> Key: PHOENIX-3189
> URL: https://issues.apache.org/jira/browse/PHOENIX-3189
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.8.0
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Blocker
> Fix For: 4.9.0, 4.8.1
>
>
> We've been doing some more testing after PHOENIX-3126 and, with the help of 
> [~arpitgupta] and [~harsha_ch], we've found an issue in a test between Storm 
> and Phoenix.
> Storm was configured to create a JDBC Bolt, specifying the principal and 
> keytab in the JDBC URL, relying on PhoenixDriver to do the Kerberos login for 
> them. After PHOENIX-3126, a ZK server blacklisted the host running the bolt, 
> and we observed that there were over 140 active ZK threads in the JVM.
> This results in a subtle change where every time the client tries to get a 
> new Connection, we end up getting a new UGI instance (because the 
> {{ConnectionQueryServicesImpl#openConnection()}} always does a new login).
> If users are correctly caching Connections, there isn't an issue (best as I 
> can presently tell). However, if users rely on the getting the same 
> connection every time (the pre-PHOENIX-3126), they will saturate their local 
> JVM with connections and crash.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3189) HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url

2016-08-17 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15425017#comment-15425017
 ] 

ASF GitHub Bot commented on PHOENIX-3189:
-

GitHub user joshelser opened a pull request:

https://github.com/apache/phoenix/pull/191

PHOENIX-3189 Perform Kerberos login before ConnectionInfo is constructed

Now that ConnectionInfo has the current User/UGI stored inside, we must
make sure that any automatic Kerberos login occurs before the ConnectionInfo
object is constructed. Otherwise, we will have multiple instances of
ConnectionInfo that differ only by the User, which will leak HBase/ZK
connections in the connectionQueryServicesMap.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/joshelser/phoenix 3189-secure-cnxninfo

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/phoenix/pull/191.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #191


commit ac4184b9ec52ab261fc4d9f73edc6676eb421d79
Author: Josh Elser 
Date:   2016-08-17T17:34:59Z

PHOENIX-3189 Perform Kerberos login before ConnectionInfo is constructed

Now that ConnectionInfo has the current User/UGI stored inside, we must
make sure that any automatic Kerberos login occurs before the ConnectionInfo
object is constructed. Otherwise, we will have multiple instances of
ConnectionInfo that differ only by the User, which will leak HBase/ZK
connections in the connectionQueryServicesMap.




> HBase/ZooKeeper connection leaks when providing principal/keytab in JDBC url
> 
>
> Key: PHOENIX-3189
> URL: https://issues.apache.org/jira/browse/PHOENIX-3189
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Josh Elser
>Assignee: Josh Elser
>Priority: Blocker
> Fix For: 4.8.1
>
>
> We've been doing some more testing after PHOENIX-3126 and, with the help of 
> [~arpitgupta] and [~harsha_ch], we've found an issue in a test between Storm 
> and Phoenix.
> Storm was configured to create a JDBC Bolt, specifying the principal and 
> keytab in the JDBC URL, relying on PhoenixDriver to do the Kerberos login for 
> them. After PHOENIX-3126, a ZK server blacklisted the host running the bolt, 
> and we observed that there were over 140 active ZK threads in the JVM.
> This results in a subtle change where every time the client tries to get a 
> new Connection, we end up getting a new UGI instance (because the 
> {{ConnectionQueryServicesImpl#openConnection()}} always does a new login).
> If users are correctly caching Connections, there isn't an issue (best as I 
> can presently tell). However, if users rely on the getting the same 
> connection every time (the pre-PHOENIX-3126), they will saturate their local 
> JVM with connections and crash.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] phoenix pull request #191: PHOENIX-3189 Perform Kerberos login before Connec...

2016-08-17 Thread joshelser
GitHub user joshelser opened a pull request:

https://github.com/apache/phoenix/pull/191

PHOENIX-3189 Perform Kerberos login before ConnectionInfo is constructed

Now that ConnectionInfo has the current User/UGI stored inside, we must
make sure that any automatic Kerberos login occurs before the ConnectionInfo
object is constructed. Otherwise, we will have multiple instances of
ConnectionInfo that differ only by the User, which will leak HBase/ZK
connections in the connectionQueryServicesMap.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/joshelser/phoenix 3189-secure-cnxninfo

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/phoenix/pull/191.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #191


commit ac4184b9ec52ab261fc4d9f73edc6676eb421d79
Author: Josh Elser 
Date:   2016-08-17T17:34:59Z

PHOENIX-3189 Perform Kerberos login before ConnectionInfo is constructed

Now that ConnectionInfo has the current User/UGI stored inside, we must
make sure that any automatic Kerberos login occurs before the ConnectionInfo
object is constructed. Otherwise, we will have multiple instances of
ConnectionInfo that differ only by the User, which will leak HBase/ZK
connections in the connectionQueryServicesMap.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (PHOENIX-3192) phoenix-spark dataframe issue with combination of "column family + column name"

2016-08-17 Thread Josh Mahonin (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424957#comment-15424957
 ] 

Josh Mahonin commented on PHOENIX-3192:
---

For the most part, the phoenix-spark code tries not to do to much w.r.t. column 
names, as most of it is offloaded to the PhoenixConfigurationUtil class:
https://github.com/apache/phoenix/blob/master/phoenix-spark/src/main/scala/org/apache/phoenix/spark/PhoenixRDD.scala#L81

All I can suggest is to try keep any fixes in the utility classes, if possible, 
so that Pig and the other MR integrations can make use of them as well. Unless 
of course the issue is within phoenix-spark itself, then change whatever you 
think is necessary.

> phoenix-spark dataframe issue with combination of "column family + column 
> name"
> ---
>
> Key: PHOENIX-3192
> URL: https://issues.apache.org/jira/browse/PHOENIX-3192
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
>
> 1. create table with different column families with same column name
> create table tbl_1 (rowkey varchar primary key, cf1.c1 varchar, cf1.c2 
> integer, cf2.c1 double, cf2.c2 boolean, cf3.c1 bigint);
> 2. insert sample data into table
> 3. create dataframe using phoenix table with different column names
> val df1 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C2"), conf = configuration )
> df1.show // this will work
> 4. create dataframe using phoenix table with same column names
> val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C1"), conf = configuration )
> df2.show// this will fail
> 5. reason currently we are not handled the dataframe solution fully (column 
> family + column name). 
> only works with (column name)
> Exception:
> scala> val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C1"), conf = configuration )
> df2: org.apache.spark.sql.DataFrame = [C1: string, C1: double]
> scala> df2.show
> 16/08/17 22:16:54 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
> scala.MatchError: 1.5 (of class java.lang.Double)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:295)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:294)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:260)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:250)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$$anonfun$createToCatalystConverter$2.apply(CatalystTypeConverters.scala:401)
>   at 
> org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
>   at 
> org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
>   at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
>   at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
>   at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>   at 
> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
>   at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
>   at scala.collection.AbstractIterator.to(Iterator.scala:1157)
>   at 
> scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
>   at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
>   at 
> scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
>   at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
>   at 
> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
>   at 
> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
>   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>   at or

[jira] [Commented] (PHOENIX-3192) phoenix-spark dataframe issue with combination of "column family + column name"

2016-08-17 Thread Kalyan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3192?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424914#comment-15424914
 ] 

Kalyan commented on PHOENIX-3192:
-

I will be working on this bug .. planning to change the DataFrame creation with 
schema mapping provided by user.

Otherwise we are not able to work with sql queries using sqlContext.sql() 
operation.

Please share any suggestions to work with DataFrame fully (column family + 
column name)

> phoenix-spark dataframe issue with combination of "column family + column 
> name"
> ---
>
> Key: PHOENIX-3192
> URL: https://issues.apache.org/jira/browse/PHOENIX-3192
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
>
> 1. create table with different column families with same column name
> create table tbl_1 (rowkey varchar primary key, cf1.c1 varchar, cf1.c2 
> integer, cf2.c1 double, cf2.c2 boolean, cf3.c1 bigint);
> 2. insert sample data into table
> 3. create dataframe using phoenix table with different column names
> val df1 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C2"), conf = configuration )
> df1.show // this will work
> 4. create dataframe using phoenix table with same column names
> val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C1"), conf = configuration )
> df2.show// this will fail
> 5. reason currently we are not handled the dataframe solution fully (column 
> family + column name). 
> only works with (column name)
> Exception:
> scala> val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
> "CF2.C1"), conf = configuration )
> df2: org.apache.spark.sql.DataFrame = [C1: string, C1: double]
> scala> df2.show
> 16/08/17 22:16:54 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
> scala.MatchError: 1.5 (of class java.lang.Double)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:295)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:294)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:260)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:250)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
>   at 
> org.apache.spark.sql.catalyst.CatalystTypeConverters$$anonfun$createToCatalystConverter$2.apply(CatalystTypeConverters.scala:401)
>   at 
> org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
>   at 
> org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
>   at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
>   at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
>   at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>   at 
> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
>   at 
> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
>   at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
>   at scala.collection.AbstractIterator.to(Iterator.scala:1157)
>   at 
> scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
>   at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
>   at 
> scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
>   at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
>   at 
> org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
>   at 
> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
>   at 
> org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
>   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>   at org.apache.spark.scheduler.Task.run(Task.scala:89)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>   at 
> java.util.concurre

[jira] [Commented] (PHOENIX-2938) HFile support for SparkSQL DataFrame saves

2016-08-17 Thread Kalyan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-2938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424898#comment-15424898
 ] 

Kalyan commented on PHOENIX-2938:
-

Thanks Josh Mahonin for your suggestions.

i will work on these changes .. after the below bug fix.
PHOENIX-3192 : phoenix-spark dataframe issue with combination of "column family 
+ column name"

otherwise i need to refactor again.

> HFile support for SparkSQL DataFrame saves
> --
>
> Key: PHOENIX-2938
> URL: https://issues.apache.org/jira/browse/PHOENIX-2938
> Project: Phoenix
>  Issue Type: Improvement
>Reporter: Chris Tarnas
>Assignee: Kalyan
>Priority: Minor
>
> Currently when saving a DataFrame in Spark it is persisted as upserts. Having 
> an option to do saves natively via HFiles, as the MapReduce loader does, 
> would be a great performance improvement for large bulk loads. The current 
> work around to reduce the load on the regionservers would be to save to csv 
> from Spark then load via the MapReduce loader.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (PHOENIX-3191) COUNT and other aggregate functions not working

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3191?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor resolved PHOENIX-3191.
---
Resolution: Invalid

This is clearly an environmental issue. Would you mind asking on the dev or 
user list and include the Phoenix version you're using? I recommend you check 
your setting for the {{phoenix.spool.directory}} config property on the 
server-side which in later versions defaults to 
{{System.getProperty("java.io.tmpdir")}} through in earlier versions defaulted 
to /tmp (which won't be valid in some environments). Try setting it to 
something you know is valid.

> COUNT and other aggregate functions not working
> ---
>
> Key: PHOENIX-3191
> URL: https://issues.apache.org/jira/browse/PHOENIX-3191
> Project: Phoenix
>  Issue Type: Bug
>Affects Versions: 4.6.0
> Environment: CentOS(Server), Windows(Client)
>Reporter: Vivek K T
>Priority: Critical
>
> Following are steps to reproduce the issue : 
> 1) Create a dummy table BUDGET : 
> create table BUDGET
> (
>   id varchar (50) primary key,
>   date timestamp,
>   amount decimal,
>   currency varchar (50)
> )
> 2) Insert data in BUDGET table : 
> upsert into BUDGET (id,date,amount,currency) values 
> ('1',current_date(),20,'INR');
> upsert into BUDGET (id,date,amount,currency) values 
> ('2',current_date(),40,'INR');
> upsert into BUDGET (id,date,amount,currency) values 
> ('3',current_date(),60,'INR');
> upsert into BUDGET (id,date,amount,currency) values 
> ('4',current_date(),80,'INR');
> upsert into BUDGET (id,date,amount,currency) values 
> ('5',current_date(),100,'INR');
> (PLEASE NOTE : TABLE CREATION AND DATA INSERTION SHOULD BE DONE FROM LINUX 
> SERVER)
> 3) Fire following query from squirrel-sql on windows: 
> select count(*) from BUDGET
> Note : Same error for other aggregate functions like sum,min,max etc. 
> -> ERROR TRACE AS FOLLOW :
> [2016-08-17 21:32:45.925 IST] ERROR [] [] [] [] [] [] [] [] [] Thread-12 
> net.sourceforge.squirrel_sql.client.session.SQLExecuterTask [126] - Error 
> reading ResultSet for SQL: select count(*) from BUDGET
> org.apache.phoenix.exception.PhoenixIOException: 
> org.apache.phoenix.exception.PhoenixIOException: The system cannot find the 
> path specified
>   at 
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)
>   at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:558)
>   at 
> org.apache.phoenix.iterate.ConcatResultIterator.getIterators(ConcatResultIterator.java:50)
>   at 
> org.apache.phoenix.iterate.ConcatResultIterator.currentIterator(ConcatResultIterator.java:97)
>   at 
> org.apache.phoenix.iterate.ConcatResultIterator.next(ConcatResultIterator.java:117)
>   at 
> org.apache.phoenix.iterate.BaseGroupedAggregatingResultIterator.next(BaseGroupedAggregatingResultIterator.java:64)
>   at 
> org.apache.phoenix.iterate.UngroupedAggregatingResultIterator.next(UngroupedAggregatingResultIterator.java:39)
>   at 
> org.apache.phoenix.jdbc.PhoenixResultSet.next(PhoenixResultSet.java:771)
>   at 
> net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetWrapper._nextOnResultSet(ResultSetWrapper.java:95)
>   at 
> net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetWrapper.next(ResultSetWrapper.java:56)
>   at 
> net.sourceforge.squirrel_sql.fw.sql.ResultSetReader.readRow(ResultSetReader.java:182)
>   at 
> net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet.createRow(ResultSetDataSet.java:238)
>   at 
> net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet._setResultSet(ResultSetDataSet.java:204)
>   at 
> net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet.setSqlExecutionTabResultSet(ResultSetDataSet.java:127)
>   at 
> net.sourceforge.squirrel_sql.client.session.mainpanel.SQLExecutionHandler.sqlResultSetAvailable(SQLExecutionHandler.java:423)
>   at 
> net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.processResultSet(SQLExecuterTask.java:549)
>   at 
> net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.processQuery(SQLExecuterTask.java:414)
>   at 
> net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.run(SQLExecuterTask.java:212)
>   at 
> net.sourceforge.squirrel_sql.fw.util.TaskExecuter.run(TaskExecuter.java:82)
>   at java.lang.Thread.run(Thread.java:744)
> Caused by: java.util.concurrent.ExecutionException: 
> org.apache.phoenix.exception.PhoenixIOException: The system cannot find the 
> path specified
>   at java.util.concurrent.FutureTask.report(FutureTask.java:122)
>   at java.util.concurrent.FutureTask.get(FutureTask.java:202)
>   at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:554)
>   ... 18 more

[jira] [Created] (PHOENIX-3192) phoenix-spark dataframe issue with combination of "column family + column name"

2016-08-17 Thread Kalyan (JIRA)
Kalyan created PHOENIX-3192:
---

 Summary: phoenix-spark dataframe issue with combination of "column 
family + column name"
 Key: PHOENIX-3192
 URL: https://issues.apache.org/jira/browse/PHOENIX-3192
 Project: Phoenix
  Issue Type: Bug
Reporter: Kalyan
Assignee: Kalyan


1. create table with different column families with same column name

create table tbl_1 (rowkey varchar primary key, cf1.c1 varchar, cf1.c2 integer, 
cf2.c1 double, cf2.c2 boolean, cf3.c1 bigint);

2. insert sample data into table

3. create dataframe using phoenix table with different column names

val df1 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
"CF2.C2"), conf = configuration )

df1.show // this will work

4. create dataframe using phoenix table with same column names

val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
"CF2.C1"), conf = configuration )

df2.show// this will fail

5. reason currently we are not handled the dataframe solution fully (column 
family + column name). 

only works with (column name)


Exception:

scala> val df2 = sqlContext.phoenixTableAsDataFrame("tbl_1", Array("CF1.C1", 
"CF2.C1"), conf = configuration )
df2: org.apache.spark.sql.DataFrame = [C1: string, C1: double]

scala> df2.show
16/08/17 22:16:54 ERROR Executor: Exception in task 0.0 in stage 0.0 (TID 0)
scala.MatchError: 1.5 (of class java.lang.Double)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:295)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:294)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:260)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StructConverter.toCatalystImpl(CatalystTypeConverters.scala:250)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTypeConverters.scala:102)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$$anonfun$createToCatalystConverter$2.apply(CatalystTypeConverters.scala:401)
at 
org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
at 
org.apache.spark.sql.SQLContext$$anonfun$6.apply(SQLContext.scala:492)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
at scala.collection.Iterator$$anon$10.next(Iterator.scala:312)
at scala.collection.Iterator$class.foreach(Iterator.scala:727)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
at 
scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
at 
scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
at scala.collection.AbstractIterator.to(Iterator.scala:1157)
at 
scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
at 
scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
at 
org.apache.spark.sql.execution.SparkPlan$$anonfun$5.apply(SparkPlan.scala:212)
at 
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
at 
org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1858)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
16/08/17 22:16:54 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, 
localhost): scala.MatchError: 1.5 (of class java.lang.Double)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:295)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$StringConverter$.toCatalystImpl(CatalystTypeConverters.scala:294)
at 
org.apache.spark.sql.catalyst.CatalystTypeConverters$CatalystTypeConverter.toCatalyst(CatalystTy

[jira] [Commented] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424873#comment-15424873
 ] 

James Taylor commented on PHOENIX-3185:
---

Agree, [~kalyanhadoop], it's an important fix. Not worried about code format 
yet. I'd just like to get a test run across all tests to make sure your change 
doesn't introduce any regressions. That'll happen if you attach a patch like 
this as described in the link I mentioned: {{git format-patch --stdout origin > 
PHOENIX-{NUMBER}.patch}}

> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Fix For: 4.8.1
>
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread James Taylor (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424873#comment-15424873
 ] 

James Taylor edited comment on PHOENIX-3185 at 8/17/16 4:44 PM:


Agree, [~kalyanhadoop], it's an important fix. Not worried about code format 
yet. I'd just like to get a test run across all tests to make sure your change 
doesn't introduce any regressions. That'll happen if you attach a patch like 
this as described in the link I mentioned:
{code}
git format-patch --stdout origin > PHOENIX-{NUMBER}.patch
{code}


was (Author: jamestaylor):
Agree, [~kalyanhadoop], it's an important fix. Not worried about code format 
yet. I'd just like to get a test run across all tests to make sure your change 
doesn't introduce any regressions. That'll happen if you attach a patch like 
this as described in the link I mentioned: {{git format-patch --stdout origin > 
PHOENIX-{NUMBER}.patch}}

> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Fix For: 4.8.1
>
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread James Taylor (JIRA)

 [ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

James Taylor updated PHOENIX-3185:
--
Fix Version/s: 4.8.1

> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Fix For: 4.8.1
>
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (PHOENIX-3191) COUNT and other aggregate functions not working

2016-08-17 Thread Vivek K T (JIRA)
Vivek K T created PHOENIX-3191:
--

 Summary: COUNT and other aggregate functions not working
 Key: PHOENIX-3191
 URL: https://issues.apache.org/jira/browse/PHOENIX-3191
 Project: Phoenix
  Issue Type: Bug
Affects Versions: 4.6.0
 Environment: CentOS(Server), Windows(Client)
Reporter: Vivek K T
Priority: Critical


Following are steps to reproduce the issue : 

1) Create a dummy table BUDGET : 

create table BUDGET
(
  id varchar (50) primary key,
  date timestamp,
  amount decimal,
  currency varchar (50)
)

2) Insert data in BUDGET table : 

upsert into BUDGET (id,date,amount,currency) values 
('1',current_date(),20,'INR');
upsert into BUDGET (id,date,amount,currency) values 
('2',current_date(),40,'INR');
upsert into BUDGET (id,date,amount,currency) values 
('3',current_date(),60,'INR');
upsert into BUDGET (id,date,amount,currency) values 
('4',current_date(),80,'INR');
upsert into BUDGET (id,date,amount,currency) values 
('5',current_date(),100,'INR');

(PLEASE NOTE : TABLE CREATION AND DATA INSERTION SHOULD BE DONE FROM LINUX 
SERVER)

3) Fire following query from squirrel-sql on windows: 

select count(*) from BUDGET

Note : Same error for other aggregate functions like sum,min,max etc. 

-> ERROR TRACE AS FOLLOW :

[2016-08-17 21:32:45.925 IST] ERROR [] [] [] [] [] [] [] [] [] Thread-12 
net.sourceforge.squirrel_sql.client.session.SQLExecuterTask [126] - Error 
reading ResultSet for SQL: select count(*) from BUDGET
org.apache.phoenix.exception.PhoenixIOException: 
org.apache.phoenix.exception.PhoenixIOException: The system cannot find the 
path specified
at 
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)
at 
org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:558)
at 
org.apache.phoenix.iterate.ConcatResultIterator.getIterators(ConcatResultIterator.java:50)
at 
org.apache.phoenix.iterate.ConcatResultIterator.currentIterator(ConcatResultIterator.java:97)
at 
org.apache.phoenix.iterate.ConcatResultIterator.next(ConcatResultIterator.java:117)
at 
org.apache.phoenix.iterate.BaseGroupedAggregatingResultIterator.next(BaseGroupedAggregatingResultIterator.java:64)
at 
org.apache.phoenix.iterate.UngroupedAggregatingResultIterator.next(UngroupedAggregatingResultIterator.java:39)
at 
org.apache.phoenix.jdbc.PhoenixResultSet.next(PhoenixResultSet.java:771)
at 
net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetWrapper._nextOnResultSet(ResultSetWrapper.java:95)
at 
net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetWrapper.next(ResultSetWrapper.java:56)
at 
net.sourceforge.squirrel_sql.fw.sql.ResultSetReader.readRow(ResultSetReader.java:182)
at 
net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet.createRow(ResultSetDataSet.java:238)
at 
net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet._setResultSet(ResultSetDataSet.java:204)
at 
net.sourceforge.squirrel_sql.fw.datasetviewer.ResultSetDataSet.setSqlExecutionTabResultSet(ResultSetDataSet.java:127)
at 
net.sourceforge.squirrel_sql.client.session.mainpanel.SQLExecutionHandler.sqlResultSetAvailable(SQLExecutionHandler.java:423)
at 
net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.processResultSet(SQLExecuterTask.java:549)
at 
net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.processQuery(SQLExecuterTask.java:414)
at 
net.sourceforge.squirrel_sql.client.session.SQLExecuterTask.run(SQLExecuterTask.java:212)
at 
net.sourceforge.squirrel_sql.fw.util.TaskExecuter.run(TaskExecuter.java:82)
at java.lang.Thread.run(Thread.java:744)
Caused by: java.util.concurrent.ExecutionException: 
org.apache.phoenix.exception.PhoenixIOException: The system cannot find the 
path specified
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:202)
at 
org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:554)
... 18 more
Caused by: org.apache.phoenix.exception.PhoenixIOException: The system cannot 
find the path specified
at 
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)
at 
org.apache.phoenix.iterate.SpoolingResultIterator.(SpoolingResultIterator.java:151)
at 
org.apache.phoenix.iterate.SpoolingResultIterator.(SpoolingResultIterator.java:86)
at 
org.apache.phoenix.iterate.SpoolingResultIterator.(SpoolingResultIterator.java:64)
at 
org.apache.phoenix.iterate.SpoolingResultIterator$SpoolingResultIteratorFactory.newIterator(SpoolingResultIterator.java:81)
at 
org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:109)
at 
org.apache.phoenix.iterate.ParallelIterat

Re: [ANNOUNCE] Josh Elser joins Phoenix PMC

2016-08-17 Thread Stephen Jiang
Congratulations!  Well deserved recognition on your contribution to Phoenix!

On Wed, Aug 10, 2016 at 11:51 PM, rajeshb...@apache.org <
chrajeshbab...@gmail.com> wrote:

> Congratulations Josh!!!
>
> On Thu, Aug 11, 2016 at 9:17 AM, Josh Elser  wrote:
>
> > Thank you all! I look forward to working more closely with everyone :D
> >
> >
> > James Taylor wrote:
> >
> >> On behalf of the Apache Phoenix PMC, I'm pleased to announce that Josh
> >> Elser has accepted our invitation to become a member of the Apache
> Phoenix
> >> PMC. Recently he found and fixed licensing-related issues in our source
> >> and
> >> binary release, and prior to that he's been the force behind the Phoenix
> >> Query Server (PQS).
> >>
> >> Really excited to have you onboard, Josh. Welcome, and looking forward
> to
> >> continued collaboration.
> >>
> >> Regards,
> >> James
> >>
> >>
>


[jira] [Commented] (PHOENIX-3185) Error: ERROR 514 (42892): A duplicate column name was detected in the object definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 (state=42892,code=514)

2016-08-17 Thread Kalyan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3185?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424782#comment-15424782
 ] 

Kalyan commented on PHOENIX-3185:
-

Hi James Taylor, i think .. i need to follow the contributor guidelines for 
code format.

my intention is to verify the patch only ..  i feel it is a major bug.


> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> -
>
> Key: PHOENIX-3185
> URL: https://issues.apache.org/jira/browse/PHOENIX-3185
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Kalyan
>Assignee: Kalyan
> Attachments: image1.png, image2.png, image3.png, 
> phoenix_duplicate_column_check.patch
>
>
> create a table with duplicate columns.
> create table test_table (rowkey varchar primary key, c1 varchar, c2 integer, 
> c3 double, c1 bigint);
> The below exception getting .. no issue with exception. But the problem is 
> phoenix is creating table with exception & later we are not able to delete 
> the table also. We need to fix this Bug
> 0: jdbc:phoenix:localhost> create table test_table (rowkey varchar primary 
> key, c1 varchar, c2 integer, c3 double, c1 bigint);
> Error: ERROR 514 (42892): A duplicate column name was detected in the object 
> definition or ALTER TABLE statement. columnName=TEST_TABLE.C1 
> (state=42892,code=514)
> org.apache.phoenix.schema.ColumnAlreadyExistsException: ERROR 514 (42892): A 
> duplicate column name was detected in the object definition or ALTER TABLE 
> statement. columnName=TEST_TABLE.C1
>   at org.apache.phoenix.schema.PTableImpl.init(PTableImpl.java:415)
>   at org.apache.phoenix.schema.PTableImpl.(PTableImpl.java:315)
>   at org.apache.phoenix.schema.PTableImpl.makePTable(PTableImpl.java:288)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:2146)
>   at 
> org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:828)
>   at 
> org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:338)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:326)
>   at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:324)
>   at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1345)
>   at sqlline.Commands.execute(Commands.java:822)
>   at sqlline.Commands.sql(Commands.java:732)
>   at sqlline.SqlLine.dispatch(SqlLine.java:808)
>   at sqlline.SqlLine.begin(SqlLine.java:681)
>   at sqlline.SqlLine.start(SqlLine.java:398)
>   at sqlline.SqlLine.main(SqlLine.java:292)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-2722) support mysql "limit,offset" clauses

2016-08-17 Thread Vishal Mahajan (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-2722?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424636#comment-15424636
 ] 

Vishal Mahajan commented on PHOENIX-2722:
-

Can we also mention @since version in phoenix docs for this feature?

> support mysql "limit,offset" clauses 
> -
>
> Key: PHOENIX-2722
> URL: https://issues.apache.org/jira/browse/PHOENIX-2722
> Project: Phoenix
>  Issue Type: New Feature
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
>Priority: Minor
> Fix For: 4.8.0
>
> Attachments: PHOENIX-2722.patch, PHOENIX-2722_formatted.patch, 
> PHOENIX-2722_v1_rebased.patch
>
>
> For serial query(query with “serial" hint or  “limit" without "order by”), we 
> can limit each scan(using page filter) to “limit+offset” instead of limit 
> earlier.
> And then, for all queries, we can forward the relevant client iterators to 
> the offset provided and then return the result.
> syntax
> {code}
> [ LIMIT { count } ]
> [ OFFSET start [ ROW | ROWS ] ]
> [ FETCH { FIRST | NEXT } [ count ] { ROW | ROWS } ONLY ]
> {code}
> Some new keywords(OFFSET,FETCH,ROW, ROWS,ONLY) are getting introduced so 
> users might need to see that they are not using them as column name or 
> something.
> WDYT, [~jamestaylor]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] phoenix issue #186: cleanup sqlline*.py using python idioms.

2016-08-17 Thread btbytes
Github user btbytes commented on the issue:

https://github.com/apache/phoenix/pull/186
  
I'm looking into this. I tried splitting the triple quoted string on space 
but that isn't working. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (PHOENIX-3190) Creating Tracing Visualization Ducumentation

2016-08-17 Thread Nishani (JIRA)
Nishani  created PHOENIX-3190:
-

 Summary: Creating Tracing Visualization Ducumentation
 Key: PHOENIX-3190
 URL: https://issues.apache.org/jira/browse/PHOENIX-3190
 Project: Phoenix
  Issue Type: Sub-task
Affects Versions: 4.8.0
Reporter: Nishani 
Assignee: Nishani 
 Fix For: 4.8.0


Documentation will contain
-Improved Search Features
-How to start the Phoenix Zipkin Web App
-How to change the zookeeper host and port 
-Features of Phoenix Zipkin Web App




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (PHOENIX-3159) CachingHTableFactory may close HTable during eviction even if it is getting used for writing by another thread.

2016-08-17 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-3159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424082#comment-15424082
 ] 

Hadoop QA commented on PHOENIX-3159:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12824096/PHOENIX-3159_v3.patch
  against master branch at commit 657917bfb15ecd5bef4616458f2c8cd3ede0cf6b.
  ATTACHMENT ID: 12824096

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 15 new 
or modified tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 
34 warning messages.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 lineLengths{color}.  The patch introduces the following lines 
longer than 100:
+ synchronized (this) { // the whole operation of closing and 
checking the count should be atomic
+  public CachingHTableFactory(HTableFactory tableFactory, Configuration conf, 
RegionCoprocessorEnvironment env) {
+  public CachingHTableFactory(HTableFactory factory, int cacheSize, 
RegionCoprocessorEnvironment env) {
+this.pool = ThreadPoolManager.getExecutor(new 
ThreadPoolBuilder("CachedHtables", env.getConfiguration())
+  public HTableInterface getTable(ImmutableBytesPtr tablename, ExecutorService 
pool) throws IOException {
+public HTableInterface getTable(ImmutableBytesPtr 
tablename,ExecutorService pool) throws IOException {
+  public HTableInterface getTable(ImmutableBytesPtr tablename, ExecutorService 
pool) throws IOException;
+   public static final String INDEX_WRITES_THREAD_MAX_PER_REGIONSERVER_KEY = 
"phoenix.index.writes.threads.max";
+public static final String INDEX_WRITER_KEEP_ALIVE_TIME_CONF_KEY = 
"index.writer.threads.keepalivetime";
+void setup(HTableFactory factory, ExecutorService pool, Abortable 
abortable, Stoppable stop, int cacheSize, RegionCoprocessorEnvironment env) {

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/520//testReport/
Javadoc warnings: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/520//artifact/patchprocess/patchJavadocWarnings.txt
Console output: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/520//console

This message is automatically generated.

> CachingHTableFactory may close HTable during eviction even if it is getting 
> used for writing by another thread.
> ---
>
> Key: PHOENIX-3159
> URL: https://issues.apache.org/jira/browse/PHOENIX-3159
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Ankit Singhal
>Assignee: Ankit Singhal
> Fix For: 4.8.1
>
> Attachments: PHOENIX-3159.patch, PHOENIX-3159_v1.patch, 
> PHOENIX-3159_v2.patch, PHOENIX-3159_v3.patch
>
>
> CachingHTableFactory may close HTable during eviction even if it is getting 
> used for writing by another thread which results in writing thread to fail 
> and index is disabled.
> LRU eviction closing HTable or underlying connection when cache is full and 
> new HTable is requested.
> {code}
> 2016-08-04 13:45:21,109 DEBUG 
> [nat-s11-4-ioss-phoenix-1-5.openstacklocal,16020,1470297472814-index-writer--pool11-t35]
>  client.ConnectionManager$HConnectionImplementation: Closing HConnection 
> (debugging purposes only)
> java.lang.Exception
> at 
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.internalClose(ConnectionManager.java:2423)
> at 
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.close(ConnectionManager.java:2447)
> at 
> org.apache.hadoop.hbase.client.CoprocessorHConnection.close(CoprocessorHConnection.java:41)
> at 
> org.apache.hadoop.hbase.client.HTableWrapper.internalClose(HTableWrapper.java:91)
> at 
> org.apache.hadoop.hbase.client.HTableWrapper.close(HTableWrapper.java:107)
> at 
> org.apache.phoenix.hbase.index.table.CachingHTableFactory$HTableInterfaceLRUMap.removeLRU(CachingHTableFactory.java:61)
> at 
> org.apache.commons.collections.map.LRUMap.addMapping(LRUMap.java:256)
> at 
> org.apache.commons.collections.map.AbstractHashedMap.put(AbstractHashedMap.java:284)
> at 
> org.apache.phoenix.hbase.index.table.CachingHTableFactory.getTable(CachingHTableFactory.java:100)
> at 
> org.apache.phoenix.hbase.index.write.ParallelWriterIndexCommitter$1.call(ParallelWriterIndexCommitter.java:

[jira] [Commented] (PHOENIX-1751) Perform aggregations, sorting, etc, in the preScannerNext instead of postScannerOpen

2016-08-17 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/PHOENIX-1751?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15424019#comment-15424019
 ] 

Hadoop QA commented on PHOENIX-1751:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12824086/PHOENIX-1751.patch
  against master branch at commit 657917bfb15ecd5bef4616458f2c8cd3ede0cf6b.
  ATTACHMENT ID: 12824086

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 
34 warning messages.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 lineLengths{color}.  The patch does not introduce lines 
longer than 100

 {color:red}-1 core tests{color}.  The patch failed these unit tests:
 
./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.ViewIT
./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.SaltedViewIT
./phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.TenantSpecificTablesDMLIT

Test results: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/519//testReport/
Javadoc warnings: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/519//artifact/patchprocess/patchJavadocWarnings.txt
Console output: 
https://builds.apache.org/job/PreCommit-PHOENIX-Build/519//console

This message is automatically generated.

> Perform aggregations, sorting, etc, in the preScannerNext instead of 
> postScannerOpen
> 
>
> Key: PHOENIX-1751
> URL: https://issues.apache.org/jira/browse/PHOENIX-1751
> Project: Phoenix
>  Issue Type: Bug
>Reporter: Lars Hofhansl
> Attachments: 1751-WIP-v2.txt, 1751-WIP-v2b.patch, 1751-WIP.txt, 
> PHOENIX-1751-v2c.patch, PHOENIX-1751.patch
>
>
> HBase retains a lease for every scanner. Then lease expires the scan will no 
> longer (be allowed to) work. The leases guard against the client going away, 
> and allow cleaning up resources if that happens.
> At various points HBase "suspends" the lease while the region server are 
> working on behalf of this scanner, so that the lease won't expire even though 
> the server is working on it.
> HBase does that during the scanning process. Crucially it suspends the leaser 
> after the scanner is opened, before next() is issued on it.
> The outcome of all this is that Phoenix executes aggregates, sorts, etc, with 
> the lease in place, and hence if these take a bit the lease can expire even 
> though the server was working on it.
> Phoenix should do this work in preScannerNext, being careful that the 
> precalculation is only performed once.
> I'll attach a sample patch soon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)