[
https://issues.apache.org/jira/browse/PHOENIX-3601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15830604#comment-15830604
]
Josh Elser commented on PHOENIX-3601:
-------------------------------------
bq. Using Josh Elser's take on the very cool
https://github.com/joshelser/phoenix-performance toolset, I generated about
114M rows of TPC-DS data on a 5 RegionServer setup. I used a load-factor of 5,
which created a 256-way split table we'll refer to as SALES. I also created a
new table, pre-salted with 5 buckets we'll call SALES2 and UPSERT SELECTed the
data over. Both tables had major compaction and UPDATE STATISTICS run on them
as well.
Sick. I'm glad you found it useful :). 30-40% decrease in execution time is
awesome!
I don't have a good control over spark myself to do some testing, but maybe
this is a good reason for me to find the time to mess around with it...
> PhoenixRDD doesn't expose the preferred node locations to Spark
> ---------------------------------------------------------------
>
> Key: PHOENIX-3601
> URL: https://issues.apache.org/jira/browse/PHOENIX-3601
> Project: Phoenix
> Issue Type: Improvement
> Affects Versions: 4.8.0
> Reporter: Josh Mahonin
> Assignee: Josh Mahonin
> Attachments: PHOENIX-3601.patch
>
>
> Follow-up to PHOENIX-3600, in order to let Spark know the preferred node
> locations to assign partitions to, we need to update PhoenixRDD to retrieve
> the underlying node location information from the splits.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)