API on a table with 4-way parallel scans and with 4 executors set for the
job. Thanks for the inputs.
Kanagha
On Thu, Aug 17, 2017 at 7:17 AM, Josh Mahonin <jmaho...@gmail.com> wrote:
> Hi,
>
> Phoenix is able to parallelize queries based on the underlying HBase
> region spli
Also, I'm using phoenixTableAsDataFrame API to read from a pre-split
phoenix table. How can we ensure read is parallelized across all executors?
Would salting/pre-splitting tables help in providing parallelism?
Appreciate any inputs.
Thanks
Kanagha
On Wed, Aug 16, 2017 at 10:16 PM, kanagha
Hi Josh,
Per your previous post, it is mentioned "The phoenix-spark parallelism is
based on the splits provided by the Phoenix query planner, and has no
requirements on specifying partition columns or upper/lower bounds."
Does it depend upon the region splits on the input table for parallelism?
Hi,
I'm trying to find what is the approach to read from multi-tenant phoenix
table via phoenix-spark plugin in java.
For scala, I see the following example.
https://github.com/apache/phoenix/blob/master/phoenix-spark/src/it/scala/org/apache/phoenix/spark/PhoenixSparkITTenantSpecific.scala
Sorry for the confusion.
metric_type,
timestamp,
metricId is defined as the primary key via Phoenix for metric_table.
Thanks
Kanagha
On Mon, Oct 3, 2016 at 3:41 PM, Michael McAllister <mmcallis...@homeaway.com
> wrote:
> >
>
> there is no indexing available
+ user@phoenix.apache.org
Kanagha
On Mon, Oct 3, 2016 at 1:31 PM, Kanagha <er.kana...@gmail.com> wrote:
> Hi,
>
> We have designed an metric_table, for ex:
>
> metric_type,
> timestamp,
> metricId
>
> in HBase using Apache Phoenix. And there is no indexing avail