Re: [PERFORM] Slow Count-Distinct Query

2014-04-01 Thread Christopher Jackson
Hi Bricklen,

   Thanks again for the feedback.  The concurrent refresh sounds cool.  I
just saw the 9.4 release is tentatively scheduled for later this year.  Do
you know what people have been doing for view refreshes in the meantime?

   Thanks


On Tue, Apr 1, 2014 at 11:48 PM, Michael Paquier
wrote:

> On Wed, Apr 2, 2014 at 1:22 PM, Christopher Jackson 
> wrote:
> >
> > Hi Bricklen,
> >
> > Thanks for the feedback.  I'll play around with materialized views.
>  My
> > understanding is they have to be manually triggered for refresh
> Yep.
>
> > and there's an exclusive lock on the view while the refresh is taking
> place.  Is this
> > your understanding as well?
> Re-yep.
>
> > I'm using PG 9.3.3.  If this is true, I'm
> > curious what clever ways people have come up with to mitigate any issues
> > with the lock.
> Kevin Grittner has implemented REFRESH MATERIALIZED VIEW CONCURRENTLY
> in 9.4. A unique index is needed on the materialized view as well to
> authorize this concurrent operation. It has the merit to allow SELECT
> operations on the matview during the refresh.
> --
> Michael
>


Re: [PERFORM] Slow Count-Distinct Query

2014-04-01 Thread Christopher Jackson
Hi Bricklen,

Thanks for the feedback.  I'll play around with materialized views.  My
understanding is they have to be manually triggered for refresh and there's
an exclusive lock on the view while the refresh is taking place.  Is this
your understanding as well?  I'm using PG 9.3.3.  If this is true, I'm
curious what clever ways people have come up with to mitigate any issues
with the lock.

   Thanks again,
  Chris


On Tue, Apr 1, 2014 at 7:34 PM, bricklen  wrote:

>
> On Sun, Mar 30, 2014 at 12:45 PM, Christopher Jackson 
> wrote:
>
>>   Hi all,
>>
>>   tl;dr - How can I speed up my count-distinct query?
>>
>
> Depending on how often you need to run that query and how important it is
> to you, if you are willing to accept a performance hit on
> INSERT/UPDATE/DELETE of the "participants" table, you could create a
> summary table containing just the count of unique email addresses or the
> list of unique email addresses populated via trigger on
> INSERT/UPDATE/DELETE of the  participants table. Another option is try out
> the new Materialized views (
> http://www.postgresql.org/docs/current/static/sql-creatematerializedview.html)
> available in 9.3.
>
>


Re: [PERFORM] Slow Count-Distinct Query

2014-03-31 Thread Christopher Jackson
   Tom and Shawn,

   Thanks for the feedback.  This has been helpful.  It's worth noting that
I was spiking this out on my local box using default memory utilization
settings.  I'll revisit this once we get our production box set up.  It's
good to know what the best practices are around the enable_bitmapscan and
seq_page_cost settings are.  Also, it's good to know that my hack wasn't
actually yielding anything.  I'll check back in once our production
environment is up and running.  For what it's worth, we're using Heroku and
we're thinking of going with the Standard Tengu tier as a start.  This will
give us 1.7GB of RAM, so hopefully bumping up the work_mem setting
shouldn't be a problem.  Does that make sense?

  Thanks for the help,
 Chris


On Mon, Mar 31, 2014 at 9:15 AM, Tom Lane  wrote:

> Christopher Jackson  writes:
> >   tl;dr - How can I speed up my count-distinct query?
>
> EXPLAIN doesn't provide a lot of visibility into what the Aggregate plan
> node is doing, but in this case what it's doing is an internal sort/uniq
> operation to implement the DISTINCT.  You didn't say what value of
> work_mem you're using, but it'd need to be probably 50-100MB to prevent
> that sort from spilling to disk (and therefore being slow).
>
> Note that the indexscan is actually *slower* than the seqscan so far as
> the table access is concerned; if the table were big enough to not fit
> in RAM, this would get very much worse.  So I'm not impressed with trying
> to force the optimizer's hand as you've done here --- it might be a nice
> plan now, but it's brittle.  See if a bigger work_mem improves matters
> enough with the regular plan.
>
> > *I'm concerned about setting the enable_bitmapscan and seq_page_cost
> values
> > because I'm not yet sure what the consequences are.  Can anyone enlighten
> > me on the recommended way to speed up this query?*
>
> Turning off enable_bitmapscan globally would be a seriously bad idea.
> Changing the cost settings to these values globally might be all right;
> it would amount to optimizing for all-in-memory cases, which might or
> might not be a good idea for your situation.  For that matter, greatly
> increasing work_mem globally is usually not thought to be smart either;
> remember that it's a per-sort-operation setting so you may need to
> provision a considerable multiple of the setting as physical RAM,
> depending on how many queries you expect to run concurrently.  So all in
> all you might be well advised to just set special values for this one
> query, whichever solution approach you use.
>
> I doubt you need the "where email=email" hack, in any case.  That isn't
> forcing the optimizer's decision in any meaningful fashion.
>
> regards, tom lane
>


[PERFORM] Slow Count-Distinct Query

2014-03-30 Thread Christopher Jackson
  Hi all,

  tl;dr - How can I speed up my count-distinct query?

  I apologize in advance if this question has been asked already.  I'm
finding the mailing list hard to navigate.  I'm trying to speed up a query
that will find a count of distinct emails with in a table using Postgres
9.3.3.  The name of the table is participants.  Our domain is set up such
that duplicate emails are allowed so long as a particular corresponding
column value is unique.


*TABLE participants*

*  id serial NOT NULL (primary key)*

*  email character varying(255)*

*  (other columns omitted)*



I have the following index defined:

*index_participants_on_email ON participants USING btree (email COLLATE
pg_catalog."default");*

The query I'm trying to run is select count(distinct email) from
participants.  I've also tried the *group by* equivalent.  *On a table size
of 2 million rows, the query takes about 1 minute to return.*  This is way
too long.  After running analyze, I see that the index is being ignored and
a full table scan is performed.

So, I tried running the following after dropping the index:
create index email_idx on participants(email) where email=email;
set enable_bitmapscan = false;
set seq_page_cost = 0.1;
set random_page_cost = 0.2;
create index email_idx_2 on participants(email);
cluster participants using email_idx_2;

With these settings in place, if I run *select count(distinct email) from
participants* I get

"Aggregate  (cost=29586.20..29586.21 rows=1 width=18) (actual
time=54243.643..54243.644 rows=1 loops=1)"
"  ->  Seq Scan on participants  (cost=0.00..24586.18 rows=208
width=18) (actual time=0.030..550.296 rows=208 loops=1)"
*"Total runtime: 54243.669 ms"*

When I run the following, I get MUCH better results
*select count(1) from (select email from participants where email=email
group by email) x;*

"Aggregate  (cost=1856.36..1856.37 rows=1 width=0) (actual
time=1393.573..1393.573 rows=1 loops=1)"
"  Output: count(1)"
"  ->  Group  (cost=0.43..1731.36 rows=1 width=18) (actual
time=0.052..1205.977 rows=208 loops=1)"
"Output: participants.email"
"->  Index Only Scan using email_idx on public.participants
 (cost=0.43..1706.36 rows=1 width=18) (actual time=0.050..625.248
rows=208 loops=1)"
"  Output: participants.email"
"  Heap Fetches: 208"
*"Total runtime: 1393.599 ms"*

This query has a weird where clause (email=email) because I'm trying to
force the analyzer's hand to use the index.

*I'm concerned about setting the enable_bitmapscan and seq_page_cost values
because I'm not yet sure what the consequences are.  Can anyone enlighten
me on the recommended way to speed up this query?*

 Thanks