On Thu, Jan 7, 2016 at 2:47 PM, Steve Singer <st...@ssinger.info> wrote:

> On Thu, 7 Jan 2016, Tory M Blue wrote:
>
>
>> So I'm backing up in a big way. I know what started it, "adding a new
>> insert slave which took 13 hours to complete (indexes
>> etc)".. But now it doesn't appear I am able to catch up. I see the slave
>> doing what it's suppose to, get a bunch of data,
>> truncate the sl_log files move on. But the master is having a hard time.
>>
>> Postgres 9.4.5 and Slony 2.2.3
>>
>> All other nodes don't have any errors or issues.
>>
>> this is Node 1 (the master)
>> node 2 is a slave
>> node 3-5 are query slaves with only 1 of 3 sets being replicated too.
>>
>> I have interval at 5 minutes and sync_group_maxsize=50
>>
>> Any suggestions on where to thump it. At some point this will cause
>> issues on my master and when I see that starting, I'll
>> have to drop node 2 again, and when i add it, it will take 13+ hours and
>> I'll be back in the same position :)
>>
>
> Bump sync_group_maxsize to be much bigger, I'm not saying that will solve
> the problem but it might help(max allowed is 10,000). I'm also suspect when
> you say your have a sync_interval of 5 minutes, since I thought 60 seconds
> was the largest allowed.
>
> My apologies

cleanup_interval="5 minutes"

my interval is 1000ms
And sync, group cites 100 is the max

# Range:  [0,100], default: 6

sync_group_maxsize=50
_______________________________________________
Slony1-general mailing list
Slony1-general@lists.slony.info
http://lists.slony.info/mailman/listinfo/slony1-general

Reply via email to