Doh my mistake, I thought it was merged.  I was just running with the
upstream 3.7 daily.  Can I use this run as my baseline and then I can run
next time on the patch to show the % improvement?  I'll wipe everything and
try on the patch, any idea when it will be merged?

-b

On Wed, Apr 29, 2015 at 5:34 AM, Susant Palai <spa...@redhat.com> wrote:

> Hi Ben
>    I checked out the glusterfs process attaching gdb and I could not find
> the newer code. Can you confirm whether you took the new patch ? patch i:
> http://review.gluster.org/#/c/9657/
>
> Thanks,
> Susant
>
>
> ----- Original Message -----
> > From: "Susant Palai" <spa...@redhat.com>
> > To: "Benjamin Turner" <bennytu...@gmail.com>, "Nithya Balachandran" <
> nbala...@redhat.com>
> > Cc: "Shyamsundar Ranganathan" <srang...@redhat.com>
> > Sent: Wednesday, April 29, 2015 1:22:02 PM
> > Subject: Re: [Gluster-devel] Rebalance improvement design
> >
> > This is how it looks for 2000 file. each 1MB. Done rebalance on 2*2 + 2.
> >
> > OLDER:
> > [root@gprfs030 ~]# gluster v rebalance test1 status
> >                                     Node Rebalanced-files          size
> >                                     scanned      failures
> >                                     skipped               status   run
> >                                     time in secs
> >                                ---------      -----------   -----------
> >                                -----------   -----------   -----------
> >                                ------------     --------------
> >                                localhost             2000         1.9GB
> >                                3325             0             0
> >                                completed              63.00
> >                            gprfs032-10ge                0        0Bytes
> >                            2158             0             0
> >                            completed               6.00
> > volume rebalance: test1: success:
> > [root@gprfs030 ~]#
> >
> >
> > NEW:
> > [root@gprfs030 upstream_rebalance]# gluster v rebalance test1 status
> >                                     Node Rebalanced-files          size
> >                                     scanned      failures
> >                                     skipped               status   run
> >                                     time in secs
> >                                ---------      -----------   -----------
> >                                -----------   -----------   -----------
> >                                ------------     --------------
> >                                localhost             2000         1.9GB
> >                                2011             0             0
> >                                completed              12.00
> >                            gprfs032-10ge                0        0Bytes
> >                            0             0             0
> >                            failed               0.00 [Failed
> >                            because of a crash which I will address in
> next
> >                            patch]
> > volume rebalance: test1: success:
> >
> >
> > Just trying out replica behaviour for rebalance.
> >
> > Here is the volume info.
> > [root@gprfs030 ~]# gluster v i
> >
> > Volume Name: test1
> > Type: Distributed-Replicate
> > Volume ID: e12ef289-86f2-454a-beaa-72ea763dbada
> > Status: Started
> > Number of Bricks: 3 x 2 = 6
> > Transport-type: tcp
> > Bricks:
> > Brick1: gprfs030-10ge:/bricks/gprfs030/brick1
> > Brick2: gprfs032-10ge:/bricks/gprfs032/brick1
> > Brick3: gprfs030-10ge:/bricks/gprfs030/brick2
> > Brick4: gprfs032-10ge:/bricks/gprfs032/brick2
> > Brick5: gprfs030-10ge:/bricks/gprfs030/brick3
> > Brick6: gprfs032-10ge:/bricks/gprfs032/brick3
> >
> >
> >
> > ----- Original Message -----
> > > From: "Susant Palai" <spa...@redhat.com>
> > > To: "Benjamin Turner" <bennytu...@gmail.com>
> > > Cc: "Gluster Devel" <gluster-devel@gluster.org>
> > > Sent: Wednesday, April 29, 2015 1:13:04 PM
> > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > >
> > > Ben, will you be able to give rebal stat for the same configuration and
> > > data
> > > set with older rebalance infra ?
> > >
> > > Thanks,
> > > Susant
> > >
> > > ----- Original Message -----
> > > > From: "Susant Palai" <spa...@redhat.com>
> > > > To: "Benjamin Turner" <bennytu...@gmail.com>
> > > > Cc: "Gluster Devel" <gluster-devel@gluster.org>
> > > > Sent: Wednesday, April 29, 2015 12:08:38 PM
> > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > >
> > > > Hi Ben,
> > > >   Yes we were using pure dist volume. Will check in to your systems
> for
> > > >   more
> > > >   info.
> > > >
> > > > Can you please update which patch set you used ? In the mean time I
> will
> > > > do
> > > > one set of test with the same configuration on a small data set.
> > > >
> > > > Thanks,
> > > > Susant
> > > >
> > > >
> > > > ----- Original Message -----
> > > > > From: "Benjamin Turner" <bennytu...@gmail.com>
> > > > > To: "Nithya Balachandran" <nbala...@redhat.com>
> > > > > Cc: "Susant Palai" <spa...@redhat.com>, "Gluster Devel"
> > > > > <gluster-devel@gluster.org>
> > > > > Sent: Wednesday, April 29, 2015 2:13:05 AM
> > > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > > >
> > > > > I am not seeing the performance you were.  I am running on 500GB of
> > > > > data:
> > > > >
> > > > > [root@gqas001 ~]# gluster v rebalance testvol status
> > > > >                                               Node Rebalanced-files
> > > > >  size       scanned      failures       skipped
>  status
> > > > >  run
> > > > > time in secs
> > > > >                                             ---------
> -----------
> > > > > -----------   -----------   -----------   -----------
> > > > > ------------
> > > > >     --------------
> > > > >                                         localhost           129021
> > > > > 7.9GB        912104             0             0          in
> progress
> > > > >     10100.00
> > > > > gqas012.sbu.lab.eng.bos.redhat.com                0        0Bytes
> > > > > 1930312             0             0          in progress
> > > > > 10100.00
> > > > > gqas003.sbu.lab.eng.bos.redhat.com                0        0Bytes
> > > > > 1930312             0             0          in progress
> > > > > 10100.00
> > > > > gqas004.sbu.lab.eng.bos.redhat.com           128903         7.9GB
> > > > >  946730             0             0          in progress
> > > > >  10100.00
> > > > > gqas013.sbu.lab.eng.bos.redhat.com                0        0Bytes
> > > > > 1930312             0             0          in progress
> > > > > 10100.00
> > > > > gqas014.sbu.lab.eng.bos.redhat.com                0        0Bytes
> > > > > 1930312             0             0          in progress
> > > > > 10100.00
> > > > >
> > > > > Based on what I am seeing I expect this to take 2 days.  Was you
> rebal
> > > > > run
> > > > > on a pure dist volume?  I am trying on 2x2 + 2 new bricks.  Any
> idea
> > > > > why
> > > > > mine is taking so long?
> > > > >
> > > > > -b
> > > > >
> > > > >
> > > > >
> > > > > On Wed, Apr 22, 2015 at 1:10 AM, Nithya Balachandran
> > > > > <nbala...@redhat.com>
> > > > > wrote:
> > > > >
> > > > > > That sounds great. Thanks.
> > > > > >
> > > > > > Regards,
> > > > > > Nithya
> > > > > >
> > > > > > ----- Original Message -----
> > > > > > From: "Benjamin Turner" <bennytu...@gmail.com>
> > > > > > To: "Nithya Balachandran" <nbala...@redhat.com>
> > > > > > Cc: "Susant Palai" <spa...@redhat.com>, "Gluster Devel" <
> > > > > > gluster-devel@gluster.org>
> > > > > > Sent: Wednesday, 22 April, 2015 12:14:14 AM
> > > > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > > > >
> > > > > > I am setting up a test env now, I'll have some feedback for you
> this
> > > > > > week.
> > > > > >
> > > > > > -b
> > > > > >
> > > > > > On Tue, Apr 21, 2015 at 11:36 AM, Nithya Balachandran
> > > > > > <nbala...@redhat.com
> > > > > > >
> > > > > > wrote:
> > > > > >
> > > > > > > Hi Ben,
> > > > > > >
> > > > > > > Did you get a chance to try this out?
> > > > > > >
> > > > > > > Regards,
> > > > > > > Nithya
> > > > > > >
> > > > > > > ----- Original Message -----
> > > > > > > From: "Susant Palai" <spa...@redhat.com>
> > > > > > > To: "Benjamin Turner" <bennytu...@gmail.com>
> > > > > > > Cc: "Gluster Devel" <gluster-devel@gluster.org>
> > > > > > > Sent: Monday, April 13, 2015 9:55:07 AM
> > > > > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > > > > >
> > > > > > > Hi Ben,
> > > > > > >   Uploaded a new patch here:
> http://review.gluster.org/#/c/9657/.
> > > > > > >   We
> > > > > > >   can
> > > > > > > start perf test on it. :)
> > > > > > >
> > > > > > > Susant
> > > > > > >
> > > > > > > ----- Original Message -----
> > > > > > > From: "Susant Palai" <spa...@redhat.com>
> > > > > > > To: "Benjamin Turner" <bennytu...@gmail.com>
> > > > > > > Cc: "Gluster Devel" <gluster-devel@gluster.org>
> > > > > > > Sent: Thursday, 9 April, 2015 3:40:09 PM
> > > > > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > > > > >
> > > > > > > Thanks Ben. RPM is not available and I am planning to refresh
> the
> > > > > > > patch
> > > > > > in
> > > > > > > two days with some more regression fixes. I think we can run
> the
> > > > > > > tests
> > > > > > post
> > > > > > > that. Any larger data-set will be good(say 3 to 5 TB).
> > > > > > >
> > > > > > > Thanks,
> > > > > > > Susant
> > > > > > >
> > > > > > > ----- Original Message -----
> > > > > > > From: "Benjamin Turner" <bennytu...@gmail.com>
> > > > > > > To: "Vijay Bellur" <vbel...@redhat.com>
> > > > > > > Cc: "Susant Palai" <spa...@redhat.com>, "Gluster Devel" <
> > > > > > > gluster-devel@gluster.org>
> > > > > > > Sent: Thursday, 9 April, 2015 2:10:30 AM
> > > > > > > Subject: Re: [Gluster-devel] Rebalance improvement design
> > > > > > >
> > > > > > >
> > > > > > > I have some rebalance perf regression stuff I have been
> working on,
> > > > > > > is
> > > > > > > there an RPM with these patches anywhere so that I can try it
> on my
> > > > > > > systems? If not I'll just build from:
> > > > > > >
> > > > > > >
> > > > > > > git fetch git:// review.gluster.org/glusterfs
> > > > > > > refs/changes/57/9657/8
> > > > > > > &&
> > > > > > > git cherry-pick FETCH_HEAD
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > I will have _at_least_ 10TB of storage, how many TBs of data
> should
> > > > > > > I
> > > > > > > run
> > > > > > > with?
> > > > > > >
> > > > > > >
> > > > > > > -b
> > > > > > >
> > > > > > >
> > > > > > > On Tue, Apr 7, 2015 at 9:07 AM, Vijay Bellur <
> vbel...@redhat.com >
> > > > > > wrote:
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > On 04/07/2015 03:08 PM, Susant Palai wrote:
> > > > > > >
> > > > > > >
> > > > > > > Here is one test performed on a 300GB data set and around
> 100%(1/2
> > > > > > > the
> > > > > > > time) improvement was seen.
> > > > > > >
> > > > > > > [root@gprfs031 ~]# gluster v i
> > > > > > >
> > > > > > > Volume Name: rbperf
> > > > > > > Type: Distribute
> > > > > > > Volume ID: 35562662-337e-4923-b862- d0bbb0748003
> > > > > > > Status: Started
> > > > > > > Number of Bricks: 4
> > > > > > > Transport-type: tcp
> > > > > > > Bricks:
> > > > > > > Brick1: gprfs029-10ge:/bricks/ gprfs029/brick1
> > > > > > > Brick2: gprfs030-10ge:/bricks/ gprfs030/brick1
> > > > > > > Brick3: gprfs031-10ge:/bricks/ gprfs031/brick1
> > > > > > > Brick4: gprfs032-10ge:/bricks/ gprfs032/brick1
> > > > > > >
> > > > > > >
> > > > > > > Added server 32 and started rebalance force.
> > > > > > >
> > > > > > > Rebalance stat for new changes:
> > > > > > > [root@gprfs031 ~]# gluster v rebalance rbperf status
> > > > > > > Node Rebalanced-files size scanned failures skipped status run
> time
> > > > > > > in
> > > > > > secs
> > > > > > > --------- ----------- ----------- ----------- -----------
> > > > > > > -----------
> > > > > > > ------------ --------------
> > > > > > > localhost 74639 36.1GB 297319 0 0 completed 1743.00
> > > > > > > 172.17.40.30 67512 33.5GB 269187 0 0 completed 1395.00
> > > > > > > gprfs029-10ge 79095 38.8GB 284105 0 0 completed 1559.00
> > > > > > > gprfs032-10ge 0 0Bytes 0 0 0 completed 402.00
> > > > > > > volume rebalance: rbperf: success:
> > > > > > >
> > > > > > > Rebalance stat for old model:
> > > > > > > [root@gprfs031 ~]# gluster v rebalance rbperf status
> > > > > > > Node Rebalanced-files size scanned failures skipped status run
> time
> > > > > > > in
> > > > > > secs
> > > > > > > --------- ----------- ----------- ----------- -----------
> > > > > > > -----------
> > > > > > > ------------ --------------
> > > > > > > localhost 86493 42.0GB 634302 0 0 completed 3329.00
> > > > > > > gprfs029-10ge 94115 46.2GB 687852 0 0 completed 3328.00
> > > > > > > gprfs030-10ge 74314 35.9GB 651943 0 0 completed 3072.00
> > > > > > > gprfs032-10ge 0 0Bytes 594166 0 0 completed 1943.00
> > > > > > > volume rebalance: rbperf: success:
> > > > > > >
> > > > > > >
> > > > > > > This is interesting. Thanks for sharing & well done! Maybe we
> > > > > > > should
> > > > > > > attempt a much larger data set and see how we fare there :).
> > > > > > >
> > > > > > > Regards,
> > > > > > >
> > > > > > >
> > > > > > > Vijay
> > > > > > >
> > > > > > >
> > > > > > > ______________________________ _________________
> > > > > > > Gluster-devel mailing list
> > > > > > > Gluster-devel@gluster.org
> > > > > > > http://www.gluster.org/ mailman/listinfo/gluster-devel
> > > > > > >
> > > > > > > _______________________________________________
> > > > > > > Gluster-devel mailing list
> > > > > > > Gluster-devel@gluster.org
> > > > > > > http://www.gluster.org/mailman/listinfo/gluster-devel
> > > > > > > _______________________________________________
> > > > > > > Gluster-devel mailing list
> > > > > > > Gluster-devel@gluster.org
> > > > > > > http://www.gluster.org/mailman/listinfo/gluster-devel
> > > > > > >
> > > > > >
> > > > >
> > > > _______________________________________________
> > > > Gluster-devel mailing list
> > > > Gluster-devel@gluster.org
> > > > http://www.gluster.org/mailman/listinfo/gluster-devel
> > > >
> > > _______________________________________________
> > > Gluster-devel mailing list
> > > Gluster-devel@gluster.org
> > > http://www.gluster.org/mailman/listinfo/gluster-devel
> > >
> >
>
_______________________________________________
Gluster-devel mailing list
Gluster-devel@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-devel

Reply via email to