[Gluster-users] errors after removing bricks
Hi, I had a 4 nodes in replication setup. After removing one of the nodes with: gluster> volume remove-brick images_1 replica 3 vmhost5:/exports/1 Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y Remove Brick commit force successful I started to see a high cpu usage on 2 of the remaining boxes, and this error in glustershd.log: [2013-04-18 12:40:47.160559] E [afr-self-heald.c:685:_link_inode_update_loc] 0-images_1-replicate-0: inode link failed on the inode (----) [2013-04-18 12:41:55.784510] I [afr-self-heald.c:1082:afr_dir_exclusive_crawl] 0-images_1-replicate-0: Another crawl is in progress for images_1-client-1 any ideas? ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Removing brick from replicated
Hi, We have a setup of 4 nodes/bricks in a replicated setup: Number of Bricks: 1 x 4 = 4 Transport-type: tcp Bricks: Brick1: vmhost2:/exports/1 Brick2: vmhost3:/exports/1 Brick3: vmhost5:/exports/1 Brick4: vmhost6:/exports/1 I'm trying to remove one of them (vmhost5): gluster> volume remove-brick images_1 replica 3 vmhost5:/exports/1 start Remove Brick start unsuccessful gluster> volume remove-brick images_1 replica 3 vmhost5:/exports/1 status Volume images_1 is not a distribute volume or contains only 1 brick. Not performing rebalance gluster> volume info images_1 Volume Name: images_1 Type: Distributed-Replicate Volume ID: 30c9c2d8-94c9-4b59-b647-0dc593d903d3 Status: Started Number of Bricks: 1 x 3 = 4 Transport-type: tcp Bricks: Brick1: vmhost2:/exports/1 Brick2: vmhost3:/exports/1 Brick3: vmhost5:/exports/1 Brick4: vmhost6:/exports/1 A few questions that hopefully someone can help me: Why does it say "Type: Distributed-Replicate" and "Number of Bricks 1 x 3 = 4" if ,after executing the "remove-brick" command, I got: unsuccessful What am I doing wrong? :) Thanks! ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Performance for KVM images (qcow)
Thanks for you help! I'll need to figure out a way to recreate the volume as replica 2 without messing with the running vms. On Wed, Apr 10, 2013 at 7:25 PM, Bryan Whitehead wrote: > I just want to say that trying to run a VM on almost any glusterfs > mounted fuse is going to suck when using 1G nic cards. > > That said your setup looks fine except you need to change "replica 4" > to "replica 2". I'm assuming you want redundancy and speed. > Replicating to all 4 bricks is probably not what you wanted. By > setting replica to 2 you'll be able to lose any 1 of the 4 bricks and > still run without data missing. > > gluster volume create replica 2 <... list all 4 bricks> > > > > On Tue, Apr 9, 2013 at 3:49 AM, Stephan von Krawczynski > wrote: > > On Tue, 09 Apr 2013 03:13:10 -0700 > > Robert Hajime Lanning wrote: > > > >> On 04/09/13 01:17, Eyal Marantenboim wrote: > >> > Hi Bryan, > >> > > >> > We have 1G nics on all our servers. > >> > Do you think that changing our design to distribute-replicate will > >> > improve the performance? > >> > > >> > Anything in the gluster performance settings that you think I should > change? > >> > >> With GlusterFS, almost all the processing is in the client side. This > >> includes replication. So, when you have "replica 4", the client will be > >> duplicating all transactions 4 times, synchronously. Your 1G ethernet > >> just became 256M. > > > > Let me drop in that no clear mind does it this way. Obviously one would > give > > the client more physical network cards, best choice as many as there are > > replications, and do the subnetting accordingly. > > > > -- > > Regards, > > Stephan > > ___ > > Gluster-users mailing list > > Gluster-users@gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ > Gluster-users mailing list > Gluster-users@gluster.org > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] Performance for KVM images (qcow)
Hi Bryan, We have 1G nics on all our servers. Do you think that changing our design to distribute-replicate will improve the performance? Anything in the gluster performance settings that you think I should change? thanks! On Tue, Apr 9, 2013 at 6:00 AM, Bryan Whitehead wrote: > This looks like you are replicating every file to all bricks? > > What is tcp running on? 1G nics? 10G? IPoIB (40-80G)? > > I think you want to have Distribute-Replicate. So 4 bricks with replica = > 2. > > Unless you are running at least 10G nics you are going to have serious > IO issues in your KVM/qcow2 VM's. > > On Mon, Apr 8, 2013 at 7:11 AM, Eyal Marantenboim > wrote: > > Hi, > > > > > > We have a set of 4 gluster nodes, all in replicated (design?) > > > > We use it to store our qcow2 images for kvm. These images have a variable > > IO, though most of them are for reading only. > > > > > > I tried to find some documentation re. performance optimization, but it's > > either unclear to me, or I couldn't find much.. so I'd copied from the > > internet and tried to adjust the config to our needs, but I'm sure it's > not > > optimized. > > > > > > We're using 3.3.1 on top of XFS. > > > > The qcow images are about 30GB (a couple of 100GB). > > > > > > Can someone please tell what would be the best paramenter for > performance to > > look at? > > > > > > here is volume info: > > > > > > Volume Name: images > > > > Type: Replicate > > > > Volume ID: > > > > Status: Started > > > > Number of Bricks: 1 x 4 = 4 > > > > Transport-type: tcp > > > > Bricks: > > > > Brick1: vmhost2:/exports/1 > > > > Brick2: vmhost3:/exports/1 > > > > Brick3: vmhost5:/exports/1 > > > > Brick4: vmhost6:/exports/1 > > > > Options Reconfigured: > > > > performance.cache-max-file-size: 1GB > > > > nfs.disable: on > > > > performance.cache-size: 4GB > > > > performance.cache-refresh-timeout: 1 > > > > performance.write-behind-window-size: 2MB > > > > performance.read-ahead: on > > > > performance.write-behind: on > > > > performance.io-cache: on > > > > performance.stat-prefetch: on > > > > performance.quick-read: on > > > > performance.io-thread-count: 64 > > > > performance.flush-behind: on > > > > features.quota-timeout: 1800 > > > > features.quota: off > > > > > > Thanks in advanced. > > > > > > ___ > > Gluster-users mailing list > > Gluster-users@gluster.org > > http://supercolony.gluster.org/mailman/listinfo/gluster-users > ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] Performance for KVM images (qcow)
Hi, We have a set of 4 gluster nodes, all in replicated (design?) We use it to store our qcow2 images for kvm. These images have a variable IO, though most of them are for reading only. I tried to find some documentation re. performance optimization, but it's either unclear to me, or I couldn't find much.. so I'd copied from the internet and tried to adjust the config to our needs, but I'm sure it's not optimized. We're using 3.3.1 on top of XFS. The qcow images are about 30GB (a couple of 100GB). Can someone please tell what would be the best paramenter for performance to look at? here is volume info: Volume Name: images Type: Replicate Volume ID: Status: Started Number of Bricks: 1 x 4 = 4 Transport-type: tcp Bricks: Brick1: vmhost2:/exports/1 Brick2: vmhost3:/exports/1 Brick3: vmhost5:/exports/1 Brick4: vmhost6:/exports/1 Options Reconfigured: performance.cache-max-file-size: 1GB nfs.disable: on performance.cache-size: 4GB performance.cache-refresh-timeout: 1 performance.write-behind-window-size: 2MB performance.read-ahead: on performance.write-behind: on performance.io-cache: on performance.stat-prefetch: on performance.quick-read: on performance.io-thread-count: 64 performance.flush-behind: on features.quota-timeout: 1800 features.quota: off Thanks in advanced. ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users