[Gluster-users] question about brick_dir/.glusterfs/00/00 is lost

2017-03-01 Thread songxin
Hi, I have question about the directory brick_dir/.glusterfs/00/00. I created a replicate gluster volume, which has two bricks, on two node. On one node I found that when I run "ls /gluster_mount_point" it show no file. But on another node when I run "ls /gluster_mount_point" it show all fil

Re: [Gluster-users] not to reconnect between client and server because of race condition

2016-11-24 Thread songxin
Hi Kaushal, It is great. This patch could fix my issue. Thanks, Xin At 2016-11-25 14:57:56, "Kaushal M" wrote: >On Fri, Nov 25, 2016 at 12:03 PM, songxin wrote: >> Hi Atin >> I found a problem, that is about client(glusterfs) will not trying to >>

Re: [Gluster-users] question about info and info.tmp

2016-11-24 Thread songxin
t 12:06 PM, songxin wrote: Hi Atin, Do you mean that you have the workaround applicable now? Or it will take time to design the workaround? If you have workaround now, could you share it to me ? If you end up in having a 0 byte info file you'd need to copy the same info file from other

Re: [Gluster-users] not to reconnect between client and server because of race condition

2016-11-24 Thread songxin
Hi Kaushal, Thank you for your reply. I will make sure whether this patch could fix my problem. Thanks, Xin At 2016-11-25 14:57:56, "Kaushal M" wrote: >On Fri, Nov 25, 2016 at 12:03 PM, songxin wrote: >> Hi Atin >> I found a problem, that is about client(glus

Re: [Gluster-users] question about info and info.tmp

2016-11-24 Thread songxin
me more time to pick this item up from my backlog. I believe we have a workaround applicable here too. On Thu, 24 Nov 2016 at 14:24, songxin wrote: Hi Atin, Actually, the glusterfs is used in my project. And our test team find this issue. So I want to make sure that whether you plan to fix

[Gluster-users] not to reconnect between client and server because of race condition

2016-11-24 Thread songxin
Hi Atin I found a problem, that is about client(glusterfs) will not trying to reconnect to server(glusterfsd) after disconnect. Actually, it seems caused by race condition. Precondition The glusterfs version is 3.7.6. I create a replicate volume using two node, A node and B node.One brick i

Re: [Gluster-users] question about info and info.tmp

2016-11-24 Thread songxin
ed a validation to fail delete request if one of the glusterd is down. I'll get back to you on this. On Mon, 21 Nov 2016 at 07:24, songxin wrote: Hi Atin, Thank you for your support. And any conclusions about this issue? Thanks, Xin 在 2016-11-16 20:59:05,"Atin Mukherjee&

[Gluster-users] question about "All subvolumes are down"

2016-11-23 Thread songxin
Hi everyone, I create a replicate volume using two nodes,A board and B board. A board ip:10.32.1.144 B board ip:10.32.0.48 One brick and mount point is on A board Another brick is on B board I found that I can't access the mount point because the disconnection happen between client and two

Re: [Gluster-users] question about info and info.tmp

2016-11-20 Thread songxin
rom the cluster. However we need to revisit this code to see if this function is anymore needed given we recently added a validation to fail delete request if one of the glusterd is down. I'll get back to you on this. On Mon, 21 Nov 2016 at 07:24, songxin wrote: Hi Atin, Thank you for you

Re: [Gluster-users] question about info and info.tmp

2016-11-20 Thread songxin
Hi Atin, Thank you for your support. And any conclusions about this issue? Thanks, Xin 在 2016-11-16 20:59:05,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 1:53 PM, songxin wrote: ok, thank you. 在 2016-11-15 16:12:34,"Atin Mukherjee" 写道: On Tue, Nov

Re: [Gluster-users] question about info and info.tmp

2016-11-16 Thread songxin
在 2016-11-16 20:59:05,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 1:53 PM, songxin wrote: ok, thank you. 在 2016-11-15 16:12:34,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 12:47 PM, songxin wrote: Hi Atin, I think the root caus

Re: [Gluster-users] question about glusterfs version migrate

2016-11-16 Thread songxin
ok,thank you for your reply. At 2016-11-16 17:59:34, "Serkan Çoban" wrote: >Below link has changes in each release. >https://github.com/gluster/glusterfs/tree/release-3.7/doc/release-notes > > >On Wed, Nov 16, 2016 at 11:49 AM, songxin wrote: >> Hi, &

[Gluster-users] question about glusterfs version migrate

2016-11-16 Thread songxin
Hi, I am planning to migrate from gluster 3.7.6 to gluster 3.7.10. So I have two questions below. 1.How could I know the changes in gluster 3.7.6 compared to gluster 3.7.10? 2.Does my application need any NBC changes? Thanks, Xin___ Gluster-users mailin

Re: [Gluster-users] question about info and info.tmp

2016-11-15 Thread songxin
ok, thank you. 在 2016-11-15 16:12:34,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 12:47 PM, songxin wrote: Hi Atin, I think the root cause is in the function glusterd_import_friend_volume as below. int32_t glusterd_import_friend_volume (dict_t *peer_data, si

Re: [Gluster-users] question about info and info.tmp

2016-11-14 Thread songxin
ny idea, Atin? Thanks, Xin 在 2016-11-15 12:07:05,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 8:58 AM, songxin wrote: Hi Atin, I have some clues about this issue. I could reproduce this issue use the scrip that mentioned in https://bugzilla.redhat.com/show_bug.cgi?id=1308487 .

Re: [Gluster-users] question about info and info.tmp

2016-11-14 Thread songxin
Hi Atin, Now I have known that the info and bricks/* are removed by the function glusterd_delete_stale_volume(). But I have not known how to solve this issue. Thanks, Xin 在 2016-11-15 12:07:05,"Atin Mukherjee" 写道: On Tue, Nov 15, 2016 at 8:58 AM, songxin wrote: Hi At

Re: [Gluster-users] question about info and info.tmp

2016-11-14 Thread songxin
写道: On Tue, Nov 15, 2016 at 8:58 AM, songxin wrote: Hi Atin, I have some clues about this issue. I could reproduce this issue use the scrip that mentioned in https://bugzilla.redhat.com/show_bug.cgi?id=1308487 . I really appreciate your help in trying to nail down this issue. While I a

Re: [Gluster-users] question about info and info.tmp

2016-11-14 Thread songxin
1 20:34:05,"Atin Mukherjee" 写道: On Fri, Nov 11, 2016 at 4:00 PM, songxin wrote: Hi Atin, Thank you for your support. Sincerely wait for your reply. By the way, could you make sure that the issue, file info is empty, cause by rename is interrupted in kernel? As per my

Re: [Gluster-users] question about info and info.tmp

2016-11-11 Thread songxin
Hi Atin, Thank you for your support. Sincerely wait for your reply. By the way, could you make sure that the issue, file info is empty, cause by rename is interrupted in kernel? Thanks, Xin 在 2016-11-11 15:49:02,"Atin Mukherjee" 写道: On Fri, Nov 11, 2016 at 1:15 PM, song

Re: [Gluster-users] question about info and info.tmp

2016-11-10 Thread songxin
在 2016-11-11 15:27:03,"Atin Mukherjee" 写道: On Fri, Nov 11, 2016 at 12:38 PM, songxin wrote: Hi Atin, Thank you for your reply. As you said that the info file can only be changed in the glusterd_store_volinfo() sequentially because of the big lock. I have found the simila

Re: [Gluster-users] question about info and info.tmp

2016-11-10 Thread songxin
Xin 在 2016-11-11 14:36:40,"Atin Mukherjee" 写道: On Fri, Nov 11, 2016 at 8:33 AM, songxin wrote: Hi Atin, Thank you for your reply. I have two questions for you. 1.Are the two files info and info.tmp are only to be created or changed in function glusterd_store_volinfo()? I did not

Re: [Gluster-users] question about info and info.tmp

2016-11-10 Thread songxin
landed up once but we couldn't reproduce it, so something is wrong with the atomic update here is what I guess. I'll be glad if you have a reproducer for the same and then we can dig into it further. On Thu, Nov 10, 2016 at 1:32 PM, songxin wrote: Hi, When I start the glusterd some

[Gluster-users] question about info and info.tmp

2016-11-10 Thread songxin
Hi, When I start the glusterd some error happened. And the log is following. [2016-11-08 07:58:34.989365] I [MSGID: 100030] [glusterfsd.c:2318:main] 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.7.6 (args: /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFO) [2016-

[Gluster-users] about directory .glusterfs/indices/xattrop/ after adding brick

2016-03-24 Thread songxin
Hi, I have a question about files in .glusterfs/indices/xattrop/. reproduce: 1.create a replicate volume using two brick 2.kill A brick process 3.create some files in mount point 4.run "gluster volume heal gv0 info" show some files need heal 5.ls B_brick/.glusterfs/indices/xattrop/ show gfid of

[Gluster-users] question about gluster volume heal info split-brain

2016-03-22 Thread songxin
Hi, I have a quesition about heal info split-brain. I know that the gfid mismatch is a kind of split-brain and the parent directory should be show split-brain. In my case the "gluster volume heal info split-brain" show that no file is split-brain, though same filename has diffetent gfid on two b

Re: [Gluster-users] question about remove-brick force

2016-03-21 Thread songxin
the brick forcefully if brick is not present there on peers? 3. Forcefully means it should not show any error even brick is present or not on the peers. Thanks, Xin At 2016-03-21 16:12:41, "Gaurav Garg" wrote: >Hi songxin, > >>> 1.what is the different between runi

Re: [Gluster-users] How to recover after one node breakdown

2016-03-21 Thread songxin
Hi, I see.Thank you for your reply. Thanks, Xin 在 2016-03-21 16:34:26,"Atin Mukherjee" 写道: > > >On 03/21/2016 01:30 PM, songxin wrote: >> Hi, >> Thank you for your reply. >> Could you help me to answer my questions as below. >> >> Now

Re: [Gluster-users] question about remove-brick force

2016-03-21 Thread songxin
Hi Gaurav, Thank you very much.It is very helpful for me. Thanks, Xin At 2016-03-21 16:12:41, "Gaurav Garg" wrote: >Hi songxin, > >>> 1.what is the different between runing "gluster volume remove-brick gv0 >>> replica 1 128.224.162.255:/data/brick

[Gluster-users] question about remove-brick force

2016-03-21 Thread songxin
Hi, When I run the command "gluster volume remove-brick gv0 replica 1 128.224.162.255:/data/brick/gv1 force" , it reture failed. my question: 1.what is the different between runing "gluster volume remove-brick gv0 replica 1 128.224.162.255:/data/brick/gv1" and "gluster volume remove-brick gv

Re: [Gluster-users] How to recover after one node breakdown

2016-03-21 Thread songxin
full work? 3."heal full" must be run on the node whose uuid is biggest in volume? Thanks, Xin At 2016-03-21 14:04:25, "Atin Mukherjee" wrote: > > >On 03/19/2016 06:50 AM, songxin wrote: >> Hi Gaurav Garg, >> >> Thank you for you reply.It is v

Re: [Gluster-users] How to recover after one node breakdown

2016-03-20 Thread songxin
er node. > > i.e: > On the unaffected node the peers directory should have an entry for the > failed node containing the uuid of the failed node. The glusterd.info file > should enable you to recreate the peer file on the failed node. > > > On 16 March 2016 at 09:25,

Re: [Gluster-users] How to recover after one node breakdown

2016-03-19 Thread songxin
;6) your volume will recover. >> >above steps are mandatory steps to recover failed node. > >Thanks, > >Regards, >Gaurav > >- Original Message - >From: "songxin" >To: "Alastair Neil" >Cc: gluster-users@gluster.org >Sent: Thursday, M

Re: [Gluster-users] about heal full

2016-03-18 Thread songxin
the command `gluster system:: uuid get` Put this uuid as well into the text file. Now execute # cat | sort The last uuid printed in this list is the one that corresponds to the highest uuid in the cluster. HTH, Krutika On Mon, Mar 14, 2016 at 12:49 PM, songxin wrote: Hi, I have cre

[Gluster-users] How to recover after one node breakdown

2016-03-18 Thread songxin
Hi, Now I face a problem. Reproduc step is as below. 1.I create a replicate volume using two brick on two board 2.start the volume 3.one board is breakdown and all files in the rootfs ,including /var/lib/glusterd/*,are lost. 4.reboot the board and ip is not change. My question: How to recovery th

Re: [Gluster-users] How to recover after one node breakdown

2016-03-18 Thread songxin
e? Thanks, Xin At 2016-03-19 02:25:58, "Gaurav Garg" wrote: >Hi songxin, > >both method are almost same for recovering the replicated volume. i forgot to >mentioned one steps: > > #gluster volume heal $vol full > >IMO this solution should also app

[Gluster-users] about heal full

2016-03-14 Thread songxin
Hi, I have create a replicate volume and I want to run "gluster volume heal gv0 full". I found that if I run "gluster volume heal gv0 full" on one board it always output err like below. Launching heal operation to perform full self heal on volume gv0 has been unsuccessful But If I

Re: [Gluster-users] about split-brain

2016-03-10 Thread songxin
gular file, correct? Could you confirm that? Answer: Yes.it is regular file. On Thu, Mar 10, 2016 at 1:03 PM, songxin wrote: Hi all, I have a file has a problem of gfid-mismatch as below. stat: cannot stat '/mnt/c//public_html/cello/ior_files/nameroot.ior': Input/output error Remo

[Gluster-users] about split-brain

2016-03-09 Thread songxin
Hi all, I have a file has a problem of gfid-mismatch as below. stat: cannot stat '/mnt/c//public_html/cello/ior_files/nameroot.ior': Input/output error Remote: getfattr -d -m . -e hex opt/lvmdir/c2/brick/public_html/cello/ior_files/nameroot.ior # file: opt/lvmdir/c2/brick/public_html/cello/

[Gluster-users] How to recovery a replicate volume

2016-03-06 Thread songxin
Hi all, I have a problem about how to recovery a replicate volume. precondition: glusterfs version:3.7.6 brick of A board :128.224.95.140:/data/brick/gv0 brick of B board:128.224.162.255:/data/brick/gv0 reproduce: 1.gluster peer probe 128.224.162.255

[Gluster-users] question about command "getfattr" in replicate volume

2016-03-02 Thread songxin
Hi, recondition: glusterfs version is 3.7.6 A node:128.224.95.140 B node:128.224.162.255 brick on A node:/data/brick/gv0 brick on B node:/data/brick/gv0 reproduce steps: 1.gluster peer probe 128.224.162.255

Re: [Gluster-users] about tail command

2016-03-02 Thread songxin
Thank you very much for your reply.It is very helpful for me. And I have one more question about "heal full" in glusterfs 3.7.6. the reproduce steps : A board:128.224.95.140 B board:128.224.162.255 1.gluster peer probe 128.224.162.255

Re: [Gluster-users] about tail command

2016-03-02 Thread songxin
:35, "Anuradha Talur" wrote: > > >- Original Message - >> From: "songxin" >> To: "gluster-user" >> Sent: Tuesday, March 1, 2016 7:19:23 PM >> Subject: [Gluster-users] about tail command >> >> Hi, >> >

[Gluster-users] about tail command

2016-03-01 Thread songxin
Hi, recondition: A node:128.224.95.140 B node:128.224.162.255 brick on A node:/data/brick/gv0 brick on B node:/data/brick/gv0 reproduce steps: 1.gluster peer probe 128.224.162.255

Re: [Gluster-users] faied start the glusterd after reboot

2016-02-28 Thread songxin
f it's a file, then it >was not placed there as part of snapshotting any volume. If it's a directory, >then did you try creating a snapshot with such a name. > >Regards, >Avra > >On 02/25/2016 05:10 PM, songxin wrote: >> If I run "reboot" on the a n

[Gluster-users] question about command "gluster volume heal gv full"

2016-02-26 Thread songxin
Hi, I want to know whether the command "gluster volume heal gv full" is sync or async. Is the volume heal complete when comand quite? If it is a async , how could I know when the heal is complete? Thanks, Xin___ Gluster-users mailing list Gluster-users

Re: [Gluster-users] faied start the glusterd after reboot

2016-02-25 Thread songxin
5日,19:05,Atin Mukherjee 写道: > > + Rajesh , Avra > >> On 02/25/2016 04:12 PM, songxin wrote: >> Thanks for your reply. >> >> Do I need check all files in /var/lib/glusterd/*? >> Must all files be same in A node and B node? > Yes, they should be iden

Re: [Gluster-users] faied start the glusterd after reboot

2016-02-25 Thread songxin
gt;/var/lib/glusterd/* from board A? > >~Atin > >On 02/25/2016 03:48 PM, songxin wrote: >> Hi, >> I have a problem as below when I start the gluster after reboot a board. >> >> precondition: >> I use two boards do this test. >> The version of glusterf

[Gluster-users] faied start the glusterd after reboot

2016-02-25 Thread songxin
Hi, I have a problem as below when I start the gluster after reboot a board. precondition: I use two boards do this test. The version of glusterfs is 3.7.6. A board ip:128.224.162.255 B board ip:128.224.95.140 reproduce steps: 1.systemctl start glusterd (A board) 2.systemctl start glusterd

Re: [Gluster-users] question about replicate volume

2016-02-23 Thread songxin
//run on A node 8.gluster volume heal gv0 info fulll At step 7, should some split-brain entries be presented? 在 2016-02-24 12:55:40,"Ravishankar N" 写道: On 02/24/2016 10:21 AM, songxin wrote: Before step 6, there are some files(a,b,c), that are created at step 5 ,

Re: [Gluster-users] question about replicate volume

2016-02-23 Thread songxin
在 2016-02-24 12:42:39,"Ravishankar N" 写道: Hello, On 02/24/2016 10:03 AM, songxin wrote: Hi, Thank you for answering my question. And I have another question to ask. If there has been some file(c, d, e) in the B node brick before step 6 as below.And the file c is diffetent with file

Re: [Gluster-users] question about replicate volume

2016-02-23 Thread songxin
and B brick? 在 2016-02-24 12:11:09,"Ravishankar N" 写道: On 02/24/2016 07:16 AM, songxin wrote: Hi all, I have a question about replicate volume as below. precondition: 1.A node ip: 128.224.162.163 2.B node ip:128.224.162.255 3.A node brick:/data/brick/gv0 4.B node brick:/data

[Gluster-users] question about replicate volume

2016-02-23 Thread songxin
Hi all, I have a question about replicate volume as below. precondition: 1.A node ip: 128.224.162.163 2.B node ip:128.224.162.255 3.A node brick:/data/brick/gv0 4.B node brick:/data/brick/gv0 reproduce step: 1.gluster peer probe 128.224.162.255

Re: [Gluster-users] two same ip addr in peer list

2016-02-19 Thread songxin
nodes. > > > following things will be very useful for analysing this issue. > > You can restart your glusterd as of now as a workaround but we need to > analysis this issue further. > > > Thanks, > > ~Gaurav > > - Original Message - > From: &

[Gluster-users] two same ip addr in peer list

2016-02-19 Thread songxin
Hi, I create a replicate volume with 2 brick.And I frequently reboot my two nodes and frequently run “peer detach” “peer detach” “add-brick” "remove-brick". A borad ip: 10.32.0.48 B borad ip: 10.32.1.144 After that, I run "gluster peer status" on A board and it show as below. Number of Peer

Re: [Gluster-users] question about sync replicate volume after rebooting one node

2016-02-17 Thread songxin
Do you mean that I will delete the info file on B node and then start the glusterd?Or copy it from A node to B node? 发自我的 iPhone > 在 2016年2月17日,14:59,Atin Mukherjee 写道: > > > >> On 02/17/2016 11:44 AM, songxin wrote: >> Hi拢卢 >> The version of glusterfs on A

Re: [Gluster-users] question about sync replicate volume after rebooting one node

2016-02-16 Thread songxin
t; >- Original Message - >> From: "songxin" >> To: "Atin Mukherjee" >> Cc: "Anuradha Talur" , gluster-users@gluster.org >> Sent: Wednesday, February 17, 2016 11:44:14 AM >> Subject: Re:Re: [Gluster-users] question about sync replicate

Re: [Gluster-users] question about sync replicate volume after rebooting one node

2016-02-16 Thread songxin
mance.readdir-ahead=on brick-0=128.224.162.255:-data-brick-gv0 brick-1=128.224.162.163:-home-wrsadmin-work-tmp-data-brick-gv0 Thanks, Xin At 2016-02-17 12:01:37, "Atin Mukherjee" wrote: > > >On 02/17/2016 08:23 AM, songxin wrote: >> Hi, >> Thank you for your immedia

Re: [Gluster-users] question about sync replicate volume after rebooting one node

2016-02-16 Thread songxin
ke glusterd and glusterfs. Thanks, Xin At 2016-02-16 18:53:03, "Anuradha Talur" wrote: > > >- Original Message - >> From: "songxin" >> To: gluster-users@gluster.org >> Sent: Tuesday, February 16, 2016 3:59:50 PM >> Subject: [Gluster-

[Gluster-users] question about sync replicate volume after rebooting one node

2016-02-16 Thread songxin
Hi, I have a question about how to sync volume between two bricks after one node is reboot. There are two node, A node and B node.A node ip is 128.124.10.1 and B node ip is 128.124.10.2. operation steps on A node as below 1.gluster peer probe 128.124.10.2 2.mkdir -p /data/brick/gv0 3.gluster

[Gluster-users] question about creating a replicate volume

2016-02-13 Thread songxin
Hi, I have a question about creating a replicate volume with two bricks as below. There are two node, A node and B node.A node ip is 128.124.10.1 and B node ip is 128.124.10.2. operation steps on A node as below 1.gluster peer probe 128.124.10.2 2.mkdir -p /data/brick/gv0 3.cteate two files,

[Gluster-users] glusterfs replica volume out-sync

2016-02-05 Thread songxin
This is regarding glusterfs(3.7.6) issue we are facing at our end. We have a logging file which saves logs of the events for two node and this file are in sync using replica volume. When we restart the nodes , we see that log file of one board is not in the sync . How to reproduce: 1.Cr

[Gluster-users] sync lost between two boards

2016-02-05 Thread songxin
Hi, I use glusterfs (version 3.7.6) in replicate mode for sync between two boards in a node. When one of the board is locked and replaced with new board and restarted we see that sync is lost between the two boards.The mounted glusterfs volume is not present on the replaced board. Output of

[Gluster-users] Glusterfs :sync lost between two boards

2016-02-04 Thread songxin
Hi, I use glusterfs (version 3.7.6) in replicate mode for sync between two boards in a node. When one of the board is locked and replaced with new board and restarted we see that sync is lost between the two boards.The mounted glusterfs volume is not present on the replaced board. Output of

[Gluster-users] replica volume is out-sync after reboot

2016-02-04 Thread songxin
This is regarding glusterfs(3.7.6) issue we are facing at our end. We have a logging file which saves logs of the events for two node and this file are in sync using replica volume. When we restart the nodes , we see that log file of one board is not in the sync . How to reproduce: 1.Cr