Hello
    
 When i start a  virtual machine,which image was stored  in Glusterfs  ,  i met 
a problem like this :
    
    
009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be 
manually corrected in backend
2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned 
error Input/output error
2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50149: 
OPEN() /rhel5-210/disk0 => -1 (Input/output error)
2009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be 
manually corrected in backend
2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned 
error Input/output error
2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50150: 
OPEN() /rhel5-210/disk0 => -1 (Input/output error)
2009-04-22 13:26:58 W [afr.c:609:afr_open] afr1: returning EIO, file has to be 
manually corrected in backend
2009-04-22 13:26:58 W [stripe.c:1869:stripe_open_cbk] bricks: afr1 returned 
error Input/output error
2009-04-22 13:26:58 E [fuse-bridge.c:662:fuse_fd_cbk] glusterfs-fuse: 50151: 
OPEN() /rhel5-210/disk0 => -1 (Input/output error)
2009-04-22 13:30:03 E [afr-self-heal-data.c:778:afr_sh_data_fix] afr1: Unable 
to resolve conflicting data of /rhel5-210/disk0. Pleas
e resolve manually by deleting the file /rhel5-210/disk0 from all but the 
preferred subvolume. Please consider 'option favorite-chil
d <>'

   so i delete the films in one side of afr mode , the log was :

 
2009-04-22 13:30:56 W 
[afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr1: path 
/rhel5-210/disk0 on subvolume client4 =>
 -1 (No such file or directory)
2009-04-22 13:30:56 W 
[afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr3: path 
/rhel5-210/disk0 on subvolume client6 =>
 -1 (No such file or directory)
2009-04-22 13:30:56 W 
[afr-self-heal-common.c:871:sh_missing_entries_lookup_cbk] afr2: path 
/rhel5-210/disk0 on subvolume client5 =>
 -1 (No such file or directory)
2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr1: 
sourcing file /rhel5-210/disk0 from client1 to other sin
ks
2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr3: 
sourcing file /rhel5-210/disk0 from client3 to other sin
ks
2009-04-22 13:30:56 W [afr-self-heal-data.c:615:afr_sh_data_open_cbk] afr2: 
sourcing file /rhel5-210/disk0 from client2 to other sin
ks
   
    At last  i found  the delete side  , file was larger than other side ,like  
that  was   3.1G  and this is  8.0G which was whole image size .
    Why the data was not the same with afr ?
 
 My  configure is this :
 
volume client1                      ##### 
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.92.184         # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick       # name of the remote volume
end-volume
volume client2                      #####
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.92.78         # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick        # name of the remote volume
end-volume
volume client3                      ##### 
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.92.190         # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick        # name of the remote volume
end-volume
volume client4                      ##### 
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.92.153        # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick        # name of the remote volume
end-volume
volume client5                      ##### 
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.93.13         # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick        # name of the remote volume
end-volume
volume client6                      ##### 
  type protocol/client
  option transport-type tcp
  option remote-host 172.20.92.209         # IP address of the remote brick
  option remote-port 6997
  option transport-timeout 10          # seconds to wait for a reply
  option remote-subvolume brick        # name of the remote volume
end-volume
volume afr1                         ##### 
  type cluster/afr
  subvolumes client1 client4
end-volume
volume afr2                         ##### 
  type cluster/afr
  subvolumes client2 client5
end-volume
volume afr3                         #####
  type cluster/afr
  subvolumes client3 client6
end-volume
volume bricks                       #####
  type cluster/stripe
  option block-size 256MB
  subvolumes afr1 afr2 afr3 
end-volume
 
2009-04-22 



eagleeyes 
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users

Reply via email to