Hi,
- 1 client
- 3 servers
One server had been down when i started writing test data. Once my writes
finished I brought up the third server and the self-heal/replication began.
However, I noticed that an 'ls -l' from the client hung. Is this normal to have
a client hang when 1 of 3 servers is healing ?
my client and server configs at the bottom.
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1859:afr_sh_entry_sync_prepare]
replicate: self-healing directory / from subvolume remote2 to 1 other
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod]
replicate: creating missing file /test_file2 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod]
replicate: creating missing file /test_file3 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod]
replicate: creating missing file /test_file on remote1
[2009-10-02 12:29:20] D [afr-self-heal-entry.c:1161:afr_sh_entry_impunge_mknod]
replicate: creating missing file /test_file4 on remote1
[2009-10-02 12:29:20] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync]
replicate: self-healing metadata of /test_file2 from remote2 to remote1
[2009-10-02 12:29:20] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare]
replicate: self-healing file /test_file2 from subvolume remote2 to 1 other
[2009-10-02 12:29:20] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync]
replicate: self-healing metadata of /test_file from remote2 to remote1
[2009-10-02 12:29:20] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare]
replicate: self-healing file /test_file from subvolume remote2 to 1 other
[2009-10-02 12:29:33] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync]
replicate: self-healing metadata of /test_file3 from remote2 to remote1
[2009-10-02 12:29:33] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare]
replicate: self-healing file /test_file3 from subvolume remote2 to 1 other
[2009-10-02 12:29:33] D [afr-self-heal-metadata.c:379:afr_sh_metadata_sync]
replicate: self-healing metadata of /test_file4 from remote2 to remote1
[2009-10-02 12:29:33] D [afr-self-heal-data.c:797:afr_sh_data_sync_prepare]
replicate: self-healing file /test_file4 from subvolume remote2 to 1 other
# client config
# file: /etc/glusterfs/glusterfs.vol
volume remote1
type protocol/client
option transport-type tcp
option remote-host rh1
option remote-subvolume brick
end-volume
volume remote2
type protocol/client
option transport-type tcp
option remote-host rh2
option remote-subvolume brick
end-volume
volume remote3
type protocol/client
option transport-type tcp
option remote-host rh3
option remote-subvolume brick
end-volume
volume replicate
type cluster/replicate
subvolumes remote1 remote2 remote3
end-volume
volume writebehind
type performance/write-behind
option window-size 1MB
subvolumes replicate
end-volume
volume cache
type performance/io-cache
option cache-size 512MB
subvolumes writebehind
end-volume
#
# server config
#
volume posix
type storage/posix
option directory /bigpartition
end-volume
volume locks
type features/locks
subvolumes posix
end-volume
volume brick
type performance/io-threads
option thread-count 8
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option auth.addr.brick.allow *
subvolumes brick
end-volume
Thanks!
=cm
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users