Please use 2.0.3 stable, or upgrade to the next rc2 until then. This has been fixed in rc2.
Avati On Mon, Jul 6, 2009 at 8:31 AM, eagleeyes<eaglee...@126.com> wrote: > HI > I use gluster2.0.3rc1 with fuse 2.8 in kernel > 2.6.30(SUSE Linux Enterprise Server 10 SP1 with kernel 2.6.30 ) . the mount > message was : > > /dev/hda4 on /data type reiserfs (rw,user_xattr) > glusterfs-client.vol.dht on /home type fuse.glusterfs (rw,allow_other,default_permissions,max_read=131072) > > > > There was some error when i "touce 111" in gluster directory ,the error was > : > /home: Transport endpoint is not connected > > pending frames: > patchset: e0db4ff890b591a58332994e37ce6db2bf430213 > signal received: 11 > configuration details:argp 1 > backtrace 1 > dlfcn 1 > fdatasync 1 > libpthread 1 > llistxattr 1 > setfsid 1 > spinlock 1 > epoll.h 1 > xattr.h 1 > st_atim.tv_nsec 1 > package-string: glusterfs 2.0.3rc1 > [0xffffe400] > /lib/glusterfs/2.0.3rc1/xlator/mount/fuse.so[0xb75c6288] > /lib/glusterfs/2.0.3rc1/xlator/performance/write-behind.so(wb_create_cbk+0xa7)[0xb75ccad7] > /lib/glusterfs/2.0.3rc1/xlator/performance/io-cache.so(ioc_create_cbk+0xde)[0xb7fbe8ae] > /lib/glusterfs/2.0.3rc1/xlator/performance/read-ahead.so(ra_create_cbk+0x167)[0xb7fc78b7] > /lib/glusterfs/2.0.3rc1/xlator/cluster/dht.so(dht_create_cbk+0xf7)[0xb75e25b7] > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(client_create_cbk+0x2ad)[0xb76004ad] > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_interpret+0x1ef)[0xb75ef8ff] > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_pollin+0xcf)[0xb75efaef] > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(notify+0x1ec)[0xb75f6ddc] > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_poll_in+0x3b)[0xb75b775b] > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_handler+0xae)[0xb75b7b8e] > /lib/libglusterfs.so.0[0xb7facbda] > /lib/libglusterfs.so.0(event_dispatch+0x21)[0xb7fabac1] > glusterfs(main+0xc2e)[0x804b6ae] > /lib/libc.so.6(__libc_start_main+0xdc)[0xb7e6087c] > glusterfs[0x8049c11] > --------- > > the server configuration > > gfs1:/ # cat /etc/glusterfs/glusterfsd-sever.vol > volume posix1 > type storage/posix # POSIX FS translator > option directory /data/data1 # Export this directory > end-volume > volume posix2 > type storage/posix # POSIX FS translator > option directory /data/data2 # Export this directory > end-volume > volume posix3 > type storage/posix # POSIX FS translator > option directory /data/data3 # Export this directory > end-volume > volume posix4 > type storage/posix # POSIX FS translator > option directory /data/data4 # Export this directory > end-volume > volume posix5 > type storage/posix # POSIX FS translator > option directory /data/data5 # Export this directory > end-volume > volume posix6 > type storage/posix # POSIX FS translator > option directory /data/data6 # Export this directory > end-volume > volume posix7 > type storage/posix # POSIX FS translator > option directory /data/data7 # Export this directory > end-volume > volume posix8 > type storage/posix # POSIX FS translator > option directory /data/data8 # Export this directory > end-volume > volume brick1 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix1 > end-volume > volume brick2 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix2 > end-volume > volume brick3 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix3 > end-volume > volume brick4 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix4 > end-volume > volume brick5 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix5 > end-volume > volume brick6 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix6 > end-volume > volume brick7 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix7 > end-volume > volume brick8 > type features/posix-locks > option mandatory-locks on # enables mandatory locking on all files > subvolumes posix8 > end-volume > ### Add network serving capability to above brick. > volume server > type protocol/server > option transport-type tcp > option transport.socket.bind-address 172.20.92.240 # Default is to listen on all interfaces > option transport.socket.listen-port 6996 # Default is 6996 > subvolumes brick1 brick2 brick3 brick4 > option auth.addr.brick1.allow * # Allow access to "brick" volume > option auth.addr.brick2.allow * # Allow access to "brick" volume > option auth.addr.brick3.allow * # Allow access to "brick" volume > option auth.addr.brick4.allow * # Allow access to "brick" volume > option auth.addr.brick5.allow * # Allow access to "brick" volume > option auth.addr.brick6.allow * # Allow access to "brick" volume > option auth.addr.brick7.allow * # Allow access to "brick" volume > option auth.addr.brick8.allow * # Allow access to "brick" volume > end-volume > > the client configuration: > > gfs1:/ # cat /etc/glusterfs/glusterfs-client.vol.dht > volume client1 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > option remote-subvolume brick1 # name of the remote volume > end-volume > volume client2 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick2 # name of the remote volume > end-volume > volume client3 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick3 # name of the remote volume > end-volume > volume client4 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick4 # name of the remote volume > end-volume > volume client5 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick1 # name of the remote volume > end-volume > volume client6 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick2 # name of the remote volume > end-volume > volume client7 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick3 # name of the remote volume > end-volume > volume client8 > type protocol/client > option transport-type tcp > option remote-host 172.20.92.240 # IP address of the remote brick2 > option remote-port 6996 > #option transport-timeout 10 # seconds to wait for a reply > option remote-subvolume brick4 # name of the remote volume > end-volume > #volume afr3 > # type cluster/afr > # subvolumes client3 client6 > #end-volume > volume dht > type cluster/dht > option lookup-unhashed yes > subvolumes client1 client2 client3 client4 > end-volume > > Could you help me ? > > > > 2009-07-06 > ________________________________ > eagleeyes > ________________________________ > 发件人: Sachidananda > 发送时间: 2009-07-04 11:39:03 > 收件人: eagleeyes > 抄送: gluster-users > 主题: Re: [Gluster-users] HELP : Files lost after DHT expansion > Hi, > eagleeyes wrote: > > When i update to gluster2.0.3 ,after dht expansion ,double directorys > > appear in the gluster directory ,why ? > > > > client configure > > volume dht > > type cluster/dht > > option lookup-unhashed yes > > option min-free-disk 10% > > subvolumes client1 client2 client3 client4 client5 client6 client7 > client8 > > #subvolumes client1 client2 client3 client4 > > end-volume > > > > > Can you please send us your server/client volume files? > -- > Sachidananda. > _______________________________________________ > Gluster-users mailing list > Gluster-users@gluster.org > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users > > _______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users