[Gluster-users] self-heal daemon not running
Hi, I have a 2 node replicated volume under ovirt 3.5. My self heal daemon is not running. I have a lot of misshealted vms on my glusterfs [root@node1 ~]# gluster volume heal g1sata info Brick node0.itsmart.cloud:/data/sata/brick/ gfid:bc0623a6-da58-4a86-8c81-f8ac67dbbe35 Number of entries: 1 Brick node1.itsmart.cloud:/data/sata/brick/ gfid:bc0623a6-da58-4a86-8c81-f8ac67dbbe35 /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/6788e53a-750d-4566-8579-37f586a0f306/2f62334e-39dc-4ffa-9102-51289588c42b - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/12ff021d-4075-4032-979c-685520dc1895/4051ffec-3dd2-495d-989b-eefb9fe92221 - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/c9dbc63e-b9a2-43aa-b433-8c53ce824492/bb0efb35-5164-4b22-9bed-5daeacf97129 - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/388c14f5-5690-4eae-a7dc-76d782ad8acc/0059a2c2-f8b1-4979-8321-41422d9a469f - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/2cb7ee4b-5c43-45e7-b13e-18aa3df0ef66/c0cd0554-ac37-4feb-803c-d1207219e3a1 - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/1bb441b8-84a2-4d5b-bd29-f57b100bbce4/095230c2-0411-44cf-a085-3c929e4ca9b6 - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/e3751092-3f6a-4aa6-b569-2a2fb4ae294a/133b2d17-2a2a-4ec3-b26a-4fd685aa2b78 - Possibly undergoing heal /fbfc7c67-ae12-4779-a5f0-42d32a3f6248/images/1535497b-d6ca-40e3-84b0-85f55217cbc9/144ddc5c-be25-4d5e-91a4-a0864ea2a10e - Possibly undergoing heal Number of entries: 9 Status of volume: g1sata Gluster process Port Online Pid -- Brick 172.16.0.10:/data/sata/brick 49152 Y 27983 Brick 172.16.0.11:/data/sata/brick 49152 Y 2581 NFS Server on localhost 2049 Y 14209 Self-heal Daemon on localhost N/A Y 14225 NFS Server on 172.16.0.10 2049 Y 27996 Self-heal Daemon on 172.16.0.10 N/A Y 28004 Task Status of Volume g1sata -- There are no active volume tasks [root@node1 ~]# rpm -qa|grep gluster glusterfs-libs-3.5.2-1.el6.x86_64 glusterfs-cli-3.5.2-1.el6.x86_64 glusterfs-rdma-3.5.2-1.el6.x86_64 glusterfs-server-3.5.2-1.el6.x86_64 glusterfs-3.5.2-1.el6.x86_64 glusterfs-api-3.5.2-1.el6.x86_64 glusterfs-fuse-3.5.2-1.el6.x86_64 vdsm-gluster-4.16.7-1.gitdb83943.el6.noarch centos 6.5 , firewall is disabled, selinux is on permissive I did a service restart on each node but that isn't helped. Also I have split-brained could someone help me? Thanks Tibor ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] NFS not start on localhost
Hi, This is the full nfs.log after delete reboot. It is refers to portmap registering problem. [root@node0 glusterfs]# cat nfs.log [2014-10-20 06:48:43.221136] I [glusterfsd.c:1959:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.5.2 (/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/567e0bba7ad7102eae3049e2ad6c3ed7.socket) [2014-10-20 06:48:43.22] I [socket.c:3561:socket_init] 0-socket.glusterfsd: SSL support is NOT enabled [2014-10-20 06:48:43.224475] I [socket.c:3576:socket_init] 0-socket.glusterfsd: using system polling thread [2014-10-20 06:48:43.224654] I [socket.c:3561:socket_init] 0-glusterfs: SSL support is NOT enabled [2014-10-20 06:48:43.224667] I [socket.c:3576:socket_init] 0-glusterfs: using system polling thread [2014-10-20 06:48:43.235876] I [rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 16 [2014-10-20 06:48:43.254087] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.254116] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.255241] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.255264] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.257279] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.257315] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.258135] I [socket.c:3561:socket_init] 0-socket.NLM: SSL support is NOT enabled [2014-10-20 06:48:43.258157] I [socket.c:3576:socket_init] 0-socket.NLM: using system polling thread [2014-10-20 06:48:43.293724] E [rpcsvc.c:1314:rpcsvc_program_register_portmap] 0-rpc-service: Could not register with portmap [2014-10-20 06:48:43.293760] E [nfs.c:332:nfs_init_versions] 0-nfs: Program NLM4 registration failed [2014-10-20 06:48:43.293771] E [nfs.c:1312:init] 0-nfs: Failed to initialize protocols [2014-10-20 06:48:43.293777] E [xlator.c:403:xlator_init] 0-nfs-server: Initialization of volume 'nfs-server' failed, review your volfile again [2014-10-20 06:48:43.293783] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed [2014-10-20 06:48:43.293789] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed pending frames: frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 11 time of crash: 2014-10-20 06:48:43configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.5.2 [root@node0 glusterfs]# systemctl status portma portma.service Loaded: not-found (Reason: No such file or directory) Active: inactive (dead) Also I have checked the rpcbind service. [root@node0 glusterfs]# systemctl status rpcbind.service rpcbind.service - RPC bind service Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled) Active: active (running) since h 2014-10-20 08:48:39 CEST; 2min 52s ago Process: 1940 ExecStart=/sbin/rpcbind -w ${RPCBIND_ARGS} (code=exited, status=0/SUCCESS) Main PID: 1946 (rpcbind) CGroup: /system.slice/rpcbind.service └─1946 /sbin/rpcbind -w okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Starting RPC bind service... okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Started RPC bind service. The restart does not solve this problem. I think this is the problem. Why are exited the portmap status? On node1 is ok: [root@node1 ~]# systemctl status rpcbind.service rpcbind.service - RPC bind service Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled) Active: active (running) since p 2014-10-17 19:15:21 CEST; 2 days ago Main PID: 1963 (rpcbind) CGroup: /system.slice/rpcbind.service └─1963 /sbin/rpcbind -w okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Starting RPC bind service... okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Started RPC bind service. Thanks in advance Tibor - Eredeti üzenet - On 10/19/2014 06:56 PM, Niels de Vos wrote: On Sat, Oct 18, 2014 at 01:24:12PM +0200, Demeter Tibor wrote: Hi, [root@node0 ~]# tail -n 20 /var/log/glusterfs/nfs.log [2014-10-18 07:41:06.136035] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed [2014-10-18 07:41:06.136040] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed pending frames: frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 11 time of crash: 2014-10-18 07:41:06configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1
Re: [Gluster-users] NFS not start on localhost
Also it's funny, because meanwhile the portmap are listening on localhost. [root@node0 log]# netstat -tunlp | grep 111 tcp0 0 0.0.0.0:111 0.0.0.0:* LISTEN 4709/rpcbind tcp6 0 0 :::111 :::*LISTEN 4709/rpcbind udp0 0 0.0.0.0:111 0.0.0.0:* 4709/rpcbind udp6 0 0 :::111 :::* 4709/rpcbind Demeter Tibor - Eredeti üzenet - Hi, This is the full nfs.log after delete reboot. It is refers to portmap registering problem. [root@node0 glusterfs]# cat nfs.log [2014-10-20 06:48:43.221136] I [glusterfsd.c:1959:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.5.2 (/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/567e0bba7ad7102eae3049e2ad6c3ed7.socket) [2014-10-20 06:48:43.22] I [socket.c:3561:socket_init] 0-socket.glusterfsd: SSL support is NOT enabled [2014-10-20 06:48:43.224475] I [socket.c:3576:socket_init] 0-socket.glusterfsd: using system polling thread [2014-10-20 06:48:43.224654] I [socket.c:3561:socket_init] 0-glusterfs: SSL support is NOT enabled [2014-10-20 06:48:43.224667] I [socket.c:3576:socket_init] 0-glusterfs: using system polling thread [2014-10-20 06:48:43.235876] I [rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 16 [2014-10-20 06:48:43.254087] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.254116] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.255241] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.255264] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.257279] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.257315] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.258135] I [socket.c:3561:socket_init] 0-socket.NLM: SSL support is NOT enabled [2014-10-20 06:48:43.258157] I [socket.c:3576:socket_init] 0-socket.NLM: using system polling thread [2014-10-20 06:48:43.293724] E [rpcsvc.c:1314:rpcsvc_program_register_portmap] 0-rpc-service: Could not register with portmap [2014-10-20 06:48:43.293760] E [nfs.c:332:nfs_init_versions] 0-nfs: Program NLM4 registration failed [2014-10-20 06:48:43.293771] E [nfs.c:1312:init] 0-nfs: Failed to initialize protocols [2014-10-20 06:48:43.293777] E [xlator.c:403:xlator_init] 0-nfs-server: Initialization of volume 'nfs-server' failed, review your volfile again [2014-10-20 06:48:43.293783] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed [2014-10-20 06:48:43.293789] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed pending frames: frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 11 time of crash: 2014-10-20 06:48:43configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.5.2 [root@node0 glusterfs]# systemctl status portma portma.service Loaded: not-found (Reason: No such file or directory) Active: inactive (dead) Also I have checked the rpcbind service. [root@node0 glusterfs]# systemctl status rpcbind.service rpcbind.service - RPC bind service Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled) Active: active (running) since h 2014-10-20 08:48:39 CEST; 2min 52s ago Process: 1940 ExecStart=/sbin/rpcbind -w ${RPCBIND_ARGS} (code=exited, status=0/SUCCESS) Main PID: 1946 (rpcbind) CGroup: /system.slice/rpcbind.service └─1946 /sbin/rpcbind -w okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Starting RPC bind service... okt 20 08:48:39 node0.itsmart.cloud systemd[1]: Started RPC bind service. The restart does not solve this problem. I think this is the problem. Why are exited the portmap status? On node1 is ok: [root@node1 ~]# systemctl status rpcbind.service rpcbind.service - RPC bind service Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; enabled) Active: active (running) since p 2014-10-17 19:15:21 CEST; 2 days ago Main PID: 1963 (rpcbind) CGroup: /system.slice/rpcbind.service └─1963 /sbin/rpcbind -w okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Starting RPC bind service... okt 17 19:15:21 node1.itsmart.cloud systemd[1]: Started RPC bind service. Thanks in advance Tibor - Eredeti üzenet - On 10/19/2014 06:56 PM, Niels de Vos wrote
Re: [Gluster-users] NFS not start on localhost
Hi, Thank you for you reply. I did your recommendations, but there are no changes. In the nfs.log there are no new things. [root@node0 glusterfs]# reboot Connection to 172.16.0.10 closed by remote host. Connection to 172.16.0.10 closed. [tdemeter@sirius-31 ~]$ ssh root@172.16.0.10 root@172.16.0.10's password: Last login: Mon Oct 20 11:02:13 2014 from 192.168.133.106 [root@node0 ~]# systemctl status nfs.target nfs.target - Network File System Server Loaded: loaded (/usr/lib/systemd/system/nfs.target; disabled) Active: inactive (dead) [root@node0 ~]# gluster volume status engine Status of volume: engine Gluster process PortOnline Pid -- Brick gs00.itsmart.cloud:/gluster/engine0 50160 Y 3271 Brick gs01.itsmart.cloud:/gluster/engine1 50160 Y 595 NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A Y 3286 NFS Server on gs01.itsmart.cloud2049Y 6951 Self-heal Daemon on gs01.itsmart.cloud N/A Y 6958 Task Status of Volume engine -- There are no active volume tasks [root@node0 ~]# systemctl status Display all 262 possibilities? (y or n) [root@node0 ~]# systemctl status nfs-lock nfs-lock.service - NFS file locking service. Loaded: loaded (/usr/lib/systemd/system/nfs-lock.service; enabled) Active: inactive (dead) [root@node0 ~]# systemctl stop nfs-lock [root@node0 ~]# systemctl restart gluster glusterd.serviceglusterfsd.service gluster.mount [root@node0 ~]# systemctl restart gluster glusterd.serviceglusterfsd.service gluster.mount [root@node0 ~]# systemctl restart glusterfsd.service [root@node0 ~]# systemctl restart glusterd.service [root@node0 ~]# gluster volume status engine Status of volume: engine Gluster process PortOnline Pid -- Brick gs00.itsmart.cloud:/gluster/engine0 50160 Y 5140 Brick gs01.itsmart.cloud:/gluster/engine1 50160 Y 2037 NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A N N/A NFS Server on gs01.itsmart.cloud2049Y 6951 Self-heal Daemon on gs01.itsmart.cloud N/A Y 6958 Any other idea? Tibor - Eredeti üzenet - On Mon, Oct 20, 2014 at 09:04:2.8AM +0200, Demeter Tibor wrote: Hi, This is the full nfs.log after delete reboot. It is refers to portmap registering problem. [root@node0 glusterfs]# cat nfs.log [2014-10-20 06:48:43.221136] I [glusterfsd.c:1959:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.5.2 (/usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/567e0bba7ad7102eae3049e2ad6c3ed7.socket) [2014-10-20 06:48:43.22] I [socket.c:3561:socket_init] 0-socket.glusterfsd: SSL support is NOT enabled [2014-10-20 06:48:43.224475] I [socket.c:3576:socket_init] 0-socket.glusterfsd: using system polling thread [2014-10-20 06:48:43.224654] I [socket.c:3561:socket_init] 0-glusterfs: SSL support is NOT enabled [2014-10-20 06:48:43.224667] I [socket.c:3576:socket_init] 0-glusterfs: using system polling thread [2014-10-20 06:48:43.235876] I [rpcsvc.c:2127:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 16 [2014-10-20 06:48:43.254087] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.254116] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.255241] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.255264] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.257279] I [socket.c:3561:socket_init] 0-socket.nfs-server: SSL support is NOT enabled [2014-10-20 06:48:43.257315] I [socket.c:3576:socket_init] 0-socket.nfs-server: using system polling thread [2014-10-20 06:48:43.258135] I [socket.c:3561:socket_init] 0-socket.NLM: SSL support is NOT enabled [2014-10-20 06:48:43.258157] I [socket.c:3576:socket_init] 0-socket.NLM: using system polling thread [2014-10-20 06:48:43.293724] E [rpcsvc.c:1314:rpcsvc_program_register_portmap] 0-rpc-service: Could not register with portmap [2014-10-20 06:48:43.293760] E [nfs.c:332:nfs_init_versions] 0-nfs: Program NLM4 registration failed The above line suggests
Re: [Gluster-users] NFS not start on localhost
Hi, I've try out these things: - nfs.disable on-of - iptables disable - volume stop-start but same. So, when I make a new volume everything is fine. After reboot the NFS won't listen on local host (only on server has brick0) Centos7 with last ovirt Regards, Tibor - Eredeti üzenet - It happens with me sometimes. Try `tail -n 20 /var/log/glusterfs/nfs.log`. You will probably find something out that will help your cause. In general, if you just wish to start the thing up without going into the why of it, try `gluster volume set engine nfs.disable on` followed by ` gluster volume set engine nfs.disable off`. It does the trick quite often for me because it is a polite way to askmgmt/glusterd to try and respawn the nfs server process if need be. But, keep in mind that this will call a (albeit small) service interruption to all clients accessing volume engine over nfs. Thanks, Anirban On Saturday, 18 October 2014 1:03 AM, Demeter Tibor tdeme...@itsmart.hu wrote: Hi, I have make a glusterfs with nfs support. I don't know why, but after a reboot the nfs does not listen on localhost, only on gs01. [root@node0 ~]# gluster volume info engine Volume Name: engine Type: Replicate Volume ID: 2ea009bf-c740-492e-956d-e1bca76a0bd3 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gs00.itsmart.cloud:/gluster/engine0 Brick2: gs01.itsmart.cloud:/gluster/engine1 Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server auth.allow: * nfs.disable: off [root@node0 ~]# gluster volume status engine Status of volume: engine Gluster process Port Online Pid -- Brick gs00.itsmart.cloud:/gluster/engine0 50158 Y 3250 Brick gs01.itsmart.cloud:/gluster/engine1 50158 Y 5518 NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A Y 3261 NFS Server on gs01.itsmart.cloud 2049 Y 5216 Self-heal Daemon on gs01.itsmart.cloud N/A Y 5223 Does anybody help me? Thanks in advance. Tibor ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] NFS not start on localhost
Hi, [root@node0 ~]# tail -n 20 /var/log/glusterfs/nfs.log [2014-10-18 07:41:06.136035] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed [2014-10-18 07:41:06.136040] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed pending frames: frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 11 time of crash: 2014-10-18 07:41:06configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.5.2 Udv: Demeter Tibor Email: tdemeter @itsmart.hu Skype: candyman_78 Phone: +36 30 462 0500 Web : www.it smart.hu IT SMART KFT. 2120 Dunakeszi Wass Albert utca 2. I. em 9. Telefon: +36 30 462-0500 Fax: +36 27 637-486 [EN] This message and any attachments are confidential and privileged and intended for the use of the addressee only. If you have received this communication in error, please notify the sender by replay e-mail and delete this message from your system. Please note that Internet e-mail guarantees neither the confidentiality nor the proper receipt of the message sent. The data deriving from our correspondence with you are included in a file of ITSMART Ltd which exclusive purpose is to manage the communications of the company; under the understanding that, in maintaining said correspondence, you authorize the treatment of such data for the mentioned purpose. You are entitled to exercise your rights of access, rectification, cancellation and opposition by addressing such written application to address above. [HUN] Ez az üzenet és annak bármely csatolt anyaga bizalmas, a nyilvános közléstol védett, kizárólag a címzett használhatja fel. Ha Ön nem az üzenet címzettje, úgy kérjük válaszüzenetben értesítse errol a feladót és törölje az üzenetet a rendszerbol. Kérjük vegye figyelembe, hogy az email-en történo információtovábbítás kockázattal járhat, nem garantálja sem a csatorna bizalmasságát, sem a kézbesítést. A levél az ITSMART Informatikai Kft. kommunikációjának eszköze, az adatokat kizárólag erre a célra használjuk. Jogosult tájékoztatást kérni személyes adatai kezelésérol, kérheti azok helyesbítését, illetve törlését írásos kérelemben a fenti e-mail címen. - Eredeti üzenet - Maybe share the last 15-20 lines of you /var/log/glusterfs/nfs.log for the consideration of everyone on the list? Thanks. From: Demeter Tibor tdeme...@itsmart.hu; To: Anirban Ghoshal chalcogen_eg_oxy...@yahoo.com; Cc: gluster-users gluster-users@gluster.org; Subject: Re: [Gluster-users] NFS not start on localhost Sent: Sat, Oct 18, 2014 10:36:36 AM Hi, I've try out these things: - nfs.disable on-of - iptables disable - volume stop-start but same. So, when I make a new volume everything is fine. After reboot the NFS won't listen on local host (only on server has brick0) Centos7 with last ovirt Regards, Tibor - Eredeti üzenet - It happens with me sometimes. Try `tail -n 20 /var/log/glusterfs/nfs.log`. You will probably find something out that will help your cause. In general, if you just wish to start the thing up without going into the why of it, try `gluster volume set engine nfs.disable on` followed by ` gluster volume set engine nfs.disable off`. It does the trick quite often for me because it is a polite way to askmgmt/glusterd to try and respawn the nfs server process if need be. But, keep in mind that this will call a (albeit small) service interruption to all clients accessing volume engine over nfs. Thanks, Anirban On Saturday, 18 October 2014 1:03 AM, Demeter Tibor tdeme...@itsmart.hu wrote: Hi, I have make a glusterfs with nfs support. I don't know why, but after a reboot the nfs does not listen on localhost, only on gs01. [root@node0 ~]# gluster volume info engine Volume Name: engine Type: Replicate Volume ID: 2ea009bf-c740-492e-956d-e1bca76a0bd3 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gs00.itsmart.cloud:/gluster/engine0 Brick2: gs01.itsmart.cloud:/gluster/engine1 Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server auth.allow: * nfs.disable: off [root@node0 ~]# gluster volume status engine Status of volume: engine Gluster process Port Online Pid -- Brick gs00.itsmart.cloud:/gluster/engine0 50158 Y 3250 Brick gs01.itsmart.cloud:/gluster/engine1 50158 Y 5518 NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N
Re: [Gluster-users] NFS not start on localhost
I'm sorry, but I dont know what do you nfs translator mean. I've followed up the ovirt hosted-engine setup howto and I installed glusterfs, etc from scratch. So it is a centos7 minimal install. The nfs-utils package are installed, but it disabled, so it does not run as service. So it is a simple gluster volume and ovirt use this as nfs store for hosted-engine-setup. When I did the whole setup everything was fine. After reboot there are no nfs on localhost (or on local ip), only on the node1. But my hosted engine could not run only from this host. Maybe is it an ovirt bug? Thanks Tibor - Eredeti üzenet - Hmmm, do you have any custom translators installed, or have you been trying out GlusterFlow? I used to get crashes of the NFS translator (looks like this) when I was getting GlusterFlow up and running, when everything wasn't quite setup correctly. If you don't have any custom translators installed (or trying out GlusterFlow), ignore this. ;) Regards and best wishes, Justin Clift - Original Message - Hi, [root@node0 ~]# tail -n 20 /var/log/glusterfs/nfs.log [2014-10-18 07:41:06.136035] E [graph.c:307:glusterfs_graph_init] 0-nfs-server: initializing translator failed [2014-10-18 07:41:06.136040] E [graph.c:502:glusterfs_graph_activate] 0-graph: init failed pending frames: frame : type(0) op(0) patchset: git://git.gluster.com/glusterfs.git signal received: 11 time of crash: 2014-10-18 07:41:06configuration details: argp 1 backtrace 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 3.5.2 Udv: Demeter Tibor Email: tdemeter @itsmart.hu Skype: candyman_78 Phone: +36 30 462 0500 Web : www.it smart.hu IT SMART KFT. 2120 Dunakeszi Wass Albert utca 2. I. em 9. Telefon: +36 30 462-0500 Fax: +36 27 637-486 [EN] This message and any attachments are confidential and privileged and intended for the use of the addressee only. If you have received this communication in error, please notify the sender by replay e-mail and delete this message from your system. Please note that Internet e-mail guarantees neither the confidentiality nor the proper receipt of the message sent. The data deriving from our correspondence with you are included in a file of ITSMART Ltd which exclusive purpose is to manage the communications of the company; under the understanding that, in maintaining said correspondence, you authorize the treatment of such data for the mentioned purpose. You are entitled to exercise your rights of access, rectification, cancellation and opposition by addressing such written application to address above. [HUN] Ez az üzenet és annak bármely csatolt anyaga bizalmas, a nyilvános közléstol védett, kizárólag a címzett használhatja fel. Ha Ön nem az üzenet címzettje, úgy kérjük válaszüzenetben értesítse errol a feladót és törölje az üzenetet a rendszerbol. Kérjük vegye figyelembe, hogy az email-en történo információtovábbítás kockázattal járhat, nem garantálja sem a csatorna bizalmasságát, sem a kézbesítést. A levél az ITSMART Informatikai Kft. kommunikációjának eszköze, az adatokat kizárólag erre a célra használjuk. Jogosult tájékoztatást kérni személyes adatai kezelésérol, kérheti azok helyesbítését, illetve törlését írásos kérelemben a fenti e-mail címen. - Eredeti üzenet - ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users -- GlusterFS - http://www.gluster.org An open source, distributed file system scaling to several petabytes, and handling thousands of clients. My personal twitter: twitter.com/realjustinclift ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] NFS not start on localhost
Hi, I have make a glusterfs with nfs support. I don't know why, but after a reboot the nfs does not listen on localhost, only on gs01. [root@node0 ~]# gluster volume info engine Volume Name: engine Type: Replicate Volume ID: 2ea009bf-c740-492e-956d-e1bca76a0bd3 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gs00.itsmart.cloud:/gluster/engine0 Brick2: gs01.itsmart.cloud:/gluster/engine1 Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server auth.allow: * nfs.disable: off [root@node0 ~]# gluster volume status engine Status of volume: engine Gluster process Port Online Pid -- Brick gs00.itsmart.cloud:/gluster/engine0 50158 Y 3250 Brick gs01.itsmart.cloud:/gluster/engine1 50158 Y 5518 NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A Y 3261 NFS Server on gs01.itsmart.cloud 2049 Y 5216 Self-heal Daemon on gs01.itsmart.cloud N/A Y 5223 Does anybody help me? Thanks in advance. Tibor ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] [ovirt-users] Raid-5 like gluster method?
Hi, Do you recommend this featrue for productive environment? Tibor - Eredeti üzenet - use 3.6 disperse feature, but it is beta2 now, you could use it when it is GA On Wed, Sep 24, 2014 at 2:55 PM, Sahina Bose sab...@redhat.com wrote: [+gluster-users] On 09/24/2014 11:59 AM, Demeter Tibor wrote: Hi, Is there any method in glusterfs, like raid-5? I have three node, each node has 5 TB of disk. I would like utilize all of space with redundancy, like raid-5. If it not possible, can I make raid-6 like redundanci within three node? (two brick/node?). Thanks in advance, Tibor ___ Users mailing list us...@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
[Gluster-users] bonding question
Hi, I made short tests with glusterfs and bonding, but I have performance issues. Environment: - bonding mode=4 (with switch support) or mode=6 - centos7 - vlans - two servers with 4 nic/node, one nic on the internet (this is the default route) and 3 nic as bonded interface - MTU 9000 on all interface (bondings, vlans, eths, etc), MTU 9216 on the switch ports - each host vlan-s can ping each host on the vlan subnets and on the non vlan subnets. - the volume uses the bonded vlans as bricks [root@node1 lock]# gluster vol info Volume Name: meta Type: Replicate Volume ID: f4d026e7-3edd-442f-9207-f0a849acebf5 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gs00.itsmart.cloud:/gluster/meta0 Brick2: gs01.itsmart.cloud:/gluster/meta1 I did this test: [root@node0 lock]# dd if=/dev/zero of=/mnt/lock/disk bs=1M count=1000 conv=fdatasync 1000+0 records in 1000+0 records out 1048576000 bytes (1,0 GB) copied, 10,3035 s, 102 MB/s I compared with local hdd speed tests: [root@node0 lock]# dd if=/dev/zero of=/home/disk bs=1M count=1000 conv=fdatasync 1000+0 records in 1000+0 records out 1048576000 bytes (1,0 GB) copied, 3,04411 s, 344 MB/s Ok, I mean this is a network based solution, but I think the 100MB/sec is possible with one nic too. I just wondering, maybe my bonding isn't working fine. What do you think, is it ok? The port utilization is minimal, there are two bigger traffic on two ports only. Thanks in advance. Tibor ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] bonding question
Hi, I would like to use glusterfs as ovirt-vmstore. I this case one vm, that is running on one compute node will use only one tcp connection? Thanks - Eredeti üzenet - Ok, I mean this is a network based solution, but I think the 100MB/sec is possible with one nic too. I just wondering, maybe my bonding isn't working fine. You should test with multiple clients/dd streams. http://serverfault.com/questions/569060/link-aggregation-lacp-802-3ad-max-throughput/ rr ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] bonding question
In glusterfs documentation the recommended mode is the mode=6. My switch (dlink dgs-1510) can be 802.3ad modes, in this case is this better than mode=6 ? Tibor - Eredeti üzenet - Indeed. Only the rr (round robin) mode will get higher performance on a single stream. It also means that packets may be received out-of-order which can cause retransmissions (so it should never be used for UDP services like SIP/RTP). AFAIK it only works with Cisco etherchannel and does not scale well. Multiple streams are balanced using the XOR of the two endpoint MAC addresses in mode 4. This can be changed to include L3 data (eg src/dest IP) but switch support is again limited for the alternate algo. I know my kit can't be changed to add L3 data. As long as you have multiple clients the default mode 4 will scale almost linearly and will be guaranteed to work across any switch that supports LACP. Cheers Alex On 29/09/14 15:03, Reinis Rozitis wrote: Ok, I mean this is a network based solution, but I think the 100MB/sec is possible with one nic too. I just wondering, maybe my bonding isn't working fine. You should test with multiple clients/dd streams. http://serverfault.com/questions/569060/link-aggregation-lacp-802-3ad-max-throughput/ rr ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] bonding question
Hi Alex, Thank you for you replys. It will be in productive environment, so I need a reliable solution. The maximum troghtput is important, but the stability is the first in this project. I don't know yet, the rr, mode=6 or the mode=4 (with switch support) is the perfect for us. (I have an dlink DGS-1510 switch, I think it could be any 802.3ad mode) Thanks a lot! Regards, Tibor - Eredeti üzenet - Yes, but even with rr it's still one tcp connection. At layer2 it gets distributed over multiple physical links. TCP doesn't care or notice (except for retransmissions as I mentioned before). This is one advantage of iSCSI/FCoE/FC/SCSI etc in that you can use multipath which is transparent, scales per-link close to linear and is part of the storage protocol (ie multiple abstract paths between initiators and targets) rather than the network stack. You could serve up iSCSI from files on a mounted via FUSE from a glusterfs cluster, which would enable multipath, but I've only ever seen a demo of this on YouTube and I was not convinced that on its own it would be crash-consistent or resistant to gluster split-brain. Anyone else that's tried this is welcome to put me right on this. Cheers Alex On 29/09/14 15:10, Demeter Tibor wrote: Hi, I would like to use glusterfs as ovirt-vmstore. I this case one vm, that is running on one compute node will use only one tcp connection? Thanks - Eredeti üzenet - Ok, I mean this is a network based solution, but I think the 100MB/sec is possible with one nic too. I just wondering, maybe my bonding isn't working fine. You should test with multiple clients/dd streams. http://serverfault.com/questions/569060/link-aggregation-lacp-802-3ad-max-throughput/ rr ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users
Re: [Gluster-users] [ovirt-users] Raid-5 like gluster method?
Hi, Could I help anybody? Tibor [+gluster-users] On 09/24/2014 11:59 AM, Demeter Tibor wrote: Hi, Is there any method in glusterfs, like raid-5? I have three node, each node has 5 TB of disk. I would like utilize all of space with redundancy, like raid-5. If it not possible, can I make raid-6 like redundanci within three node? (two brick/node?). Thanks in advance, Tibor ___ Users mailing list us...@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users