Re: [Gluster-users] Upgrade from 3.5 to 3.6: issue with socket

2015-03-08 Thread Joe Julian
I've seen this when the subprocesses running glustershd and nfs were not 
restarted after the upgrade. I've also seen it if nfs is disabled on all 
volumes.


On 03/08/2015 09:10 PM, Krishnan Parthasarathi wrote:

I just upgraded from 3.5.3 to 3.6.2 and have issues mounting my volume on a
client. On the server side I found this error message which might be the
cause of my issues:

Could you describe the issues you are facing?


[2015-03-08 13:22:36.383715] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/6b8f1f2526c6af8a87f1bb611ae5a86f.socket failed (Invalid
argument)

Anyony has any idea what that could be ?

This log message indicates that there was no (server) process listening on
the socket identified by the path, when glusterd tried to connect(3). This
happens if the connect(3) happened sooner than the server process starts 
listening.
glusterd will attempt a (re)connect within the next 3 seconds and should 
successfully
establish connection with the server. This wouldn't result in any sort of
'service disruption' in the I/O path.

Could you confirm that this log snippet is from glusterd's log file?
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Upgrade from 3.5 to 3.6: issue with socket

2015-03-08 Thread Krishnan Parthasarathi
 
 I just upgraded from 3.5.3 to 3.6.2 and have issues mounting my volume on a
 client. On the server side I found this error message which might be the
 cause of my issues:

Could you describe the issues you are facing?

 
 [2015-03-08 13:22:36.383715] W [socket.c:611:__socket_rwv] 0-management:
 readv on /var/run/6b8f1f2526c6af8a87f1bb611ae5a86f.socket failed (Invalid
 argument)
 
 Anyony has any idea what that could be ?

This log message indicates that there was no (server) process listening on
the socket identified by the path, when glusterd tried to connect(3). This
happens if the connect(3) happened sooner than the server process starts 
listening.
glusterd will attempt a (re)connect within the next 3 seconds and should 
successfully
establish connection with the server. This wouldn't result in any sort of
'service disruption' in the I/O path.

Could you confirm that this log snippet is from glusterd's log file?
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster errors create zombie processes [LOGS ATTACHED]

2015-03-08 Thread Przemysław Mroczek
I don't have volfiles, they are not on our machines as I said previously we
don't have impact on gluster servers.

I saw some graph that looks similiar to volume file on logs. I will paste
it here but we don't really have any impact on that. We are just using
client to connect to gluster servers, we are not in control of.

*1: volume drslk-prod-client-0*
*  2: type protocol/client*
*  3: option ping-timeout 20*
*  4: option remote-host brick13.gluster.iadm*
*  5: option remote-subvolume /GLUSTERFS/drslk-prod*
*  6: option transport-type socket*
*  7: option frame-timeout 60*
*  8: option send-gids true*
*  9: end-volume*
* 10:  *
* 11: volume drslk-prod-client-1*
* 12: type protocol/client*
* 13: option ping-timeout 20*
* 14: option remote-host brick14.gluster.iadm*
* 15: option remote-subvolume /GLUSTERFS/drslk-prod*
* 16: option transport-type socket*
* 17: option frame-timeout 60*
* 18: option send-gids true*
* 19: end-volume*
* 20:  *
* 21: volume drslk-prod-client-2*
* 22: type protocol/client*
* 23: option ping-timeout 20*
* 24: option remote-host brick15.gluster.iadm*
* 25: option remote-subvolume /GLUSTERFS/drslk-prod*
* 26: option transport-type socket*
* 27: option frame-timeout 60*
* 28: option send-gids true*
* 29: end-volume*
* 30:  *
* 31: volume drslk-prod-replicate-0*
* 32: type cluster/replicate*
* 33: option read-hash-mode 2*
* 34: option data-self-heal-window-size 128*
* 35: option quorum-type auto*
* 36: subvolumes drslk-prod-client-0 drslk-prod-client-1
drslk-prod-client-2*
* 37: end-volume*
* 38:  *
* 39: volume drslk-prod-client-3*
* 40: type protocol/client*
* 41: option ping-timeout 20*
* 42: option remote-host brick16.gluster.iadm*
* 43: option remote-subvolume /GLUSTERFS/drslk-prod*
* 44: option transport-type socket*
* 45: option frame-timeout 60*
* 46: option send-gids true*
* 47: end-volume*
* 48:  *
* 49: volume drslk-prod-client-4*
* 50: type protocol/client*
* 51: option ping-timeout 20*
* 52: option remote-host brick17.gluster.iadm*
* 53: option remote-subvolume /GLUSTERFS/drslk-prod*
* 54: option transport-type socket*
* 55: option frame-timeout 60*
* 56: option send-gids true*
* 57: end-volume*
* 58:  *
* 59: volume drslk-prod-client-5*
* 60: type protocol/client*
* 61: option ping-timeout 20*
* 62: option remote-host brick18.gluster.iadm*
* 63: option remote-subvolume /GLUSTERFS/drslk-prod*
* 64: option transport-type socket*
* 65: option frame-timeout 60*
* 66: option send-gids true*
* 67: end-volume*
* 68:  *
* 69: volume drslk-prod-replicate-1*
* 70: type cluster/replicate*
* 71: option read-hash-mode 2*
* 72: option data-self-heal-window-size 128*
* 73: option quorum-type auto*
* 74: subvolumes drslk-prod-client-3 drslk-prod-client-4
drslk-prod-client-5*
* 75: end-volume*
* 76:  *
* 77: volume drslk-prod-client-6*
* 78: type protocol/client*
* 79: option ping-timeout 20*
* 80: option remote-host brick19.gluster.iadm*
* 81: option remote-subvolume /GLUSTERFS/drslk-prod*
* 82: option transport-type socket*
* 83: option frame-timeout 60*
* 84: option send-gids true*
* 85: end-volume*
* 86:  *
* 87: volume drslk-prod-client-7*
* 88: type protocol/client*
* 89: option ping-timeout 20*
* 90: option remote-host brick20.gluster.iadm*
* 91: option remote-subvolume /GLUSTERFS/drslk-prod*
* 92: option transport-type socket*
* 93: option frame-timeout 60*
* 94: option send-gids true*
* 95: end-volume*
* 96:  *
* 97: volume drslk-prod-client-8*
* 98: type protocol/client*
* 99: option ping-timeout 20*
*100: option remote-host brick21.gluster.iadm*
*101: option remote-subvolume /GLUSTERFS/drslk-prod*
*102: option transport-type socket*
*103: option frame-timeout 60*
*104: option send-gids true*
*105: end-volume*
*106:  *
*107: volume drslk-prod-replicate-2*
*108: type cluster/replicate*
*109: option read-hash-mode 2*
*110: option data-self-heal-window-size 128*
*111: option quorum-type auto*
*112: subvolumes drslk-prod-client-6 drslk-prod-client-7
drslk-prod-client-8*
*113: end-volume*
*114:  *
*115: volume drslk-prod-client-9*
*116: type protocol/client*
*117: option ping-timeout 20*
*118: option remote-host brick22.gluster.iadm*
*119: option remote-subvolume /GLUSTERFS/drslk-prod*
*120: option transport-type socket*
*121: option frame-timeout 60*
*122: option send-gids true*
*123: end-volume*
*124:  *
*125: volume drslk-prod-client-10*
*126: type protocol/client*
*127: option ping-timeout 20*
*128: option remote-host brick23.gluster.iadm*
*129: option remote-subvolume /GLUSTERFS/drslk-prod*
*130: option transport-type socket*
*131: option frame-timeout 60*
*132: option send-gids true*
*133: end-volume*
*134:  *

[Gluster-users] Upgrade from 3.5 to 3.6: issue with socket

2015-03-08 Thread Uli Zumbuhl
Hello,

I just upgraded from 3.5.3 to 3.6.2 and have issues mounting my volume on a 
client. On the server side I found this error message which might be the cause 
of my issues: 

[2015-03-08 13:22:36.383715] W [socket.c:611:__socket_rwv] 0-management: readv 
on /var/run/6b8f1f2526c6af8a87f1bb611ae5a86f.socket failed (Invalid argument)

Anyony has any idea what that could be ?

Regards
Uli
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] no data available

2015-03-08 Thread Tamas Papp


On 03/06/2015 02:34 AM, Niels de Vos wrote:
This should be addressed with the next 3.6 update. See bug 1188064 for 
more details: https://bugzilla.redhat.com/1188064 Niels 



Thanks,
tamas
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster errors create zombie processes [LOGS ATTACHED]

2015-03-08 Thread Vijay Bellur

On 03/08/2015 09:36 AM, Przemysław Mroczek wrote:

I don't have volfiles, they are not on our machines as I said previously
we don't have impact on gluster servers.

I saw some graph that looks similiar to volume file on logs. I will
paste it here but we don't really have any impact on that. We are just
using client to connect to gluster servers, we are not in control of.



I would recommend to not alter the default for frame timeout.



Btw, do you think that different versions of gluster client and gluster
server could be an issue here?



It can potentially be. What versions are you using on the servers and 
the client?


-Vijay


2015-03-08 1:29 GMT+01:00 Vijay Bellur vbel...@redhat.com
mailto:vbel...@redhat.com:

On 03/07/2015 06:20 PM, Przemysław Mroczek wrote:

Hi guys,

We have rails app, which is using gluster for our distributed file
system. The glusters servers are hosted independently as part of
deal
with other, we don't have any impact on them, we are connected o
them by
using gluster native client.

We tried to resolve this issue using help from the admins of the
company
that is hosting our gluster servers, but they say that's the client
issue and we ran out of ideas how that's possible if we are not
doing
anything special here.

Information about independent gluster servers:
-version: 3.6.0.42.1
- They are using red hat
-They are enterprise so the are always using older versions

Our servers:
System version: Ubuntu 14.04
Our gluster client version: 3.6.2

The exact problem is that it often happens(couple times a week) that
errors in gluster causes proceses to become zombies. It happens
with our
application server(unicorn), nginx and our crawling script that
is run
as daemon.

Our fstab file:

10.10.11.17:/drslk-prod /mnt/storage  glusterfs
defaults,_netdev,nobootwait,__fetch-attempts=10 0 0
10.10.11.17:/drslk-backup /mnt/backup  glusterfs
defaults,_netdev,nobootwait,__fetch-attempts=10 0 0

Logs from gluster:

2015-02-18 12:36:12.375695] E
[rpc-clnt.c:362:saved_frames___unwind] (--

/usr/lib/x86_64-linux-gnu/__libglusterfs.so.0(_gf_log___callingfn+0x186)[__0x7fb41ddeada6]
(--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(saved_frames___unwind+0x1de)[0x7fb41d
bc1c7e] (--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(saved_frames___destroy+0xe)[0x7fb41dbc1d8e]
(--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(rpc_clnt___connection_cleanup+0x82)[__0x7fb41dbc3602]
(-- /usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(rpc
_clnt_notify+0x48)[__0x7fb41dbc3d98] )
0-drslk-prod-client-10: forced
unwinding frame type(GlusterFS 3.3) op(LOOKUP(27)) called at
2015-02-18
12:36:12.361489 (xid=0x5d475da)
[2015-02-18 12:36:12.375765] W
[client-rpc-fops.c:2766:__client3_3_lookup_cbk]
0-drslk-prod-client-10:
remote operation failed: Transport endpoint is not connected. Path:
/system/posts/00/00/71/77/59.__jpg
(2ad81c2b-a141-478d-9dd4-__253345edbce
b)
[2015-02-18 12:36:12.376288] E
[rpc-clnt.c:362:saved_frames___unwind] (--

/usr/lib/x86_64-linux-gnu/__libglusterfs.so.0(_gf_log___callingfn+0x186)[__0x7fb41ddeada6]
(--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(saved_frames___unwind+0x1de)[0x7fb41d
bc1c7e] (--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(saved_frames___destroy+0xe)[0x7fb41dbc1d8e]
(--

/usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(rpc_clnt___connection_cleanup+0x82)[__0x7fb41dbc3602]
(-- /usr/lib/x86_64-linux-gnu/__libgfrpc.so.0(rpc
_clnt_notify+0x48)[__0x7fb41dbc3d98] )
0-drslk-prod-client-10: forced
unwinding frame type(GlusterFS 3.3) op(LOOKUP(27)) called at
2015-02-18
12:36:12.361858 (xid=0x5d475db)
[2015-02-18 12:36:12.376355] W
[client-rpc-fops.c:2766:__client3_3_lookup_cbk]
0-drslk-prod-client-10:
remote operation failed: Transport endpoint is not connected. Path:
/system/posts/00/00/08 (f5c33a99-719e-4ea2-ad1f-__33b893af103d)
[2015-02-18 12:36:12.376711] I
[socket.c:3292:socket_submit___request]
0-drslk-prod-client-10: not connected (priv-connected = 0)
[2015-02-18 12:36:12.376749] W [rpc-clnt.c:1562:rpc_clnt___submit]
0-drslk-prod-client-10: failed to submit rpc-request (XID: 0x5d475dc
Program: GlusterFS 3.3, ProgVers: 330, Proc: 27) to rpc-transport
(drslk-prod-client-10)
[2015-02-18 12:36:12.376814] W
[client-rpc-fops.c:2766:__client3_3_lookup_cbk]
0-drslk-prod-client-10:
remote operation failed: