[Gluster-users] Heal failed entries glsuter 3.6.1

2015-06-03 Thread Félix de Lelelis
Hi,

I have a question about heal process. Sometimes appear heal entries failed
after execute gluster volume heal Volname statistics.  When can happen
these cases?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Upgrade gluster 3.6.1 to 3.7.1

2015-06-02 Thread Félix de Lelelis
Hi,

We want upgrade  gluster 3.6.1 to 3.7.1. I have a question about quota,
must take into account the change in configuration? In this case, what
steps should follow?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Monitorig gluster 3.6.1

2015-05-31 Thread Félix de Lelelis
Hi,

I have monitoring gluster with scripts that lunch scripts. All scripts are
redirected to a one script that check if is active any process glusterd and
if the repsonse its false, the script lunch the check.

All checks are:

   - gluster volume volname info
   - gluster volume heal volname info
   - gluster volume heal volname split-brain
   - gluster volume volname status detail
   - gluster volume volname statistics

Since I enable the monitoring in our pre-production gluster, the gluster is
down 2 times. We  suspect that the monitoring are overloading but should
not.

The question is, there any way to check those states otherwise?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Gluster 3.6.1

2015-05-29 Thread Félix de Lelelis
Hi,

I have a cluster with 3 nodes on pre-production. Yesterday, one node was
down. The errror that I have seen is that:


[2015-05-28 19:04:27.305560] E [glusterd-syncop.c:1578:gd_sync_task_begin]
0-management: Unable to acquire lock for cfe-gv1
The message "I [MSGID: 106006]
[glusterd-handler.c:4257:__glusterd_nodesvc_rpc_notify] 0-management: nfs
has disconnected from glusterd." repeated 5 times between [2015-05-28
19:04:09.346088] and [2015-05-28 19:04:24.349191]
pending frames:
frame : type(0) op(0)
patchset: git://git.gluster.com/glusterfs.git
signal received: 11
time of crash:
2015-05-28 19:04:27
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.6.1
/usr/lib64/libglusterfs.so.0(_gf_msg_backtrace_nomem+0xb2)[0x7fd86e2f1232]
/usr/lib64/libglusterfs.so.0(gf_print_trace+0x32d)[0x7fd86e30871d]
/usr/lib64/libc.so.6(+0x35640)[0x7fd86d30c640]
/usr/lib64/glusterfs/3.6.1/xlator/mgmt/glusterd.so(glusterd_remove_pending_entry+0x2c)[0x7fd85f52450c]
/usr/lib64/glusterfs/3.6.1/xlator/mgmt/glusterd.so(+0x5ae28)[0x7fd85f511e28]
/usr/lib64/glusterfs/3.6.1/xlator/mgmt/glusterd.so(glusterd_op_sm+0x237)[0x7fd85f50f027]
/usr/lib64/glusterfs/3.6.1/xlator/mgmt/glusterd.so(__glusterd_brick_op_cbk+0x2fe)[0x7fd85f53be5e]
/usr/lib64/glusterfs/3.6.1/xlator/mgmt/glusterd.so(glusterd_big_locked_cbk+0x4c)[0x7fd85f53d48c]
/usr/lib64/libgfrpc.so.0(rpc_clnt_handle_reply+0x90)[0x7fd86e0c50b0]
/usr/lib64/libgfrpc.so.0(rpc_clnt_notify+0x171)[0x7fd86e0c5321]
/usr/lib64/libgfrpc.so.0(rpc_transport_notify+0x23)[0x7fd86e0c1273]
/usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0x8530)[0x7fd85d17d530]
/usr/lib64/glusterfs/3.6.1/rpc-transport/socket.so(+0xace4)[0x7fd85d17fce4]
/usr/lib64/libglusterfs.so.0(+0x76322)[0x7fd86e346322]
/usr/sbin/glusterd(main+0x502)[0x7fd86e79afb2]
/usr/lib64/libc.so.6(__libc_start_main+0xf5)[0x7fd86d2f8af5]
/usr/sbin/glusterd(+0x6351)[0x7fd86e79b351]
-


That is a problem with software? is a bug ?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Upgrade Glsuter 3.7 : Issue with quota

2015-05-21 Thread Félix de Lelelis
Hi,

I have upgraded gluster from release 6.3, with quota enable, to release
3.7, and when I have restarted the nodes, the log lunch a error:

 E [MSGID: 106012] [glusterd-utils.c:2670:glusterd_compare_friend_volume]
0-management: Cksums of quota configuration of volume cfe-gv1 differ. local
cksum = 2736284911, remote  cksum = 4092896828 on peer srv-vln-des3

Where is the problem?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Metadata filesystem XFS gluster 3.6

2015-03-31 Thread Félix de Lelelis
Hi,

I had a problem with a filesystem xfs on gluster. The filesystem metadata
was filled:

Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: space map metadata:
unable to allocate new metadata block
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2: metadata
operation 'dm_thin_insert_block' failed: error = -28
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2: aborting
current metadata transaction
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin: 252:2: switching
pool to read-only mode
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error: block
0x701830 ("xfs_buf_iodone_callbacks") error 5 numblks 8
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error: block
0x68047c ("xlog_iodone") error 5 numblks 64
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): xfs_do_force_shutdown(0x2)
called from line 1170 of file fs/xfs/xfs_log.c.  Return address =
0xa012a4c1
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): Log I/O Error Detected.
Shutting down filesystem
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): Please umount the
filesystem and rectify the problem(s)
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error: block
0x6804bc ("xlog_iodone") error 5 numblks 64
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): xfs_do_force_shutdown(0x2)
called from line 1170 of file fs/xfs/xfs_log.c.  Return address =
0xa012a4c1
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): xfs_log_force: error 5
returned.
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: attempt to access beyond end of device
Mar 27 13:58:18 srv-vln-des2 kernel: dm-0: rw=0, want=562056, limit=24576
Mar 27 13:58:18 srv-vln-des2 kernel: device-mapper: thin:
process_bio_read_only: dm_thin_find_block() failed: error = -5
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): metadata I/O error: block
0x6804fc ("xlog_iodone") error 5 numblks 64
Mar 27 13:58:18 srv-vln-des2 kernel: XFS (dm-4): xfs_do_force_shutdown(0x2)
called from line 1170 of file fs/xfs/xfs_log.c.  Return address =
0xa012a4c



After that, gluster was shutdown and with it the 2 server are shtudown too.
The lvm partition was missing and so far I haven't been able restore the
file system. All data is missing??
I don't understand the situation and I don't know if it's due a xfs
filesystem or glusterfs fail. Someone it has been this situation?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] "Out of memory" Gluster 3.6.2

2015-03-24 Thread Félix de Lelelis
Hi,

Today, Glusterd daemon has been killed  due to excessive memory consumption:

 [3505254.762715] Out of memory: Kill process 7780 (glusterd) score 581 or
sacrifice child
[3505254.763451] Killed process 7780 (glusterd) total-vm:3537640kB,
anon-rss:1205240kB, file-rss:672kB

I have installed gluster 3.6.2 on centos 7. There in any way to avoid this
whithout the need to kill the process?Simply need more memory?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Issue glusterfs 3.6.2

2015-03-18 Thread Félix de Lelelis
Hi,

Sorry, I have not explained well. I don't mean that I have lunch this
comand simultaneously, I am using zabbix and I have defined diferent time
intervals for this but comes a point at wich I can't luch anythig command
over the cluster, such as "gluster volume status". Is there any way around
this?

Thanks



2015-03-18 11:36 GMT+01:00 Atin Mukherjee :

>
>
> On 03/18/2015 03:24 PM, Félix de Lelelis wrote:
> > Hi,
> >
> > I have a problem with glusterfs 3.6. I am monitoring it with scripts that
> > lunch "gluster volume status VONAME detail" and "gluster volume profile
> > VOLNAME info". When this scripts are running about 1-2 hours, with check
> > every 1 minute, gluster is blocked and  the node generates "Another
> > transaction is in progress". The only way to restore the situation is
> > restart glusterd.
> >
> > Can you help me please.
> This is expected. Concurrent transactions on a same volumes are not
> allowed. There is a cluster wide volume lock maintained by glusterd. In
> this case since both these commands are for VOLNAME, any one of two
> would throw an error saying "Another transaction is in progress". If you
> ensure that there is a time gap between these two commands in your
> script, you can get rid of this situation.
>
> ~Atin
> >
> >
> > Thanks
> >
> >
> >
> > ___
> > Gluster-users mailing list
> > Gluster-users@gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-users
> >
>
> --
> ~Atin
>
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Issue glusterfs 3.6.2

2015-03-18 Thread Félix de Lelelis
Hi,

I have a problem with glusterfs 3.6. I am monitoring it with scripts that
lunch "gluster volume status VONAME detail" and "gluster volume profile
VOLNAME info". When this scripts are running about 1-2 hours, with check
every 1 minute, gluster is blocked and  the node generates "Another
transaction is in progress". The only way to restore the situation is
restart glusterd.

Can you help me please.


Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Gluster 3.6 issue

2015-03-16 Thread Félix de Lelelis
Hi,

I have a cluster with 2 nodes, and sometimes when I lunch gluster volume
status, appear a error on the log:

[2015-03-16 17:24:25.215352] E
[glusterd-utils.c:7364:glusterd_add_inode_size_to_dict] 0-management:
xfs_info exited with non-zero exit status
[2015-03-16 17:24:25.215379] E
[glusterd-utils.c:7390:glusterd_add_inode_size_to_dict] 0-management:
failed to get inode size

That you can be due?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] FOPS gluster 3.6

2015-03-12 Thread Félix de Lelelis
Hi,

Which is the difference between these types of fops: read, readdir and
readdirp. I am monitoring and I am interesting on read and write operations
but when lunch cat comand over the client only is modified readdirp on the
profile.

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Monitoring gluster 3.6

2015-03-04 Thread Félix de Lelelis
Hi,

Someone know  how obtain  stats over fops or I/O operations in gluster? The
idea is integrate this scripts with zabbix.

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Geo-replication session

2015-02-23 Thread Félix de Lelelis
Hi,

I am testing geo-replication in gluster-3.6. I have created 2 sessions but
when I deleted the last one glusterd yet has register it:

 [2015-02-23 08:51:25.440521] I
[glusterd-geo-rep.c:3907:glusterd_get_gsync_status_mst_slv] 0-:
geo-replication status prueba srv-vln-des3-priv1::back-cfe :session is not
active
[2015-02-23 08:51:25.440584] W
[glusterd-geo-rep.c:1840:glusterd_get_statefile_name] 0-: Config file
(/var/lib/glusterd/geo-replication/prueba_srv-vln-des3-priv1_back-cfe/gsyncd.conf)
missing. Looking for template config file
(/var/lib/glusterd/geo-replication/gsyncd_template.conf)
[2015-02-23 08:51:25.440598] I
[glusterd-geo-rep.c:1849:glusterd_get_statefile_name] 0-: Using default
config template(/var/lib/glusterd/geo-replication/gsyncd_template.conf).
[2015-02-23 08:51:25.524719] I
[glusterd-geo-rep.c:3926:glusterd_get_gsync_status_mst_slv] 0-:
/var/lib/glusterd/geo-replication/prueba_srv-vln-des3-priv1_back-cfe/ssh%3A%2F%2Froot%40192.168.64.207%3Agluster%3A%2F%2F127.0.0.1%3Aback-cfe.status
statefile not present.
[2015-02-23 08:51:25.524880] I
[glusterd-handler.c:1280:__glusterd_handle_cli_get_volume] 0-glusterd:
Received get vol req
[2015-02-23 08:51:25.627568] I
[glusterd-handler.c:3803:__glusterd_handle_status_volume] 0-management:
Received status volume req for volume prueba
[2015-02-23 08:51:25.627948] W [glusterd-locks.c:550:glusterd_mgmt_v3_lock]
(--> /usr/lib64/libglusterfs.so.0(_gf_log_callingfn+0x186)[0x7fecabad94c6]
(-->
/usr/lib64/glusterfs/3.6.2/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_lock+0x4cb)[0x7feca157510b]
(-->
/usr/lib64/glusterfs/3.6.2/xlator/mgmt/glusterd.so(gd_sync_task_begin+0x57d)[0x7feca15726bd]
(-->
/usr/lib64/glusterfs/3.6.2/xlator/mgmt/glusterd.so(glusterd_op_begin_synctask+0x2c)[0x7feca15729ec]
(-->
/usr/lib64/glusterfs/3.6.2/xlator/mgmt/glusterd.so(__glusterd_handle_status_volume+0x152)[0x7feca14d9852]
) 0-management: Lock for prueba held by
af083e0a-737e-4aa0-9f00-0e0efcd477df
[2015-02-23 08:51:25.628044] E [glusterd-syncop.c:1578:gd_sync_task_begin]
0-management: Unable to acquire lock for prueba


How I can reset this situation?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Changelogs on gluster 3.6

2015-02-20 Thread Félix de Lelelis
Hi,

There is anyway to take information about the last changelog that is
applied on slave and master  node in geo-replication?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Split-brain info gluster 3.6

2015-02-20 Thread Félix de Lelelis
Hi,

I generated a split-brain condition, and I solved it but with "gluster
volume heal vol_name info split-brain"  yet I can see past entries solved:

Number of entries: 3
atpath on brick
---
2015-02-19 17:13:08 /split
2015-02-19 17:14:09 /split
2015-02-19 17:15:10 /split

Brick srv-vln-des2-priv1:/gfs-to-snap/prueba/brick1/brick
Number of entries: 4
atpath on brick
---
2015-02-19 17:09:32 /split
2015-02-19 17:13:08 /split
2015-02-19 17:14:09 /split
2015-02-19 17:15:10 /split

How can I reset that entries?

Thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Splitmount on gluster 3.6

2015-02-18 Thread Félix de Lelelis
Hi,

I wolud like  to know if splitmount utility is useful for gluster version
upper 3.3. I think that since 3.3 version I need to delete the hardlink
also in split-brain case, it's true?

Thank's
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Simulate split-brain on gluster 3.6

2015-02-16 Thread Félix de Lelelis
Hi,

I am simulating a split brain condition on my cluster but I don't be able
it. I have disconnected the nodes and creating a file with the same name
and different contents but always the self-heal process take the last copy
of the file.

How can create thos condition?

thanks
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Warning: "W [socket.c:611:__socket_rwv] 0-management: readv on "

2015-02-15 Thread Félix de Lelelis
Hi,

The last week upgraded us cluster to 3.6 version. I noticed in the log the
following error:

W [socket.c:611:__socket_rwv] 0-management: readv on
/var/run/f3fcde54ca5d30115274155a37baa079.socket failed (Invalid argument)

It is due a nfs daemon?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Changelog dir in gluster 3.6

2015-02-15 Thread Félix de Lelelis
Hi,

I don't understand if the changelog is needed to replica and geo-replica
situations, so the filesystem is fill of a lot of that files in
/.glusterfs/changelogs. Is there anyway to reduce or waive directory?


Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] upgrade to gluster 3.6

2015-02-11 Thread Félix de Lelelis
Hi,

I don't know if this way it's the correct, so if I reported to other
account, please let me know.

I have a problem after upgrade my cluster to version 3.6 from version 3.5.
When I have started the volumes, I have seen a lot of errors:

[2015-02-11 11:23:18.231142] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:21.231749] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:24.232092] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:27.232648] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:30.233272] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:33.233672] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:36.234255] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:39.234688] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:42.235269] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:45.235867] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:48.236499] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)
[2015-02-11 11:23:51.237021] W [socket.c:611:__socket_rwv] 0-management:
readv on /var/run/032613d904825273481612e900965093.socket failed (Invalid
argument)


Can you help me please?

Thanks.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users