[Gluster-devel] geo-replication issue

2014-12-08 Thread Jan H Holtzhausen
Hi
I just know someone must have seen this before 
Geo-replication on 3.6.1, epel, Centos 6.6

Popen: ssh [2014-12-08 22:57:31.433468] E 
[socket.c:2267:socket_connect_finish] 0-glusterfs: connection to 
127.0.0.1:24007 failed (Connection refused)

Popen: ssh [2014-12-08 22:57:31.433484] T 
[socket.c:723:__socket_disconnect] 0-glusterfs: disconnecting 0x22c6550, 
state=0 gen=0 sock=6

Popen: ssh [2014-12-08 22:57:31.433497] D 
[socket.c:704:__socket_shutdown] 0-glusterfs: shutdown() returned -1. 
Transport endpoint is not connected

Popen: ssh [2014-12-08 22:57:31.433506] D 
[socket.c:2344:socket_event_handler] 0-transport: disconnecting now

Popen: ssh [2014-12-08 22:57:31.433541] T [cli.c:270:cli_rpc_notify] 
0-glusterfs: got RPC_CLNT_DISCONNECT



Is my google-fu deserting me in my hour of need?



BR

Jan



From:  Gong XiaoHui xhg...@wind.com.cn
Date:  Monday 08 December 2014 at 7:45 AM
To:  'gluster-devel@gluster.org' gluster-devel@gluster.org
Subject:  [Gluster-devel] some issues about geo-replication and gfapi

 

Hi

 I have two questions with glusterfs:

1.When I use libgfapi to write a file to a volume, which is configured as 
a geo-replication masterVol. It cannot work well, the geo status is 
faulty, following is the log: 

The glusterfs version is 3.6.1

 

[2014-12-08 13:12:11.708616] I [master(/data_xfs/geo2-master):1330:crawl] 
_GMaster: finished hybrid crawl syncing, stime: (1418010314, 0)

[2014-12-08 13:12:11.709706] I 
[master(/data_xfs/geo2-master):480:crawlwrap] _GMaster: primary master 
with volume id d220647c-5730-4cef-a89b-932470c914d2 ...

[2014-12-08 13:12:11.735719] I 
[master(/data_xfs/geo2-master):491:crawlwrap] _GMaster: crawl interval: 3 
seconds

[2014-12-08 13:12:11.811722] I [master(/data_xfs/geo2-master):1182:crawl] 
_GMaster: slave's time: (1418010314, 0)

[2014-12-08 13:12:11.840826] E [repce(/data_xfs/geo2-master):207:__call__] 
RepceClient: call 8318:139656072095488:1418015531.84 (entry_ops) failed on 
peer with OSError

[2014-12-08 13:12:11.840990] E 
[syncdutils(/data_xfs/geo2-master):270:log_raise_exception] top: FAIL:

Traceback (most recent call last):

  File /usr/libexec/glusterfs/python/syncdaemon/gsyncd.py, line 164, in 
main

main_i()

  File /usr/libexec/glusterfs/python/syncdaemon/gsyncd.py, line 643, in 
main_i

local.service_loop(*[r for r in [remote] if r])

  File /usr/libexec/glusterfs/python/syncdaemon/resource.py, line 1344, 
in service_loop

g2.crawlwrap()

  File /usr/libexec/glusterfs/python/syncdaemon/master.py, line 529, in 
crawlwrap

self.crawl(no_stime_update=no_stime_update)

  File /usr/libexec/glusterfs/python/syncdaemon/master.py, line 1194, in 
crawl

self.process(changes)

  File /usr/libexec/glusterfs/python/syncdaemon/master.py, line 946, in 
process

self.process_change(change, done, retry)

  File /usr/libexec/glusterfs/python/syncdaemon/master.py, line 910, in 
process_change

self.slave.server.entry_ops(entries)

  File /usr/libexec/glusterfs/python/syncdaemon/repce.py, line 226, in 
__call__

return self.ins(self.meth, *a)

  File /usr/libexec/glusterfs/python/syncdaemon/repce.py, line 208, in 
__call__

raise res

OSError: [Errno 12] Cannot allocate memory

[2014-12-08 13:12:11.842178] I 
[syncdutils(/data_xfs/geo2-master):214:finalize] top: exiting.

[2014-12-08 13:12:11.843421] I [repce(agent):92:service_loop] RepceServer: 
terminating on reaching EOF.

[2014-12-08 13:12:11.843682] I [syncdutils(agent):214:finalize] top: 
exiting.

[2014-12-08 13:12:12.103255] I [monitor(monitor):222:monitor] Monitor: 
worker(/data_xfs/geo2-master) died in startup phase

 

 

 

2.Another question, I cannot configure a geo-replication in 3.6.1 with the 
method in 3.4.1.

Any response is appreciate.

产品及时服务:请在Wind资讯终端上按“F1”键或致电客服专线400-820-Wind(9463)
---
宫晓辉

技术X01部

上海万得信息技术股份有限公司(Wind资讯)
Shanghai Wind Information Co., Ltd. 

上海市浦东新区福山路33号建工大厦9楼 200120
9/F Jian Gong Mansion,33 Fushan Road, Pudong New Area, 
Shanghai, P.R.C. 200120
Tel: (0086 21)6888 2280*8310
Fax: (0086 21)6888 2281
Email: xhg...@wind.com.cn 
http://www.wind.com.cn

 
___ Gluster-devel mailing list 
Gluster-devel@gluster.org 
http://supercolony.gluster.org/mailman/listinfo/gluster-devel
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
Are you referring to something else in your request? Meaning, you want 
/myfile, /dir1/myfile and /dir2/dir3/myfile to fall onto the same 
 bricks/subvolumes and that perchance is what you are looking for?

That is EXACTLY what I am looking for.
What are my chances?

BR
Jan

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
I think I have it.
Unless I’m totally confused, I can hash ONLY on the filename with:
glusterfs --volfile-server=a_server --volfile-id=a_volume \
   --xlator-option a_volume-dht.extra_hash_regex='.*[/]' \
   /a/mountpoint
Correct?

Jan

From:  Jan H Holtzhausen j...@holtztech.info
Date:  Tuesday 25 November 2014 at 9:06 PM
To:  gluster-devel@gluster.org
Subject:  Re: [Gluster-devel] EHT / DHT

Are you referring to something else in your request? Meaning, you want 
/myfile, /dir1/myfile and /dir2/dir3/myfile to fall onto the same 
 bricks/subvolumes and that perchance is what you are looking for?

That is EXACTLY what I am looking for.
What are my chances?

BR
Jan
___ Gluster-devel mailing list 
Gluster-devel@gluster.org 
http://supercolony.gluster.org/mailman/listinfo/gluster-devel
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
Hmm
Then something is wrong, 
If I upload 2 identical files, with different paths they only end up on 
the same server 1/4 of the time (I have 4 bricks).
I’ll test the regex quickly.

BR
Jan




On 2014/11/25, 7:55 PM, Shyam srang...@redhat.com wrote:

On 11/25/2014 02:28 PM, Jan H Holtzhausen wrote:
 I think I have it.
 Unless I’m totally confused, I can hash ONLY on the filename with:

 glusterfs --volfile-server=a_server --volfile-id=a_volume \
 --xlator-option a_volume-dht.extra_hash_regex='.*[/]' \
 /a/mountpoint

 Correct?

The hash of a file does not include the full path, it is on the file 
name _only_. So any regex will not work when the filename remains 
constant like myfile.

As Jeff explains the option is really to prevent using temporary parts 
of the name in the hash computation (for rename optimization). In this 
case, you do not seem to have any tmp parts to the name, like myfile 
and myfile~ should evaluate to the same hash, so remove all trailing 
'~' from the name.

So I am not sure the above is the option you are looking for.


 Jan

 From: Jan H Holtzhausen j...@holtztech.info 
mailto:j...@holtztech.info
 Date: Tuesday 25 November 2014 at 9:06 PM
 To: gluster-devel@gluster.org mailto:gluster-devel@gluster.org
 Subject: Re: [Gluster-devel] EHT / DHT

Are you referring to something else in your request? Meaning, you want

/myfile, /dir1/myfile and /dir2/dir3/myfile to fall onto the same

 bricks/subvolumes and that perchance is what you are looking for?


 That is EXACTLY whatI  am looking for.

 What are my chances?

As far as I know not much out of the box. As Jeff explained, the 
directory distribution/layout considers the GFID of the directory, hence 
each of the directories in the above example would/could get different 
ranges.

The file on the other hand remains constant myfile so its hash value 
remains the same, but due to the distribution range change as above for 
the directories, it will land on different bricks and not the same one.

Out of curiosity, why is this functionality needed?

Shyam
___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
So in a distributed cluster, the GFID tells all bricks what a files 
preceding directory structure looks like?
Where the physical file is saved is a function of the filename ONLY.
Therefore My requirement should be met by default, or am I being dense?

BR
Jan



On 2014/11/25, 8:15 PM, Shyam srang...@redhat.com wrote:

On 11/25/2014 03:11 PM, Jan H Holtzhausen wrote:
 STILL doesn’t work … exact same file ends up on 2 different bricks …
 I must be missing something.
 All I need is for:
 /directory1/subdirectory2/foo
 And
 /directory2/subdirectoryaaa999/foo


 To end up on the same brick….

This is not possible is what I was attempting to state in the previous 
mail. The regex filter is not for this purpose.

The hash is always based on the name of the file, but the location is 
based on the distribution/layout of the directory, which is different 
for each directory based on its GFID.

So there are no options in the code to enable what you seek at present.

Why is this needed?

Shyam

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
As to the why.
Filesystem cache hits.
Files with the same name tend to be the same files.

Regards
Jan




On 2014/11/25, 8:42 PM, Jan H Holtzhausen j...@holtztech.info wrote:

So in a distributed cluster, the GFID tells all bricks what a files 
preceding directory structure looks like?
Where the physical file is saved is a function of the filename ONLY.
Therefore My requirement should be met by default, or am I being dense?

BR
Jan



On 2014/11/25, 8:15 PM, Shyam srang...@redhat.com wrote:

On 11/25/2014 03:11 PM, Jan H Holtzhausen wrote:
 STILL doesn’t work … exact same file ends up on 2 different bricks …
 I must be missing something.
 All I need is for:
 /directory1/subdirectory2/foo
 And
 /directory2/subdirectoryaaa999/foo


 To end up on the same brick….

This is not possible is what I was attempting to state in the previous 
mail. The regex filter is not for this purpose.

The hash is always based on the name of the file, but the location is 
based on the distribution/layout of the directory, which is different 
for each directory based on its GFID.

So there are no options in the code to enable what you seek at present.

Why is this needed?

Shyam

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


Re: [Gluster-devel] EHT / DHT

2014-11-25 Thread Jan H Holtzhausen
I could tell you… 
But Symantec wouldn’t like it…..

From:  Poornima Gurusiddaiah pguru...@redhat.com
Date:  Wednesday 26 November 2014 at 7:16 AM
To:  Jan H Holtzhausen j...@holtztech.info
Cc:  gluster-devel@gluster.org
Subject:  Re: [Gluster-devel] EHT / DHT

Out of curiosity, what back end and deduplication solution are you using?

Regards,
Poornima

From: Jan H Holtzhausen j...@holtztech.info
To: Anand Avati av...@gluster.org, Shyam srang...@redhat.com, 
gluster-devel@gluster.org
Sent: Wednesday, November 26, 2014 3:43:36 AM
Subject: Re: [Gluster-devel] EHT / DHT

Yes we have deduplication at the filesystem layer

BR
Jan

From:  Anand Avati av...@gluster.org
Date:  Wednesday 26 November 2014 at 12:11 AM
To:  Jan H Holtzhausen j...@holtztech.info, Shyam srang...@redhat.com, 
gluster-devel@gluster.org
Subject:  Re: [Gluster-devel] EHT / DHT

Unless there is some sort of de-duplication under the covers happening in 
the brick, or the files are hardlinks to each other, there is no cache 
benefit whatsoever by having identical files placed on the same server.

Thanks,
Avati

On Tue Nov 25 2014 at 12:59:25 PM Jan H Holtzhausen j...@holtztech.info 
wrote:
As to the why.
Filesystem cache hits.
Files with the same name tend to be the same files.

Regards
Jan




On 2014/11/25, 8:42 PM, Jan H Holtzhausen j...@holtztech.info wrote:

So in a distributed cluster, the GFID tells all bricks what a files
preceding directory structure looks like?
Where the physical file is saved is a function of the filename ONLY.
Therefore My requirement should be met by default, or am I being dense?

BR
Jan



On 2014/11/25, 8:15 PM, Shyam srang...@redhat.com wrote:

On 11/25/2014 03:11 PM, Jan H Holtzhausen wrote:
 STILL doesn’t work … exact same file ends up on 2 different bricks …
 I must be missing something.
 All I need is for:
 /directory1/subdirectory2/foo
 And
 /directory2/subdirectoryaaa999/foo


 To end up on the same brick….

This is not possible is what I was attempting to state in the previous
mail. The regex filter is not for this purpose.

The hash is always based on the name of the file, but the location is
based on the distribution/layout of the directory, which is different
for each directory based on its GFID.

So there are no options in the code to enable what you seek at present.

Why is this needed?

Shyam

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel

___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel


___
Gluster-devel mailing list
Gluster-devel@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-devel