Re: [Gluster-users] moving drives containing bricks from one server to another

2017-07-18 Thread Ted Miller

  
  
It would help if you say what kind of
  volumes you have, as they all work a little differently.
  
  You should NOT destroy the old volumes.
  
  It should be quite possible to move the bricks to a new server and
  get them running as part of THE SAME gluster system that you have
  now.
  
  Until you tell us how many bricks, how many servers have what
  bricks, exactly which servers you are replacing (you mention only
  one server, but that is not a normal gluster configuration), we
  can't give you a whole lot of help.  The usual routine involves
  working with one brick server at a time, while keeping the rest of
  the array untouched.  With some replicated arrays and some careful
  work, you don't have to take the array offline at all.  However,
  replicated arrays get handled VERY differently from distributed,
  so we can't help much until you tell us more.
  
  Ted Miller
  Elkhart, IN
  
  On 07/18/2017 06:18 AM, Andy Tai wrote:


  

  hi,  I did not see a reply to my problem.  Let me ask it
in a different way...

  
  If I have bricks from a previous glusterfs volume and that
  volume is now gone because of the old machine was replaced,
  now I tried to create a new volume and add the old bricks to
  the new volume with the "force" opinion to "volume create". 
  The old data files are still in the bricks but when I mount
  the new volume the new volume shows it is empty.
  

Is it possible for glusterfs to recognize the old data files on
the bricks in some way to essentially re-create the same view as
the old volume, via some kind of healing or meta data re-scan or
re-creation?  The bricks are in good shape with no data
corruption and are in sync (same data were replicated).  Thanks
  
  
On Thu, May 4, 2017 at 10:03 PM, Andy
  Tai <a...@atai.org>
  wrote:
  

  
Hi, I have a gluster volume with bricks spread over
  several physical drives.   I now want to upgrade my
  server to a new system and plan to move the drives
  from the old server to the new server, with a
  different host name and IP address.  Can I shut down
  the gluster volume on the old server, move and install
  the physical drives containing the bricks to the new
  server, and then create a new gluster volume on the
  new server, and add the bricks to the new volume in
  the same way reflecting the previous organization of
  the volume on the old server, and expect everything to
  work (all files preserved and accessible via
  glusterfs, with no data loss?  
  

The gluster volume on the old server would be retired
and I want to let the new server taking over the role of
serving the gluster volume.

  
  Thanks
  
  

  

  -- 

  

  

  

  

  Andy Tai, a...@atai.org,
Skype: licheng.tai,
Line: andy_tai, WeChat:
andytai1010
Year 2017 民國106年
自動的精神力是信仰與覺悟
自動的行為力是勞動與技能

  

  

  

  

  

  

  

  




-- 

  

  

  

  

  Andy Tai, a...@atai.org,
   

Re: [Gluster-users] Moving a physical disk from one machine to another

2017-02-07 Thread Ted Miller

On 02/06/2017 11:11 AM, Scott Hazelhurst wrote:

Apologies if this is obvious but I’ve searched the documentation and mailing 
list and can’t see an answer.

One of my  servers which hosts a gluster brick has died beyond possible saving. 
The physical disk of the brick itself is good and I’d like not to lose anything.

Can I just add it as a brick on another server? Unfortunately, I am not able to 
host the disk on a machine with the same IP address/host name. I suppose I 
could just add a new empty disk and then copy the data from the old disk but 
I’d prefer to do things as easily as possible.

Scott





This communication is intended for the addressee only. It is confidential. If you have received this communication in error, please notify 
us immediately and destroy the original message. You may not copy or disseminate this communication without the permission of the University. Only authorised signatories are 
competent to enter into agreements on behalf of the University and recipients are thus advised that the content of this message may not be legally binding on the University 
and may contain the personal views and opinions of the author, which are not necessarily the views and opinions of The University of the Witwatersrand, Johannesburg. All 
agreements between the University and outsiders are subject to South African Law unless the University agrees in writing to the contrary. 

http://lists.gluster.org/mailman/listinfo/gluster-users


Did this a few months ago, when I had a server die.

Assuming it is a replicated volume, first thing is to remove old brick from 
your volume.  It has been a while since I had to do this, but I think you 
just do a "gluster remove-brick  ".  It will probably require a 'force' 
since the server is gone.


You should be able to install the disk in another server (with gluster 
installed).


Then just add that brick to the volume, like you did when setting up the 
volume.  You will again have to force, because the volume will already have 
a volume signature.


After that do a full heal, and wait.  It will save a good bit of time that 
the files are already on the disk.  It should add any files that have been 
changed, and then go on its way.  You may want to be sure that any files 
deleted since the server died don't get put back into the volume.  Have had 
that happen, but don't remember if this was the combination that triggered 
it, or something else.


Ted Miller
Indiana, USA





___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Mount gluster volume inside other mounted volume...

2016-08-25 Thread Ted Miller

  
  
On 08/25/2016 08:11 AM, Gilberto Nunes
  wrote:


  Hello list


I have two volumes, DATA and WORK.


DATA has size 500 GB 
WORK has size 1,2 TB


I can mount DATA with this command:


 mount -t glusterfs -o acl localhost:DATA /home


Everything is ok with that.
But, when I mont WORK volume inside /home/work, like this


mount -t glusterfs -o acl localhost:WORK /home/work


I realize that /home and /home/work has the same size:
df -h

  
localhost:WORK  539G   45M  539G   1% /home/work
localhost:DATA      539G   45M  539G   1% /home
  

  

Are the bricks for WORK and DATA both subdirectories in the same
partition on at one of your computers?  If the bricks are just
subdirectories, rather than dedicated partitions, df will look like
that.  I have some "odd" volumes that are subdirectories (for now)
and they look like that.  Anything I add to either volume will
increase the USED and decrease the AVAILABLE for both volumes.

What df shows is the figures for the partition that these are part
of, not for the particular subdirectory that you turned into a
brick.

If that isn't your setup, someone else will have to help you.

Ted Miller
Elkhart, IN, USA

  

  
  
  Is there a way to workaround this??
  
  
  Perhaps this is no a issue related to glusterfs, but I
need just so advice where found a possible solution
  
  
  Thanks anyway
  
  
  Best regards
  

  

  

  

  

  

  

  Gilberto Ferreira
  

  

  

  

  

  

  

  

  
  ___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users



  



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] 'nofail' fails miserably

2016-08-24 Thread Ted Miller
Now that most distros are switching to systemd, there seems to be a problem 
in the mount command's handling of the 'nofail' option.


Setup: using glusterfs 3.7.13 on Centos 7 (up to date) from the Centos 
Storage SIG repo.


The man page for systemd.mount says:

nofail
With nofail this mount will be only wanted, not required, by 
local-fs.target or
remote-fs.target. This means that the boot will continue even if this mount 
point is not
mounted successfully.

It works fine during bootup -- things end up mounted, but they don't timeout 
and throw the server into maintenance mode if they are a little slow.


However, if I do a 'mount -a' from the command line, and any gluster volume 
needs to be mounted, mount (I assume mount.glusterfs) throws the response:


Invalid option nofail

Somebody needs to take responsibility for either filtering out the 'nofail' 
option so that mount.glusterfs doesn't see it, or else glusterfs needs to be 
smart enough to recognize that, even though it is in the options space, it is 
OK to ignore it.  If it is OK for systemd, and the only place you can put it 
is in the mount options, then mount.glusterfs needs to be OK with that.


I have not tried the other options that systemd.mount uses, so some of them 
may also cause problems.  man systemd.mount lists these options as being used:

x-systemd.requires=
x-systemd.requires-mounts-for=
x-systemd.automount
x-systemd.device-timeout=
noauto
auto
nofail
x-initrd.mount

'noauto' has been around forever, so it is probably handled OK, (I think I 
used it a while back) but not all the new options are handled right, or else 
the documentation is bad.


I hope somebody can stick a couple of lines of code in so we can mount 
'nofail' volumes after boot.


Thank you,
Ted Miller
Elkhart, Indiana, USA


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Gluster replica over WAN...

2016-08-03 Thread Ted Miller

On 8/2/2016 8:14 AM, Gilberto Nunes wrote:


Hello list...
This is my first post on this list.

I have here two IBM Server, with 9 TB of hardisk on which one.
Between this servers, I have a WAN connecting two offices,let say OFFICE1 
and OFFICE2.

This WAN connection is over fibre channel.
When I setting up gluster with replica with two bricks, and mount the 
gluster volume in other folder, like this:


mount -t glusterfs localhost:/VOLUME /STORAGE


and when I go to that folder, and try to access the content, I get a lot of 
timeout... Even a single ls give a lot of time to return the list.


This folder, /STORAGE is access by many users, through samba share.

So, when OFFICE1 access the gluster server access the files over 
\\server\share, has a long delay to show the files Sometimes, get time out.


My question is: is there some way to improve the gluster to work faster in 
this scenario??


Thanks a lot.

Best regards

--

Gilberto Ferreira
+55 (47) 9676-7530
Skype: gilberto.nunes36



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users
I probably can't help, but I can tell you the kind of things that will help 
people to understand your problem better.


What is the bandwidth between sites?

What is the ping time between sites?

If you run "ping -c1 " from OFFICE1, how many pings fail?

What else can you tell us about the WAN connection that would help us 
understand your situation?


How many files are there in the \\server\share directory?

Are people at both sites actively writing files to the shared storage?  If 
not, you may need to look at gluster geo-replication. It is one way, so all 
writing would have to be done to OFFICE1, with OFFICE2 having read-only 
access.  (Some things that works wonderfully for, other things it doesn't 
work at all.)


I can also tell you that no one who has tried it will endorse running a 
production replica 2 cluster over WAN.  You are asking for either downtime or 
split-brain or both.  Replica 3 is the minimum for production use of a 
replicate cluster, even over LAN.


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] 3.7.13 two node ssd solid rock

2016-08-03 Thread Ted Miller

On 8/3/2016 11:13 AM, Leno Vo wrote:

One of my gluster 3713 is on two nodes only with samsung ssd 1tb pro raid 5 
x3,it already crashed two time because of brown out and block out, it got 
production vms on it, about 1.3TB.


Never got split-brain, and healed quickly.  Can we say 3.7.13 two nodes 
with ssd is solid rock or just lucky?


My other gluster is on 3 nodes 3713, but one node never got up (old server 
proliant wants to retire), ssh raid 5 with combination sshd lol laptop 
seagate, it never healed about 586 occurences but there's no split-brain 
too.  and vms are intact too, working fine and fast.


ahh never turned on caching on the array, the esx might not come up right 
away, u need to go to setup first to make it work and restart and then you 
can go to array setup (hp array F8) and turned off caching.  then esx 
finally boot up.



___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users
I would say you are very lucky.  I would not use anything less that replica 3 
in production.


Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Problem with Glusterfs - taking over CPU time on clients

2016-07-01 Thread Ted Miller

On 7/1/2016 5:22 AM, Milos Kurtes wrote:


Hi,

I have 7 clients on AWS in ELB and every of them is connected to glusterfs 
servers group of two.


On clients, version of glusterfs is glusterfs 3.7.1 built on Nov 22 2015 
17:39:20 and
on server side both of servers have same glusterfs 3.6.2 built on Jan 22 
2015 12:58:10


The major and urgent problem is the situation after disconnecting and 
connecting one of the servers, on one of the clients (and that is changing 
which one after few minutes) are overtaken with glusterfs process (all 
apache processes are on sleep in the same moment) which is responsible for 
rw session files. This is a very frequent job.


Without disconnection, everything working except filling log files with 
a big amount of messages (probably incompatibility versions of glusterfs on 
servers and clients side why I tried yesterday to get the update for server 
side).


Do anybody have a solution for this overtaking situation?

Milos Kurtes, System Administrator, Alison Co.
I'm no expert, but it sounds like the heal process hogging resources healing 
(or checking for differences) the replicated volume back into the cluster.

Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] About Gluster cluster availability when one of out of two nodes is down

2016-06-30 Thread Ted Miller
Is it not the default behavior that if a volume looses quorum, the files are 
still available in read-only mode?  If so, it makes sense to me that this 
behavior would continue after a reboot. Otherwise the user is going to be 
very confused.


Ted Miller
Elkhart, IN, USA


On 6/30/2016 2:10 AM, Atin Mukherjee wrote:
Currently on a two node set up, if node B goes down and node A is rebooted 
brick process(es) on node A doesn't come up to avoid split brains. However 
we have had concerns/bugs from different gluster users on the availability 
with this configuration. So we can solve this issue by starting the brick 
process(es) if quorum is not enabled. If quorum is enabled we'd not. 
Although quorum option really doesn't make sense in a two node cluster, but 
we can leverage this option to get rid of this specific situation.


I'd like to know your feedback on this and then I push a patch right away.

~Atin


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] selinux status on RHEL/Centos 7

2016-06-29 Thread Ted Miller
What is the status of selinux tagging on Centos 7?  I have read enough to 
know that this is a chain-like process requiring changes in the client, the 
server, FUSE, and the kernel to make it all work.  What is the current status 
of this process on Centos 7?


My use-case: I need to allow Apache to access files that are stored on 
gluster and mounted using FUSE.  What are my options (besides shutting down 
selinux for the Apache process)?


Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] reading from local replica?

2015-06-09 Thread Ted Miller

On 6/8/2015 5:55 PM, Brian Ericson wrote:

Am I misunderstanding cluster.read-subvolume/cluster.read-subvolume-index?

I have two regions, "A" and "B" with servers "a" and "b" in, respectfully, 
each region.  I have clients in both regions. Intra-region communication is 
fast, but the pipe between the regions is terrible.  I'd like to minimize 
inter-region communication to as close to glusterfs write operations only 
and have reads go to the server in the region the client is running in.


I have created a replica volume as:
gluster volume create gv0 replica 2 a:/data/brick1/gv0 b:/data/brick1/gv0 
force


As a baseline, if I use scp to copy from the brick directly, I get -- for a 
100M file -- times of about 6s if the client scps from the server in the 
same region and anywhere from 3 to 5 minutes if I the client scps the 
server in the other region.


I was under the impression (from something I read but can't now find) that 
glusterfs automatically picks the fastest replica, but that has not been my 
experience; glusterfs seems to generally prefer the server in the other 
region over the "local" one, with times usually in excess of 4 minutes.


I've also tried having clients mount the volume using the "xlator" options 
cluster.read-subvolume and cluster.read-subvolume-index, but neither seem 
to have any impact.  Here are sample mount commands to show what I'm 
attempting:


mount -t glusterfs -o xlator-option=cluster.read-subvolume=gv0-client-<0 or 
1> a:/gv0 /mnt/glusterfs
mount -t glusterfs -o xlator-option=cluster.read-subvolume-index=<0 or 1> 
a:/gv0 /mnt/glusterfs


Am I misunderstanding how glusterfs works, particularly when trying to 
"read locally"?  Is it possible to configure glusterfs to use a local 
replica (or the "fastest replica") for reads? 
I am not a developer, nor intimately familiar with the insides of glusterfs, 
but here is how I understand that glusterfs-fuse file reads work.
First, all replica bricks are read, to make sure they are consistent.  (If 
not, gluster tries to make them consistent before proceeding).
After consistency is established, then the actual read occurs from the brick 
with the shortest response time.  I don't know when or how the response time 
is measured, but it seems to work for most people most of the time.  (If the 
client is on one of the brick hosts, it will almost always read from the 
local brick.)


If the file reads involve a lot of small files, the consistency check may be 
what is killing your response times, rather than the read of the file 
itself.  Over a fast LAN, the consistency checks can take many times the 
actual read time of the file.


Hopefully others will chime in with more information, but if you can supply 
more information about what you are reading, that will help too.  Are you 
reading entire files, or just reading in a lot of "snippets" or what?


Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] The strange behavior whose common denominator is gluster

2015-06-09 Thread Ted Miller

On 6/9/2015 8:30 AM, Pablo Silva wrote:

Hi Ted!

Thanks for your reply, for the implementation of Service N2 (As2 
Mendelson B45 + gluster 3.3.1-15.el6.x86_64 (1 brick)

we have implemented  one brick


As you can see, Mendelson has a Brick, and use it for to process teh 
Purchase Ordes feed in SOA21 by MULE process


[root@mendelson ~]# gluster volume info

Volume Name: m2mas2
Type: Distribute
Volume ID: 328d1e48-bb0a-4b41-92ac-699f78d2dca7
Status: Started
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: mendelson.X.b2b:/opt/mendelson/messages
Options Reconfigured:
auth.allow: 10.200.20.*,10.200.22.*
storage.owner-uid: 500
storage.owner-gid: 500



[root@mendelson ~]# cat /etc/fstab

#
# /etc/fstab
# Created by anaconda on Wed Oct 23 11:42:22 2013
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
/dev/mapper/VolGroup-lv_root / ext4defaults,acl1 1
UUID=df46586a-2f06-4ada-b2e6-523f7ec3967b /boot   ext4
defaults1 2

/dev/mapper/VolGroup-lv_swap swap swapdefaults0 0
tmpfs   /dev/shm tmpfs   defaults0 0
devpts  /dev/pts devpts  gid=5,mode=620  0 0
sysfs   /sys sysfs   defaults0 0
proc/proc procdefaults0 0


Here is your problem.  You are not mounting m2mas2. /opt/mendelson/messages 
is NOT the same as m2mas2.  You can NEVER write (or read) to 
/opt/mendelson/messages.  You must pretend that is _does not exist_.  You 
must mount the Gluster volume, and use only the path to the gluster volume, 
never to the brick.  When you write to the brick you are "going behind 
gluster's back" and gluster does not know that anything has changed, so it 
does not make those files available to the gluster clients.  It may 
eventually notice them, and make them available, but it is not anything 
dependable or predictable.


First, make sure you have the glusterfs-fuse package installed.

You need to add a line to /etc/fstab that is like:

mendelson:/m2mas2/opt/m2m.as2   glusterfs  defaults   0 0


Then change everything that you are now reading and writing to 
/opt/mendelson/messages so that it writes to /opt/m2m.as2 (or whatever mount 
point you have chosen).


Many of us follow this pattern. All bricks are located at /bricks/xxx/xxx.  
All gluster files systems are mounted to /gluster/xxx.  That makes the 
separation between the bricks and the gluster volumes very clear, and very 
hard to mess up when writing file paths in programs.  No trying to remember 
"Which is the brick, and which is the gluster volume?"


On your Apache system, you again need to mount the glusterfs volume in a 
separate place, and make sure that the glusterfs location is the only one 
that Apache knows about or uses in any way whatsoever.


Ted Miller



In this machine run MULE process for to feed with Purchase Orders to Mendelson.

[root@soa21ifh ~]# mount
/dev/mapper/VolGroup00-LogVol00 on / type ext3 (rw,acl)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
/dev/xvda1 on /boot type ext3 (rw)
tmpfs on /dev/shm type tmpfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw)
glusterfs#mendelson:/m2mas2 on /opt/m2m.as2 type fuse 
(rw,allow_other,max_read=131072)


maybe you.'re right and we are wrong with the way of architecture


Thanks in Advance

-Pablo

On Mon, Jun 8, 2015 at 6:06 PM, Ted Miller <mailto:tmil...@sonsetsolutions.org>> wrote:


Are you sure you have mounted the gluster volume, and are writing to
the gluster volume, and NOT to the brick?  What you describe can happen
when you write to the brick instead of the gluster volume.  You can see
here:
http://www.gluster.org/community/documentation/index.php/QuickStart in
steps 6 and 7.  If you do not understand the difference, include the
output of the 'mount' command from one of your servers.
Ted Miller
Elkhart, IN, USA


On 6/5/2015 8:46 AM, Pablo Silva wrote:

Dear Colleagues:

We are using gluster in 3.3.1-15.el6.x86_64 and GlusterFS-3.6.2-1.el5
versions, we have two types of service:

1) Apache httpd-2.2.3-91.el5.centos + GlusterFS-3.6.2-1.el5 (two bricks)

2) AS2 Mendelson B45 + gluster 3.3.1-15.el6.x86_64 (two bricks)

It is different services, a common problem, which I will explain

Service N1 (Apache httpd-2.2.3-91.el5.centos + GlusterFS-3.6.2-1.el5
(two bricks))

---
We have a high-availability architecture, in which there are two
Apache servers see a directory that is hosted on a gluster long ago we
had a problem where an Apache s

Re: [Gluster-users] nfs-ganesha/samba vfs and replica redundancy

2015-06-08 Thread Ted Miller

On 6/3/2015 3:15 AM, Benjamin Kingston wrote:
Can someone give me a hint on the best way to maintain data availability to 
a share on a third system using nfs-ganesha and samba?


I currently have a round-robbin dns entry that nfs ganesha/samba uses, 
however even with a short ttl, there's brief downtime when a replica node 
fails. I can't see in the samba VFS or ganesha fsal syntax where a 
secondary address can be provided.


I've tried comma seperated, space seperated, with/without quotes for 
multiple IP's and only seen issues.
Any reason you aren't using a "floating IP address"?  This isn't the newest 
talk, but the concepts have not changed: 
http://events.linuxfoundation.org/sites/events/files/lcjpcojp13_nakai.pdf


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] The strange behavior whose common denominator is gluster

2015-06-08 Thread Ted Miller
Are you sure you have mounted the gluster volume, and are writing to the 
gluster volume, and NOT to the brick?  What you describe can happen when you 
write to the brick instead of the gluster volume. You can see here: 
http://www.gluster.org/community/documentation/index.php/QuickStart in steps 
6 and 7.  If you do not understand the difference, include the output of the 
'mount' command from one of your servers.

Ted Miller
Elkhart, IN, USA

On 6/5/2015 8:46 AM, Pablo Silva wrote:

Dear Colleagues:

We are using gluster in 3.3.1-15.el6.x86_64 and GlusterFS-3.6.2-1.el5 
versions, we have two types of service:


1) Apache httpd-2.2.3-91.el5.centos + GlusterFS-3.6.2-1.el5 (two bricks)

2) AS2 Mendelson B45 + gluster 3.3.1-15.el6.x86_64 (two bricks)

It is different services, a common problem, which I will explain

Service N1 (Apache httpd-2.2.3-91.el5.centos + GlusterFS-3.6.2-1.el5 (two 
bricks))

---
We have a high-availability architecture, in which there are two Apache 
servers see a directory that is hosted on a gluster long ago we had a 
problem where an Apache server could list the files and submit them for 
download, while the other Apache server that is watching the same directory 
with the same files gluster indicated that there were no files for download.


Feeding gluster files to that directory, MULE performed asynchronously.In 
summary, an Apache server could access files and another did not give aware 
of their existence, as the directory and the same files.


Service N2 (As2 Mendelson B45 + gluster 3.3.1-15.el6.x86_64 (two bricks) )
--
We have only one Mendelson AS2 Server B45 running with gluster (two bricks),
The operations of mendelson is quite simple, is to observe the presence of 
files in a directory every 5 seconds and sent to the partner, the directory 
is hosted in gluster, the issue that every certain amount of time not 
Mendelson AS2 takes cognizance the existence of files in the directory, 
even if you enter the directory notes of its existence


In both cases, different services being the only common denominator is 
gluster, someone else is experiencing this problem?


Have we not set the service gluster well and we are repeating the same 
mistake ?, or is it a bug?


Thanks in advance
Pablo


___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


--
*Ted Miller*, Design Engineer
*SonSet Solutions*
(formerly HCJB Global Technology Center)
my desk +1 574.970.4272
receptionist +1 574.972.4252
http://sonsetsolutions.org

/Technology for abundant life!/
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster 3.7.0 released

2015-05-27 Thread Ted Miller

responses below
Ted Miller

On 5/26/2015 12:01 AM, Atin Mukherjee wrote:


On 05/26/2015 03:12 AM, Ted Miller wrote:

From: Niels de Vos 
Sent: Monday, May 25, 2015 4:44 PM

On Mon, May 25, 2015 at 06:49:26PM +, Ted Miller wrote:


From: Humble Devassy Chirammal 
Sent: Monday, May 18, 2015 9:37 AM
Hi All,

GlusterFS 3.7.0 RPMs for RHEL, CentOS, Fedora and packages for Debian are available 
at download.gluster.org<http://download.gluster.org> [1].

[1] http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.0/

--Humble


On Thu, May 14, 2015 at 2:49 PM, Vijay Bellur 
mailto:vbel...@redhat.com>> wrote:

Hi All,

I am happy to announce that Gluster 3.7.0 is now generally available. 3.7.0 
contains several

[snip]

Cheers,
Vijay

[snip]

What happened to packages for RHEL/Centos 5?  I have the (probably
unusual--added gluster to existing servers) setup of running a replica
3 cluster where two nodes run on Centos 6 and one is still on Centos
5.  This is a personal setup, and I have been using
http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-5/x86_64/repodata/repomod.xml
as my repo.  It has worked fine for a while, but this time the two
Centos 6 nodes updated to 3.7, but the Centos 5 node got left behind
at 3.6.3.

Packages for RHEL/CentOS-5 are not available yet. These will follow
later. Thare are some changes needed to be able to build the packages on
EL5. Because we are currently stabilizing our CI/regression tests, we do
not merge any other changes. Until we provide packages in our
repository, you could apply patch http://review.gluster.org/10803
yourself and build the EL5 version. I expect that we will do a release
in 2-3 weeks which will have EL5 RPMs too.

I have no idea about the problem below, it sounds like something the
GlusterD developers could help with.

Niels


Command 'gluster volume status' on the C5 machine makes everything
look fine:

Status of volume: ISO2
Gluster process   PortOnline  Pid
--
Brick 10.x.x.2:/bricks/01/iso249162   Y   4679
Brick 10.x.x.4:/bricks/01/iso249183   Y   6447
Brick 10.x.x.9:/bricks/01/iso249169   Y   1985

But the same command on either of the C6 machines shows the C5 machine
(10.x.x.2) missing in action (though it does recognize that there are
NFS and heal daemons there):

Status of volume: ISO2
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick 10.41.65.4:/bricks/01/iso249183 0  Y   6447
Brick 10.41.65.9:/bricks/01/iso249169 0  Y   1985
NFS Server on localhost 2049  0  Y   2279
Self-heal Daemon on localhost   N/A   N/AY   2754
NFS Server on 10.41.65.22049  0  Y   4757
Self-heal Daemon on 10.41.65.2  N/A   N/AY   4764
NFS Server on 10.41.65.42049  0  Y   6543
Self-heal Daemon on 10.41.65.4  N/A   N/AY   6551

So, is this just an oversight (I hope), or has support for C5 been dropped?
If support for C5 is gone, how do I downgrade my Centos6 machines back
to 3.6.x? (I know how to change the repo, but the actual sequence of
yum commands and gluster commands is unknown to me).

Could you attach the glusterd log file of 10.x.x.2 machine
attached as etc-glusterfs-glusterd.vol.log.newer.2, starting from last 
machine reboot

  and the node from where you triggered volume status.

attached as etc-glusterfs-glusterd.vol.log.newer4 starting same time as .2 log

Could you also share gluster volume info output of all the nodes?

I have several volumes, so I chose the one that shows up first on the listings:

*from 10.41.65.2:*

[root@office2 /var/log/glusterfs]$ gluster volume info

Volume Name: ISO2
Type: Replicate
Volume ID: 090da4b3-c666-41fe-8283-2c029228b3f7
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: 10.41.65.2:/bricks/01/iso2
Brick2: 10.41.65.4:/bricks/01/iso2
Brick3: 10.41.65.9:/bricks/01/iso2

[root@office2 /var/log/glusterfs]$ gluster volume status ISO2
Status of volume: ISO2
Gluster process PortOnline  Pid
--
Brick 10.41.65.2:/bricks/01/iso2 49162   Y   4463
Brick 10.41.65.4:/bricks/01/iso2 49183   Y   6447
Brick 10.41.65.9:/bricks/01/iso2 49169   Y   1985
NFS Server on localhost 2049Y   4536
Self-heal Daemon on localhost N/A Y   4543
NFS Server on 10.41.65.9 2049Y   2279
Self-heal Daemon on 10.41.65.9 N/A Y   2754
NFS Server on 10.41.65.4 2049Y   6543
Self-heal Da

Re: [Gluster-users] [Gluster-devel] Gluster 3.7.0 released

2015-05-25 Thread Ted Miller

From: Niels de Vos 
Sent: Monday, May 25, 2015 4:44 PM

On Mon, May 25, 2015 at 06:49:26PM +, Ted Miller wrote:
>
> 
> From: Humble Devassy Chirammal 
> Sent: Monday, May 18, 2015 9:37 AM
> Hi All,
>
> GlusterFS 3.7.0 RPMs for RHEL, CentOS, Fedora and packages for Debian are 
> available at download.gluster.org<http://download.gluster.org> [1].
>
> [1] http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.0/
>
> --Humble
>
>
> On Thu, May 14, 2015 at 2:49 PM, Vijay Bellur 
> mailto:vbel...@redhat.com>> wrote:
>
> Hi All,
>
> I am happy to announce that Gluster 3.7.0 is now generally available. 3.7.0 
> contains several
>
> [snip]
>
> Cheers,
> Vijay
>
> [snip]
>
> What happened to packages for RHEL/Centos 5?  I have the (probably
> unusual--added gluster to existing servers) setup of running a replica
> 3 cluster where two nodes run on Centos 6 and one is still on Centos
> 5.  This is a personal setup, and I have been using
> http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-5/x86_64/repodata/repomod.xml
> as my repo.  It has worked fine for a while, but this time the two
> Centos 6 nodes updated to 3.7, but the Centos 5 node got left behind
> at 3.6.3.

Packages for RHEL/CentOS-5 are not available yet. These will follow
later. Thare are some changes needed to be able to build the packages on
EL5. Because we are currently stabilizing our CI/regression tests, we do
not merge any other changes. Until we provide packages in our
repository, you could apply patch http://review.gluster.org/10803
yourself and build the EL5 version. I expect that we will do a release
in 2-3 weeks which will have EL5 RPMs too.

I have no idea about the problem below, it sounds like something the
GlusterD developers could help with.

Niels

> Command 'gluster volume status' on the C5 machine makes everything
> look fine:
>
> Status of volume: ISO2
> Gluster process   PortOnline  Pid
> --
> Brick 10.x.x.2:/bricks/01/iso249162   Y   4679
> Brick 10.x.x.4:/bricks/01/iso249183   Y   6447
> Brick 10.x.x.9:/bricks/01/iso249169   Y   1985
>
> But the same command on either of the C6 machines shows the C5 machine
> (10.x.x.2) missing in action (though it does recognize that there are
> NFS and heal daemons there):
>
> Status of volume: ISO2
> Gluster process TCP Port  RDMA Port  Online  Pid
> --
> Brick 10.41.65.4:/bricks/01/iso249183 0  Y   6447
> Brick 10.41.65.9:/bricks/01/iso249169 0  Y   1985
> NFS Server on localhost 2049  0  Y   2279
> Self-heal Daemon on localhost   N/A   N/AY   2754
> NFS Server on 10.41.65.22049  0  Y   4757
> Self-heal Daemon on 10.41.65.2  N/A   N/AY   4764
> NFS Server on 10.41.65.42049  0  Y   6543
> Self-heal Daemon on 10.41.65.4  N/A   N/AY   6551
>
> So, is this just an oversight (I hope), or has support for C5 been dropped?
> If support for C5 is gone, how do I downgrade my Centos6 machines back
> to 3.6.x? (I know how to change the repo, but the actual sequence of
> yum commands and gluster commands is unknown to me).
>
> Ted Miller
> Elkhart, IN, USA


Thanks for the information.  As long as I know it is coming, I can improvise 
and hang on.

I am assuming that the problem with the .2 machine not being seen is a result 
of running a cluster with a version split.

Ted Miller
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] [Gluster-devel] Gluster 3.7.0 released

2015-05-25 Thread Ted Miller


From: Humble Devassy Chirammal 
Sent: Monday, May 18, 2015 9:37 AM
Hi All,

GlusterFS 3.7.0 RPMs for RHEL, CentOS, Fedora and packages for Debian are 
available at download.gluster.org<http://download.gluster.org> [1].

[1] http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.0/

--Humble


On Thu, May 14, 2015 at 2:49 PM, Vijay Bellur 
mailto:vbel...@redhat.com>> wrote:

Hi All,

I am happy to announce that Gluster 3.7.0 is now generally available. 3.7.0 
contains several

[snip]

Cheers,
Vijay

[snip]

What happened to packages for RHEL/Centos 5?  I have the (probably 
unusual--added gluster to existing servers) setup of running a replica 3 
cluster where two nodes run on Centos 6 and one is still on Centos 5.  This is 
a personal setup, and I have been using 
http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-5/x86_64/repodata/repomod.xml
 as my repo.  It has worked fine for a while, but this time the two Centos 6 
nodes updated to 3.7, but the Centos 5 node got left behind at 3.6.3.  Command 
'gluster volume status' on the C5 machine makes everything look fine:

Status of volume: ISO2
Gluster process   PortOnline  Pid
--
Brick 10.x.x.2:/bricks/01/iso249162   Y   4679
Brick 10.x.x.4:/bricks/01/iso249183   Y   6447
Brick 10.x.x.9:/bricks/01/iso249169   Y   1985

But the same command on either of the C6 machines shows the C5 machine 
(10.x.x.2) missing in action (though it does recognize that there are NFS and 
heal daemons there):

Status of volume: ISO2
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick 10.41.65.4:/bricks/01/iso249183 0  Y   6447
Brick 10.41.65.9:/bricks/01/iso249169 0  Y   1985
NFS Server on localhost 2049  0  Y   2279
Self-heal Daemon on localhost   N/A   N/AY   2754
NFS Server on 10.41.65.22049  0  Y   4757
Self-heal Daemon on 10.41.65.2  N/A   N/AY   4764
NFS Server on 10.41.65.42049  0  Y   6543
Self-heal Daemon on 10.41.65.4  N/A   N/AY   6551

So, is this just an oversight (I hope), or has support for C5 been dropped?
If support for C5 is gone, how do I downgrade my Centos6 machines back to 
3.6.x? (I know how to change the repo, but the actual sequence of yum commands 
and gluster commands is unknown to me).

Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] ... i was able to produce a split brain...

2015-02-03 Thread Ted Miller


On 2/3/2015 2:44 PM, Joe Julian wrote:


On 02/03/2015 11:34 AM, Ted Miller wrote:


On 2/3/2015 12:23 PM, Joe Julian wrote:



That brought another thought to mind (have not had reason to try it):
How does gluster cope if you go behind its back and rename a "rejected" 
file?  For instance, in my example above, what if I go directly on the 
brick and rename the host-2 copy of the file to hair-pulling.txt-dud?  
The ideal scenario would seem to be that if user does a heal it would 
treat the copy as new file, see no dupe for hair-pulling.txt, and create 
a new dupe on host-2.  Since hair-pulling.txt-dud is also a new file, a 
dupe would be created on host-1.  User could then access files, verify 
correctness, and then delete hair-pulling.txt-dud.
This should cause you to have two files with the same gfid. This will 
create the hardlink in .glusterfs again, and the heal will then re-create 
the .txt file also with that same gfid. Since both files will have the 
same gfid (stored in extended attributes) and be hard linked to the same 
file under .glusterfs you should then end up with both files being 
split-brain. 
Joe, I moved you comments up to be closest to the proposal they seem 
relevant to.


*
A not-officially-sanctioned way that I dealt with a split-brain a few 
versions back:

1. decided I wanted to keep file on host-2
2. log onto host-2
3. cp /brick/data1/hair-pulling.txt /gluster/data1/hair-pulling.txt-dud
4. rm /brick/data1/hair-pulling.txt
5. follow some Joe Julian blog stuff to delete the "invisible fork" of file
6. gluster volume heal data1 all
If you note, in the above scenario _I copied from the brick to the mounted 
gluster volume_.  I believe that this forces the breaking of any linkage 
between the old file and the new one.  Am I missing something there?

Yep, I missed that. I seem to be suffering from split-brain, myself, today.

Don't sweat it, your blog posts have dug me out of more than one hole.

Ted Miller
Elkhart, IN, USA


.

//
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] 答复: Re: replica position

2015-02-03 Thread Ted Miller


On 1/29/2015 8:13 PM, yang.bi...@zte.com.cn wrote:


example: brick A ,brick B,brick C compose one volume

When client(fuse) in brick A write data into volume,for some reason ,we 
want the data  wrote into brick B or brick C.
Are you considering a replicated volume?  If so, the _client_ will write to 
all three volumes at the same time.


If you are looking at a distributed volume, then each file ends up on only 
one node, but file writing is distributed across different nodes.


Ted Miller
Elkhart, IN, USA




发件人: Pranith Kumar Karampuri 
收件人: yang.bi...@zte.com.cn, gluster-users@gluster.org,
日期: 2015/01/28 16:40
主题: Re: [Gluster-users] replica position
-




On 01/27/2015 02:43 PM, _yang.bi...@zte.com.cn_ 
<mailto:yang.bi...@zte.com.cn>wrote:


Hi,

Can the replica position be configured ?
Could you give an example?

Pranith

Regards.

Bin.Yang








ZTE Information Security Notice: The information contained in this mail 
(and any attachment transmitted herewith) is privileged and confidential 
and is intended for the exclusive use of the addressee(s).  If you are not 
an intended recipient, any disclosure, reproduction, distribution or other 
dissemination or use of the information contained is strictly prohibited. 
 If you have received this mail in error, please delete it and notify us 
immediately.






___
Gluster-users mailing list
_Gluster-users@gluster.org_ <mailto:Gluster-users@gluster.org>
_http://www.gluster.org/mailman/listinfo/gluster-users_




ZTE Information Security Notice: The information contained in this mail (and 
any attachment transmitted herewith) is privileged and confidential and is 
intended for the exclusive use of the addressee(s).  If you are not an intended 
recipient, any disclosure, reproduction, distribution or other dissemination or 
use of the information contained is strictly prohibited.  If you have received 
this mail in error, please delete it and notify us immediately.




___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


--
*Ted Miller*, Design Engineer
*SonSet Solutions*
(formerly HCJB Global Technology Center)
my desk +1 574.970.4272
receptionist +1 574.972.4252
http://sonsetsolutions.org

/Technology for abundant life!/
___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] ... i was able to produce a split brain...

2015-02-03 Thread Ted Miller


On 2/3/2015 12:23 PM, Joe Julian wrote:



That brought another thought to mind (have not had reason to try it):
How does gluster cope if you go behind its back and rename a "rejected" 
file?  For instance, in my example above, what if I go directly on the 
brick and rename the host-2 copy of the file to hair-pulling.txt-dud?  The 
ideal scenario would seem to be that if user does a heal it would treat 
the copy as new file, see no dupe for hair-pulling.txt, and create a new 
dupe on host-2.  Since hair-pulling.txt-dud is also a new file, a dupe 
would be created on host-1.  User could then access files, verify 
correctness, and then delete hair-pulling.txt-dud.
This should cause you to have two files with the same gfid. This will 
create the hardlink in .glusterfs again, and the heal will then re-create 
the .txt file also with that same gfid. Since both files will have the same 
gfid (stored in extended attributes) and be hard linked to the same file 
under .glusterfs you should then end up with both files being split-brain. 

Joe, I moved you comments up to be closest to the proposal they seem relevant 
to.


*
A not-officially-sanctioned way that I dealt with a split-brain a few 
versions back:

1. decided I wanted to keep file on host-2
2. log onto host-2
3. cp /brick/data1/hair-pulling.txt /gluster/data1/hair-pulling.txt-dud
4. rm /brick/data1/hair-pulling.txt
5. follow some Joe Julian blog stuff to delete the "invisible fork" of file
6. gluster volume heal data1 all
If you note, in the above scenario I copied from the brick to the mounted 
gluster volume.  I believe that this forces the breaking of any linkage 
between the old file and the new one.  Am I missing something there?


Ted Miller
Elkhart, IN

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] ... i was able to produce a split brain...

2015-02-03 Thread Ted Miller


On 1/31/2015 12:47 AM, Pranith Kumar Karampuri wrote:


On 01/30/2015 06:28 PM, Jeff Darcy wrote:

Pranith and I had a discussion regarding this issue and here is what we have
in our mind right now.

We plan to provide the user commands to execute from mount so that he can
access the files in split-brain. This way he can choose which copy is to be
used as source. The user will have to perform a set of getfattrs and
setfattrs (on virtual xattrs) to decide which child to choose as source and
inform AFR with his decision.

A) To know the split-brain status :
getfattr -n trusted.afr.split-brain-status 

This will provide user with the following details -
1) Whether the file is in metadata split-brain
2) Whether the file is in data split-brain

It will also list the name of afr-children to choose from. Something like :
Option0: client-0
Option1: client-1

We also tell the user what the user could do to view metadata/data info; 
like

stat to get metadata etc.

B) Now the user has to choose one of the options (client-x/client-y..) to
inspect the files.
e.g., setfattr -n trusted.afr.split-brain-choice -v client-0 
We save the read-child info in inode-ctx in order to provide the user access
to the file in split-brain from that child. Once the user inspects the file,
he proceeds to do the same from the other child of replica pair and makes an
informed decision.

C) Once the above steps are done, AFR is to be informed with the final 
choice

for source. This is achieved by -
(say the fresh copy is in client-0)
e.g., setfattr -n trusted.afr.split-brain-heal-finalize -v client-0

This child will be chosen as source and split-brain resolution will be done.
May I suggest another possible way to get around the difficulty in 
determining which of the files is the one to keep?


What if each of the files were to be renamed by appending the name of the 
brick-host that it lives on?

For example, in a replica 2 system:
brick-1: data1
host-1: host1
brick-2: date1
host-2: host2
file name: hair-pulling.txt

after running script/command to resolve split-brain, file system would have 
two files:

hair-pulling.txt__host-1_data1
hair-pulling.txt__host-2_data1

the user would then delete the unwanted file and rename the wanted file back 
to hair-pulling.txt.


The only problem would come with a very large file with a large number of 
replicas (like the replica 5 system I am working with). You might run out of 
space for all the copies.


Otherwise, this seems (to me) to present a user-friendly way to do this.  If 
the user has doubts (and disk space), user can choose to keep the rejected 
file around for a while, "just in case" it happens to have something useful 
in it that is missing from the "accepted" file.



That brought another thought to mind (have not had reason to try it):
How does gluster cope if you go behind its back and rename a "rejected" 
file?  For instance, in my example above, what if I go directly on the brick 
and rename the host-2 copy of the file to hair-pulling.txt-dud?  The ideal 
scenario would seem to be that if user does a heal it would treat the copy as 
new file, see no dupe for hair-pulling.txt, and create a new dupe on host-2.  
Since hair-pulling.txt-dud is also a new file, a dupe would be created on 
host-1.  User could then access files, verify correctness, and then delete 
hair-pulling.txt-dud.


*
A not-officially-sanctioned way that I dealt with a split-brain a few 
versions back:

1. decided I wanted to keep file on host-2
2. log onto host-2
3. cp /brick/data1/hair-pulling.txt /gluster/data1/hair-pulling.txt-dud
4. rm /brick/data1/hair-pulling.txt
5. follow some Joe Julian blog stuff to delete the "invisible fork" of file
6. gluster volume heal data1 all
I believe that this did work for me at that time.  I have not had to do it on 
a recent gluster version.


Ted Miller
Elkhart, IN

___
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] how to restrict client connection to server to only one IP address

2014-10-20 Thread Ted Miller


On 10/16/2014 2:48 PM, Łukasz Zygmański wrote:

Hello,

I am new to this list and new to GlusterFS, so I would be grateful if you 
could help me.


I am trying to do this setup:

client1(10.75.2.45)
   |
   |   MTU 1500
   V
(10.75.2.41)
  gluster1gluster2
(10.75.2.43)  --->  (10.75.2.44)
  <---
  MTU 9000

In words, I have two glusterfs servers (in replication): gluster1 and 
gluster2 and a glusterfs client client1.

The gluster1 has two network interfaces: 10.75.2.41 and 10.75.2.43.
I would like gluster1 to communicate with gluster2 using jumbo frames and 
connection would be between interfaces 10.75.2.43 and 10.75.2.44.
Since the client1 can only use default packet size (MTU 1500) I would like 
it to connect with gluster1 using only other network interface: 10.75.2.41.


Is it possible?

At the moment on gluster1 I have:

eno16780032: flags=4163 mtu 1500
inet 10.75.2.43  netmask 255.255.255.0  broadcast 10.75.2.255
eno33559296: flags=4163  mtu 1500
inet 10.75.2.41  netmask 255.255.255.0  broadcast 10.75.2.255

and when I mount from client1 using:
mount -t glusterfs 10.75.2.41:/vol1 /mnt/glusterfs

it still uses connection to 10.75.2.43:
# netstat -natup | egrep '(2.41|2.43)'
tcp0  0 10.75.2.45:1020 10.75.2.43:49152 ESTABLISHED 
10856/glusterfs
tcp0  0 10.75.2.45:1022 10.75.2.41:24007 ESTABLISHED 
10856/glusterfs


Is there a way to restrict communication from client1 to gluster1 using 
only one IP address: 10.75.2.41?


Any help would be much appreciated.

Best regards
Lukasz

PS
GlusterFS version on client:
glusterfs-3.5.2-1.el7.x86_64
glusterfs-fuse-3.5.2-1.el7.x86_64

GlusterFS version on server:
glusterfs-server-3.5.2-1.el7.x86_64
glusterfs-3.5.2-1.el7.x86_64
Since no one has answered this in a few days, I will try to do so, or at 
least start the process.


You do not mention how the client connects.

1. If it is using gluster-fuse, what you are trying to do is futile, because 
the connections are not as you think.  The data does not flow from client1 -> 
gluster1 -> gluster2.  The way it really works is that client1 connects 
directly to both gluster1 and gluster2, and sends the data to both of them at 
the same time.  The only time any volume of data transfers directly from 
gluster1 to gluster2 is during a heal operation.  Unfortunately, gluster does 
not understand the concept of a separate "storage network" that the servers 
use to talk to each other.  It only has one address, and that address is the 
one that the clients connect to.


2. If the client uses NFS, then you have something more like what you drew.  
The data passes client1 -> gluster1 via NFS, and then gluster1 -> gluster2.  
I am not using NFS, so I can't help you with if it is possible to have NFS on 
one network connection and gluster on a different connection, or what is 
required to accomplish this (if it can be done at all).


Ted Miller

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] replaceing bricks from a cluster with replication

2014-10-15 Thread Ted Miller


On 10/15/2014 3:46 PM, John G. Heim wrote:


You don't say, so I will assume you have 2 bricks per machine, resulting 
in 2 names returned when you execute a

 gluster volume list
 
command.  I will assume that your volumes are named 'Dick' and 'Jane' and 
one brick per machine.


Well, maybe I'm missing something but I have only one volume.  Was I 
supposed to do it differently?
You didn't miss anything at all :)  It all depends on what you are trying to 
store on gluster, and how that storage needs (or doesn't need) to be 
separated into different volumes.  If your data can all live happily together 
in one volume, that is just fine.  Not everyone has storage requirements that 
are that simple.  When they want to do something like this, they have to 
repeat for each separate volume.


In my explanation, you just can ignore step 4!

However, before you try my way, Joe Julian posted what looks like a simpler 
way to accomplish the same thing.  You still have to work brick by brick, but 
he does the add and remove a brick in one step instead of my two.


Also, he (very wisely) includes the warning about waiting until each brick is 
fully healed before going on to the next one.  Depending on the number and 
size of your files, the heal may take quite a while.  The heal may also 
interfere with normal access to the stored information, because it can 
completely tie up your network connection shoving data across it.


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] replaceing bricks from a cluster with replication

2014-10-14 Thread Ted Miller


On 10/9/2014 4:30 PM, John G. Heim wrote:
How can I replace machines in a gluster cluster with triple replication? 
I've had crashes and I've replaces single machines before but always with 
one with the same name. Now I need to take machine1, machine2, and machine3 
out of the cluster and replace it with machineA, machineB, and machineC.


Gluster will not let me detach machine1 all by itself. I have to detach
 machine1, machine2, and machine3 all at the same time. But is that going 
to cause data loss?


It seems to me that detaching all 3 machines in a brick on a triple 
replicated cluster would be like removing a single brick from a 
non-replicated cluster. And you can lose data in that case, right? 
I am just a user with a replica = 3 cluster running, but I will try to 
outline how I would to it.  If I goof something up (or explain it badly) I 
hope that others will jump in and correct me.


You don't say, so I will assume you have 2 bricks per machine, resulting in 2 
names returned when you execute a


gluster volume list

command.  I will assume that your volumes are named 'Dick' and 'Jane'.

The only way I know how to do what you want to do is a rather tedious 
process, but here it is.  I will not try to get all the commands word 
perfect, more like give you "pseudocode".


The basic premise is that you cannot add "machines", you can only add and 
subtract bricks from a volume.


1. Add machineA to the cluster as a peer.
2. Add brick "DickA" on machineA to volume "Dick", resulting in a "replica 4" 
volume.
3. Remove brick "Dick1" on machine1 from volume "Dick", putting you back at a 
"replica 3" volume.

4. Repeat steps 2 & 3 for "JaneA" and "Jane1".
5. At this point you no longer have any data on machine1, so remove machine1 
from the cluster.


Repeat steps 1-5, substituting machineB and machine2.

Repeat steps 1-5, substituting machineC and machine3.

At that point you should be migrated from machine1,2,3  to machineA,B,C.

Is that clear as mud?

It is probably more tedious than what you wanted to do (especially if you 
have 5 or more volumes), but it is the only way I know to do it.


Have fun,
Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Hosed installation

2014-10-09 Thread Ted Miller

On 10/7/2014 1:56 PM, Ryan Nix wrote:

Hello,

I seem to have hosed my installation while trying to replace a failed 
brick.  The instructions for replacing the brick with a different host 
name/IP on the Gluster site are no longer available so I used the 
instructions from the Redhat Storage class that I attended last week, which 
assumed the replacement had the same host name.


http://community.gluster.org/q/a-replica-node-has-failed-completely-and-must-be-replaced-with-new-empty-hardware-how-do-i-add-the-new-hardware-and-bricks-back-into-the-replica-pair-and-begin-the-healing-process/

It seems the working brick (I had two servers with simple replication only) 
will not release the DNS entry of the failed brick.


Is there any way to simply reset Gluster completely?
The simple way to "reset gluster completely" would be to delete the volume 
and start over.  Sometimes this is the quickest way, especially if you only 
have one or two volumes.


If nothing has changed, deleting the volume will not affect the data on the 
brick.


You can either:
Find and follow the instructions to delete the "markers" that glusterfs puts 
on the brick, in which case the create process should be the same as any new 
volume creation.
Otherwise, when you do the "volume create..." step, it will give you an 
error, something like 'brick already in use'.  You used to be able to 
override that by adding --force to the command line.  (Have not needed it 
lately, so don't know if it still works.)


Hope this helps
Ted Miller
Elkhart, IN


Just to confirm, if I delete the volume so I can start over, deleting the 
volume will not delete the data.  Is this correct?  Finally, once the 
volume is deleted, do I have to do what Joe Julian recommended here? 
http://joejulian.name/blog/glusterfs-path-or-a-prefix-of-it-is-already-part-of-a-volume/


Thanks for any insights.

- Ryan


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Gluster pool help

2014-09-10 Thread Ted Miller


On 9/9/2014 7:12 PM, Michael Bushey wrote:
I'm currently using GlusterFS 3.5.2 on a pair of production servers to 
share uploaded files and it's been reliable and working well with just the 
two servers. I've done some local tests of trying to add and remove servers 
and the results have not been good. I'm starting to think I have the wrong 
idea of what GlusterFS does and I'm trying to stick the square peg in the 
round hole. Is there some config to say "number of replicas = number of 
servers"?


My need is to have a copy of the data local on each server. If I have 5 
servers, I want five copies of the data. Basically each server should have 
it's own copy like it's a local ext4/zfs file-system, except it will 
immediately sync files added or removed on the other servers. I'm not sure 
the what the gluster definition of brick is, but I believe I want the 
number of bricks to equal the number of servers (at least for this 
particular file-system).


I've played around a bit and lost my data every time I've tried to add or 
remove a node. There is plenty of documentation, but it all says the same 
thing and doesn't answer the basics.


From `gluster volume info`: Number of Bricks: 1 x 2 = 2

I'm assuming the middle 2 is the number of bricks. I have no clue why we're 
multiplying by 1 to get itself.
http://gluster.org/community/documentation/index.php/Gluster_3.2:_Displaying_Volume_Information 
- does not show this magic multiplication. The page for 3.3 does not exist.


We're cloud based and we want to be able to add and remove servers on the 
fly. If for example I want to scale from 5 to 7 servers, it looks like I 
need to create a new storage pool with 7 replicas. Would it be the most 
logical to name my brick with the number of replicas? For example server1 
to server5 already have /var/gluster/files5. I could then create 
/var/gluster/files7 on server1 to server7, create the pool with 7 replicas, 
then pick a machine like server1 to copy the data from /var/gluster/files5 
to /var/gluster/files7, then destroy /var/gluster/files5. This seems 
extremely convoluted but it does not appear possible to expand a storage 
pool and expand the number of replicants with it.


Thanks in advance for help and your time to read this.
Michael
I have done this a few times, but it is not real recent, so I will try to 
explain in a way that is both correct and understandable.


1. As to the question about the "`gluster volume info`: Number of Bricks: 1 x 
2 = 2"
The first number is the number of bricks that the data is distributed 
across.  If that number is 3, each of three bricks has 1/3 of the data.  This 
is the "Distribute" mode of operating gluster.  From what you describe, you 
are not interested in that. You are interested in the "Replicate" mode.  The 
numbering was adopted because gluster makes it possible to use various 
combinations of Distribute and Replicate at the same time, and this numbering 
describes the arrangement very succinctly.


2. As to how to you add a brick to your volume, so you can add a server to 
your system.  You use the 'gluster volume add-brick' command.  You are 
telling gluster: You are currently a two-brick system and I want to make you 
a 3-brick system (each brick being on a different server) by adding 
'server3'.  Typing 'gluster volume add-brick help' returns


gluster volume add-brick  [ ]  ... 
[force]

Assuming that the brick is ready and freshly formatted with xfs, the command 
ends up looking something like:


gluster volume add-brick upload replica 3 server3:/bricks/upload/upload

This will create the gluster structure on the brick, and will start the 
process of replicating the gluster data onto the brick.  The gluster mount 
should be usable at that point, pulling the data from other bricks if it has 
not been replicated onto the local brick yet.


If this server has been in service before, it is probably best to reformat 
the brick before putting it into service.  Otherwise gluster will see 
remnants of the past use on the brick, and not want to use it, assuming there 
may be valuable information already stored there.


CAUTION: Unless they have fixed this issue, the replication process can bring 
a server to its knees until it is fully replicated.


3. To remove a server from your system, the process is basically reversed.  
If you want to go from 5 servers to 4, you issue the command:


gluster volume remove-brick upload replica 4 server5:/bricks/upload/upload

This tells gluster to drop the brick on server5 from the volume, and 
reassures gluster that you know that you will now have a replicated volume 
with 4 replicas on it.


Hope this helps.

Ted Miller
Elkhart, IN

P.S. I am not reading the list regularly, so if you need more info, best to 
copy me when sending to the list.  Just saw your request and said "I'v

Re: [Gluster-users] nfs

2014-05-28 Thread Ted Miller

You do not create anything on "one volume".  That will not replicate properly.

Mount the /gluster/ volume -- in a different place -- and create the volume 
on the /gluster/ volume.  It will then replicate to both bricks.


Line in /etc/fstab should look like:
primary:str  /mnt/glusterfs *glusterfs*default 0 0
or
primary:/str  /mnt/glusterfs *nfs* default0 0

When you mount the /gluster/ volume, it does not "replicate from one brick to 
another" it is created on both bricks at the same time".


If you mounted the volume correctly, you can just look at the bricks on each 
server, and you can see what is replicated.  The entire file structure will 
be there.


Ted Miller
Elkhart, IN, USA

P.S.  This is one of the most common mistakes made by newbies. Don't feel bad 
if you made it.  If you did not, congratulate yourself.



On 5/28/2014 5:13 AM, yalla.gnan.ku...@accenture.com wrote:


Hi All,

I got it. Its due to lack of hostnames in /etc/hosts.  Now it is working. J

I am facing another problem.  I have used replicated volumes in the two 
node gluster.  I have used these  replicated volume in openstack.


I have created file system on one volume.  But how to verify that the same 
content in one node is replicated on the gluster disk of another node ?


Thanks

Kumar

*From:*Gnan Kumar, Yalla
*Sent:* Wednesday, May 28, 2014 2:27 PM
*To:* gluster-users@gluster.org
*Subject:* nfs

Hi All,

I have a two node gluster storage.  The gluster version is 3.2.5.  But 
while mounting from a remote server as nfs volume, it is failing.


--

root@compute:~# showmount -e primary

Export list for primary:

/rpl *

/str *

/dst *

root@compute:~# mount -vvv -o mountproto=tcp -t nfs  primary:/str  
/mnt/glusterfs


mount: fstab path: "/etc/fstab"

mount: mtab path:  "/etc/mtab"

mount: lock path:  "/etc/mtab~"

mount: temp path:  "/etc/mtab.tmp"

mount: UID:0

mount: eUID:   0

mount: spec:  "primary:/str"

mount: node:  "/mnt/glusterfs"

mount: types: "nfs"

mount: opts:  "mountproto=tcp"

mount: external mount: argv[0] = "/sbin/mount.nfs"

mount: external mount: argv[1] = "primary:/str"

mount: external mount: argv[2] = "/mnt/glusterfs"

mount: external mount: argv[3] = "-v"

mount: external mount: argv[4] = "-o"

mount: external mount: argv[5] = "rw,mountproto=tcp"

mount.nfs: timeout set for Wed May 28 06:13:38 2014

mount.nfs: trying text-based options 
'mountproto=tcp,addr=10.211.203.66,mountaddr=10.211.203.66'


mount.nfs: prog 13, trying vers=3, prot=6

mount.nfs: trying 10.211.203.66 prog 13 vers 3 prot TCP port 38467

mount.nfs: prog 15, trying vers=3, prot=6

mount.nfs: trying 10.211.203.66 prog 15 vers 3 prot TCP port 38465

mount.nfs: mount(2): Permission denied

mount.nfs: access denied by server while mounting primary:/str

What is the reason ?

Thanks

Kumar


-

This message is for the designated recipient only and may contain 
privileged, proprietary, or otherwise confidential information. If you have 
received it in error, please notify the sender immediately and delete the 
original. Any other use of the e-mail by you is prohibited. Where allowed 
by local law, electronic communications with Accenture and its affiliates, 
including e-mail and instant messaging (including content), may be scanned 
by our systems for the purposes of information security and assessment of 
internal compliance with Accenture policy.

__

www.accenture.com


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


--
"He is no fool who gives what he cannot keep, to gain what he cannot lose." - - 
Jim Elliot
For more information about Jim Elliot and his unusual life, see 
http://www.christianliteratureandliving.com/march2003/carolyn.html.

Ted Miller
Design Engineer
HCJB Global Technology Center, a ministry of Reach Beyond
2830 South 17th St
Elkhart, IN  46517
574--970-4272 my desk
574--970-4252 receptionist

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] services needed for nfs [SOLVED]

2014-05-23 Thread Ted Miller


On 5/22/2014 12:13 PM, Ted Miller wrote:

I have a running replica 3 setup, running on Centos 6, fully updated.

I was trying to remove some unneeded services on a node (and did not 
document what I stopped).  That node is now showing that NFS is N/A on that 
node.


Since I already had one node showing NFS as N/A due to a "native" NFS share 
that was already in place, that leaves me with only one operation NFS server.


What services need to be running to make the gluster-nfs shares available?

What services should be turned off (or other changes made), in order to not 
have conflict between gluster-nfs and "native" NFS?


Ted Miller
Elkhart, IN, USA 
In my case, it was the nfslock service that was missing.  I suspect there are 
others needed, but nfslock is definitely needed.

Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Proposal for improvements for heal commands

2014-05-22 Thread Ted Miller

On 5/8/2014 7:12 AM, Pranith Kumar Karampuri wrote:

[snip]

3) "gluster volume heal  info split-brain" will be re-implemented to 
print all the files that are in split-brain instead of the limited 1024 entries.
- One constant complaint is that even after the file is fixed from 
split-brain, it may still show up in the previously cached output. In this 
implementation the goal is to remove all the caching and compute the results 
afresh.
Pranith, I missed your three-day deadline, because I was on 
vacation/holiday.  If not for this time, perhaps for the next iteration.


If at all possible, can we have the gfids in the split-brain output replaced 
by file names?  No matter how you go about fixing a split-brain, not knowing 
the file name when faced with a list of possibilities is an infuriating 
situation.  You can't tell log files from databases, so you can't even do triage.


Thanks for your work,
Ted Miller
Elkhart, IN

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] services needed for nfs

2014-05-22 Thread Ted Miller

I have a running replica 3 setup, running on Centos 6, fully updated.

I was trying to remove some unneeded services on a node (and did not document 
what I stopped).  That node is now showing that NFS is N/A on that node.


Since I already had one node showing NFS as N/A due to a "native" NFS share 
that was already in place, that leaves me with only one operation NFS server.


What services need to be running to make the gluster-nfs shares available?

What services should be turned off (or other changes made), in order to not 
have conflict between gluster-nfs and "native" NFS?


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Replace dead nodes/bricks

2014-05-21 Thread Ted Miller

A few things are not clear to me.  Comments inline below.

On 5/15/2014 5:19 AM, Lyle Plaatjes wrote:
I've just started at a new company and I came across this problem. They 
have webservers peering using gluster 3.2.7.


I take it that the gluster storage is on the same machines as the web-server 
software?


Was this a replica-4 setup, where there is a replicated brick on each of 4 
nodes?

The problem comes in where they upgraded the VM's to newer versions of 
Ubuntu. They didn't replace the bricks before decommissioning the other 
nodes. Only one node is actually still running so that is 1 brick that 
actually exists and is being replicated to the new hosts.


So there really is no true replication going on, just all files being served 
from the one gluster brick that still works?  (And if that brick dies, the 
whole site disappears until restored from backup {if there is a backup})


Now when one of the hosts is rebooted then gluster doesn't mount the volume 
because it's looking at the 3 dead peers and one that is still fine.


Are your new nodes (without bricks) peers in the gluster cluster?
Are your mounts of the form localhost: ?

What I need to do is replace the old dead peers with the new ones so that 
the gluster volume will actually mount if a host is rebooted.

Based on my guesses as to what your setup is, I think this is what I would do.

 * Get all web servers operating as peers in the trusted pool
 o It is not clear whether the new web servers even have gluster installed
 * change /etc/fstab so that mounts are of the form localhost:
 o so that it doesn't matter what other node is up or down, as long as
   the volume is active
 o I don't' know what exact commands Ubuntu is using, but in Centos 6 I
   use the "nofail" option in the fourth column of /etc/fstab (where
   'default' is a common entry).
 + This allows the bootup to proceed, even though the volume may not
   be mountable yet.
 # During (Centos) bootup, mount gets a second (or third) chance
   to mount things
 o make sure that the last two columns in /etc/fstab are "0 0"
 + so that it doesn't try to do a filesystem check during bootup
 * set up bricks on new servers
 o if the machine names are the same as the old machines, use a
   different path from the old brick path
 + to see old brick path, run "gluster volume info "
 o put brick in /etc/fstab, so it gets mounted
 * run "gluster volume add-brick  replica  :/" on each node
 o this adds the new bricks to the volume
 o you may need to wait until one brick has healed (synchronized) before
   adding the next brick
 + even synching one brick can saturate a network link, and bring
   things to their knees
 o repeat until you have 4 bricks active
 * run "gluster volume remove-brick  replica  :/" to remove "historic" bricks
 o don't do the remove bricks before adding any bricks
 + taking a replicated volume down to one brick and then trying to
   bring it back to two bricks can be problematic on some (maybe
   all) versions of gluster

Remember, I am assuming:

 * you have 4 web servers that should also all be gluster brick nodes
 * you were running a replica 4 configuration

Ted Miller
Elkhart, IN, USA

I am not an official part of gluster, just another user who has added and 
removed bricks a few times.


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Regarding Replicated Volume

2014-03-20 Thread Ted Miller

On 3/19/2014 3:06 PM, Cary Tsai wrote:

Hi There:
New to the GlusterFS and could not find the answer from the document
so hope I can get the answer form the mailing list.

Let's say we have two web servers:
One in Seattle, WA and another one is in Chapel Hill, NC.
So I create a 'replicated' volume which one brick in WA and another brick 
in NC.

I assume the web server in both WA and NC can mount the 'replicated' volume.
There are 2 HTTP/Get calls from CA and NY.
We assume CA's HTTP/Get is sent to web server in WA and
NY's HTTP/Get is sent to web server in NC.

My question is does the web server in WA definitely gets the data
from the brick in WA? If not, is any way to configure so the
web server in WA definitely gets data from the brick in WA?
The answer to your basic question is either "yes" or "they're working on 
it".  I know a while back it was on the "to do" list, but I am not sure if 
the patch is done, and if so, has it made it into production code.  But, the 
last I heard, yes, we were heading that direction.  Contrary to what someone 
else said, no, your scenario is not the only one where this is desirable.  In 
most any active situation using mostly-read files, reading from your local 
replicated disk is much faster, and also reduces network activity.


The big make-or-break questions as to whether this will work for you are:
* How much do you write? (more=more problem)
* Are these files sort-of/almost WORM files (Write Once, Read Many). WORM is 
better, RAM is worse

* Do both servers write?  (Only one is better)
* Do you modify files?  (More modification = more headaches)
* Do you replace/update files?  (Yes = more grief)

The critical issue is timing.  Gluster has various operations where it has to 
communicate with all nodes, and the process cannot move forward until all 
nodes answer.  Gluster is designed for all nodes to be connected by 1GB or 
faster networking, so your cross-continental link is outside the use-case the 
developers are using.  This always applies to writes, i.e. when a write 
occurs, it has to finish on both servers, probably with several commands 
issued, and each time it cannot go on to the next step until the distant 
server finishes.  There are certain read operations where gluster checks to 
make sure that things match between all servers. I hear reference to the 
"stat" call as being one that can be slow, but I can't say I fully understand 
what it does.  I think I understand that an 'ls' command does not include the 
'stat' call, but the 'ls -l' does include the 'stat' call, so a 'ls -l' 
command on a directory with hundreds or thousands of files can take MUCH 
longer than an 'ls' call to that same directory.


IF your web site is doing read-only access to your file system, and it is not 
triggering any calls that make gluster do a difference check between your two 
servers, it might work.


If
1. You do not require absolute real-time synchronization between the servers
AND
2. You can do all the writes on one of the two servers
then
you should probably look at Geo-replication.  Geo-replication is a one-way 
process, where all the changes happen on one end, and they are reflected on 
the other end.  It is designed to handle slower network links, and allows you 
to keep the two sites in close-to real-time synchronization.  How close to 
real time will depend on your server write load, and you would have to 
describe what you are doing and let some of the folks here give you their 
experience in similar situations.  At least you are within the intended 
use-case, so the developers will be receptive to any problems you have, and 
they may get fixed.


Another caution (based on painfully learned experience).  If you decide to 
try a regular (not Geo-Replicated) system, I advise that you store your data 
on a third machine somewhere, ESPECIALLY if both machines are updating files 
at the same time.  Otherwise, it seems that it is only a matter of time 
before you will be struggling with a split-brain situation.  When you face 
your first split-brain, you will wish you had never run into one.


Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Bring up a brick after disk failure

2014-03-19 Thread Ted Miller

On 3/19/2014 7:17 AM, teg...@renget.se wrote:

Hi,

One of my bricks suffered from complete raid failure, (3 disks on raid6). I 
created a new raid, and wanted to bring the brick back up in the volume. 
Did the following

1. Removed it with
gluster volume remove-brick gluster s1:/mnt/raid6 start
gluster volume remove-brick gluster s1:/mnt/raid6 commit

2. Detached with
gluster peer detach s1

3. Probed anew:
gluster peer probe s1

4. Tried to add with
gluster volume add-brick gluster s1:/mnt/raid6

But last one fails with
Stage failed on operation 'Volume Add brick'

I'm using 3.4.2.

Thanks,

/jon
WARNING: I'm not even sure if this is true, or if I changed something else, 
but I give you my experience to try.


I was having trouble adding an additional brick to my replica array, and I 
was getting the same (or a similar) message.  I then tried adding the brick 
from the command line on the machine where the brick is located.  All of the 
sudden I began to get meaningful error messages.  As far as I know, the only 
thing I changed was which computer I sshed into, but it made all the 
difference in the world. The error messages were quite specific, and easy to 
understand.


I am curious if others have experienced the same thing.

Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] missing dependency on glusterfs-??.rpm

2014-03-13 Thread Ted Miller
I actually copied it from an oVirt node, where they get it from. Content of 
the  /etc/yum.repos.d/glusterfs-epel.repo file is:


# Place this file in your /etc/yum.repos.d/ directory

[glusterfs-epel]

name=GlusterFS is a clustered file-system capable of scaling to several 
petabytes.

baseurl=http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-$releasever/$basearch/

enabled=1

skip_if_unavailable=1

gpgcheck=0

[glusterfs-noarch-epel]

name=GlusterFS is a clustered file-system capable of scaling to several 
petabytes.

baseurl=http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-$releasever/noarch

enabled=1

skip_if_unavailable=1

gpgcheck=0

[glusterfs-source-epel]

name=GlusterFS is a clustered file-system capable of scaling to several 
petabytes. - Source

baseurl=http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-$releasever/SRPMS

enabled=0

skip_if_unavailable=1

gpgcheck=0

One of the reasons I chose that file is because I was extending a replica 3 
volume on the ovirt nodes to a 4th (non-ovirt) node.


"yum list gluster*" now spits back

# yum list gluster*

Loaded plugins: fastestmirror, presto

Loading mirror speeds from cached hostfile

 * base: centos.mirror.nac.net

 * extras: mirror.cs.uwp.edu

 * updates: mirror.cs.vt.edu

Installed Packages

glusterfs.x86_64   3.4.2-1.el6   @glusterfs-epel

glusterfs-cli.x86_64   3.4.2-1.el6   @glusterfs-epel

glusterfs-fuse.x86_64  3.4.2-1.el6   @glusterfs-epel

glusterfs-libs.x86_64  3.4.2-1.el6   @glusterfs-epel

glusterfs-server.x86_643.4.2-1.el6   @glusterfs-epel
Available Packages

glusterfs-api.x86_64   3.4.2-1.el6   glusterfs-epel

glusterfs-api-devel.x86_64 3.4.2-1.el6   glusterfs-epel

glusterfs-debuginfo.x86_64 3.4.2-1.el6   glusterfs-epel

glusterfs-devel.x86_64 3.4.2-1.el6   glusterfs-epel
glusterfs-geo-replication.x86_64   3.4.2-1.el6   glusterfs-epel

glusterfs-rdma.x86_64  3.4.2-1.el6   glusterfs-epel

glusterfs-resource-agents.noarch   3.4.2-1.el6   glusterfs-noarch-epel



On 3/13/2014 7:44 PM, Carlos Capriotti wrote:

What link did you use to fetch your .repo file ?


On Thu, Mar 13, 2014 at 11:27 PM, Ted Miller <mailto:tmil...@hcjb.org>> wrote:


Sequence of events:

spun up new Centos 6.5 VM
yum -y update
yum install glusterfs
this installed gluster 3.4.0

find out there is no gluster-server package in CentOS repos.

added glusterfs-epel, glusterfs-noarch-epel

yum -y install gluster-server

service glusterd start   [*FAILED*]

yum list gluster*
note that all gluster* packages were updated to 3.4.2 /except/ yum
update glusterfs-libs

yum update glusterfs-libs

service glusterd start       [*OK *]

Ted Miller
Elkhart, IN


___
Gluster-users mailing list
Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
http://supercolony.gluster.org/mailman/listinfo/gluster-users




--
"He is no fool who gives what he cannot keep, to gain what he cannot lose." - - 
Jim Elliot
For more information about Jim Elliot and his unusual life, see 
http://www.christianliteratureandliving.com/march2003/carolyn.html.

Ted Miller
Design Engineer
HCJB Global Technology Center, a ministry of Reach Beyond
2830 South 17th St
Elkhart, IN  46517
574--970-4272 my desk
574--970-4252 receptionist

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] PLEASE READ ! We need your opinion. GSOC-2014 and the Gluster community

2014-03-13 Thread Ted Miller



[snip]

2) We have a recurring issue with split-brain solution. There is an entry 
on trello asking/suggesting a mechanism that arbitrates this resolution 
automatically. I pretty much think this could come together with another 
solution that is file replication consistency check.



Anything to improve split-brain resolution would get my vote.

Ted Miller
Elkhart, IN

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] missing dependency on glusterfs-??.rpm

2014-03-13 Thread Ted Miller

Sequence of events:

spun up new Centos 6.5 VM
yum -y update
yum install glusterfs
this installed gluster 3.4.0

find out there is no gluster-server package in CentOS repos.

added glusterfs-epel, glusterfs-noarch-epel

yum -y install gluster-server

service glusterd start   [*FAILED*]

yum list gluster*
note that all gluster* packages were updated to 3.4.2 /except/ yum update 
glusterfs-libs


yum update glusterfs-libs

service glusterd start   [*  OK *]

Ted Miller
Elkhart, IN

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] One node goes offline, the other node loses its connection to its local Gluster volume

2014-03-11 Thread Ted Miller


On 3/6/2014 7:48 PM, Greg Scott wrote:

In your real-life concern, the interconnect would not interfere with the 
existence of either
machines' ip address so after the ping-timeout, operations would resume in a 
split-brain
configuration. As long as no changes were made to the same file on both 
volumes, when the
connection is reestablished, the self-heal will do exactly what you expect.

Except that's not what happens.  If I ifdown that interconnect NIC, I should 
see the file system after 42 seconds, right?  But I don't.  Email butchers the 
output below, but what it shows is, I can look at my /firewall-scripts 
directory just fine when things are steady state.  I ifdown the interconnect 
NIC, that directory goes away.  I wait more than 2 minutes and it still doesn't 
come back.  And then when I ifup the NIC, everything goes back to normal after 
a few seconds.

[root@stylmark-fw1 ~]# ls /firewall-scripts
allow-all   etc  initial_rc.firewall  rcfirewall.conf   
 var
allow-all-with-nat  failover-monitor.sh  rc.firewall  
start-failover-monitor.sh
[root@stylmark-fw1 ~]# date
Thu Mar  6 18:39:42 CST 2014
[root@stylmark-fw1 ~]# ifdown enp5s4
[root@stylmark-fw1 ~]# ls /firewall-scripts
ls: cannot access /firewall-scripts: Transport endpoint is not connected
[root@stylmark-fw1 ~]# date
Thu Mar  6 18:41:50 CST 2014
[root@stylmark-fw1 ~]# ls /firewall-scripts
ls: cannot access /firewall-scripts: No such file or directory
[root@stylmark-fw1 ~]# ifup enp5s4
[root@stylmark-fw1 ~]# ls /firewall-scripts
ls: cannot access /firewall-scripts: No such file or directory
[root@stylmark-fw1 ~]# df -h
Filesystem   Size  Used Avail Use% Mounted on
/dev/mapper/fedora-root   17G  2.3G   14G  14% /
devtmpfs 989M 0  989M   0% /dev
tmpfs996M 0  996M   0% /dev/shm
tmpfs996M  524K  996M   1% /run
tmpfs996M 0  996M   0% /sys/fs/cgroup
tmpfs996M 0  996M   0% /tmp
/dev/sda2477M   87M  362M  20% /boot
/dev/sda1200M  9.6M  191M   5% /boot/efi
/dev/mapper/fedora-gluster--fw1  9.8G   33M  9.8G   1% /gluster-fw1
192.168.253.1:/firewall-scripts  9.8G   33M  9.8G   1% /firewall-scripts
[root@stylmark-fw1 ~]# ls /firewall-scripts
allow-all   etc  initial_rc.firewall  rcfirewall.conf   
 var
allow-all-with-nat  failover-monitor.sh  rc.firewall  
start-failover-monitor.sh
[root@stylmark-fw1 ~]#


You can avoid the split-brain using a couple of quorum techniques, the one that 
would seem to satisfy your
requirements leaving your volume read-only during the duration of the outage.

I like this idea - how do I do it?

I don't see a follow-up here, so I will put in (only) my two cents worth.

If I understand correctly, you get the read-only condition by using 
client-side quorum.  The behavior you describe above sounds like that 
produced by server-side quorum -- the volume goes offline until a quorum is 
present.


I have suffered through a couple of split-brain situations, and I agree that 
you do not want to run a two-node setup without quorum.


You may have gotten an answer that I did not see, but even so, I'll leave 
this here for the next guy who has a question.


Ted Miller
Elkhart, IN


- Greg
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] overzealous quota?

2014-01-29 Thread Ted Miller

After concluding that:
Replica 2 + quorum = Low Availability
Replica 2 - quorum = split brain (learned the hard way)

I switched to replica 3 for my oVirt 2 host setup.

Since I have only 2 hosts under the control of oVirt (which wants a LOT of 
control), I ended up with

Replica 3 + quorum
Brick 1 on host 1
Brick 2 on host 2
Brick 3 on host 2

Not ideal, but at least there are 3 bricks to vote, and host 1 should be able 
to go down and still have a quorum.


What I find is that the situation is the same as replica 2 + quorum -- both 
hosts must be up in order for the volume to be accessible.

Symptom
gluster volume info: reports volume is started.
gluster volume status: reports 2 bricks, but no ports allocated.
files on gluster mount inaccessible.

I realize that this is a corner case, but is there some reason I can't have 
my quorum all on one host?


My setup: oVirt on Centos 6.5, updated yesterday.

RFE #1: an easy, positive way to find out if a volume is available. Currently 
the only way I know is to look at "volume info" to see if it is started, then 
at "volume status" to see if ports are assigned.


RFE #2: A way to designate that a volume would become read-only on lack of 
quorum, rather than taking it offline.  I am sure I am not the only one with 
an almost-WORM use case, where having the volume go read-only would allow 
operations could continue as normal, just no new content could be added.  In 
my use case I will be serving media files in real time.  Under the current 
setup, I am considering mounting a brick as a read-only source, because that 
would allow operation to continue when quorum is lost.


Ted Miller
Elkhart, IN, USA
___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] could I disable updatedb in brick server?

2014-01-27 Thread Ted Miller


On 1/26/2014 6:13 AM, Mingfan Lu wrote:
we have lots of (really) files in our gluser brick servers and every day, 
we will generate lots, the number of files increase very quickly. could I 
disable updatedb in brick servers? if that, glusterfs servers will be impacted?
I don't have the docs open now, but there is a setup file where you can tell 
updatedb what to index and what to not index.  I altered my configuration to 
add some things that it ignores by default, but it should be just as simple 
to tell it not to index your brick(s).


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] intresting issue of replication and self-heal

2014-01-27 Thread Ted Miller


On 1/21/2014 11:05 PM, Mingfan Lu wrote:

I have a volume (distribute-replica (*3)), today i found an interesting problem

node22 node23 and node24 are the replica-7 from client A
but the annoying thing is when I create dir or write file from client to 
replica-7,


 date;dd if=/dev/zero of=49 bs=1MB count=120
Wed Jan 22 11:51:41 CST 2014
120+0 records in
120+0 records out
12000 bytes (120 MB) copied, 1.96257 s, 61.1 MB/s

but I could only find node23 & node24 have the find
---
node23,node24
---
/mnt/xfsd/test-volume/test/49

in clientA, I use find command

I use another machine as client B, and mount the test volume (newly mounted)
to run*find /mnt/xfsd/test-volume/test/49*

from Client A, the  three nodes have the file now.

---
node22,node23.node24
---
/mnt/xfsd/test-volume/test/49

but in Client A, I delete the file /mnt/xfsd/test-volume/test/49, node22 
still have the file in brick.


---
node22
---
/mnt/xfsd/test-volume/test/49

but if i delete the new created files from Client B )
my question is why node22 have no newly created/write dirs/files? I have to 
use find to trigger the self-heal to fix that?


from ClientA's log, I find something like:

 I [afr-self-heal-data.c:712:afr_sh_data_fix] 0-test-volume-replicate-7: no 
active sinks for performing self-heal on file /test/49


It is harmless for it is information level?

I also see something like:
[2014-01-19 10:23:48.422757] E 
[afr-self-heal-entry.c:2376:afr_sh_post_nonblocking_entry_cbk] 
0-test-volume-replicate-7: Non Blocking entrylks failed for  
/test/video/2014/01.
[2014-01-19 10:23:48.423042] E 
[afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 
0-test-volume-replicate-7: background  entry self-heal failed on 
/test/video/2014/01
From the paths you are listing, it looks like you may be mounting the 
bricks, not the gluster volume.


You MUST mount the gluster volume, not the bricks that make up the volume.  
In your example, the mount looks like it is mounting the xfs volume.  Your 
mount command should be something like:


   mount :test volume /mount/gluster/test-volume

If a brick is part of a gluster volume, the brick must NEVER be written to 
directly.  Yes, what you write MAY eventually be duplicated over to the other 
nodes, but if and when that happens is unpredictable.  It will give the 
unpredictable replication results that you are seeing.


The best way to test is to run "mount".  If the line where you are mounting 
the gluster volume doesn't say "glusterfs" on it, you have it wrong.  Also, 
the line you use in /etc/fstab must say "glusterfs", not "xfs" or "ext4".


If you are in doubt, include the output of "mount" in your next email to the 
list.


Ted Miller
Elkhart, IN, USA

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] engine-iso-uploader -- REST API not usable?

2014-01-17 Thread Ted Miller
I ran into this problem when I tried to use engine-iso-uploader, but reading on 
the lists makes it sound like it may be a more general problem.  There was a 
bug that caused this, but that was back in the ver. 3.0/3.1 days, and doesn't 
seem common since then.

Back on Dec 24 I was able to upload an ISO file OK, so I am not sure what has 
changed since then.

I am running a test setup, fully up to date:
office2a  host w/ glusterfs Centos 6
office4a  host w/ glusterfs Centos 6
ov-eng01 engine on Centos 6 VM (not hosted on oVirt)
office9  KVM host (not oVirt) for ov-eng01

whether I log in to ov-eng01 by ssh or execute the command from the console, I 
get:

# engine-iso-uploader list -v
Please provide the REST API password for the admin@internal oVirt Engine user 
(CTRL+D to abort):
ERROR: Problem connecting to the REST API.  Is the service available and does 
the CA certificate exist?

checking on some things suggested on a thread about engine-iso-uploader back in 
March, I get:

# ls -la /etc/pki/ovirt-engine/ca.pem
-rw-r--r--. 1 root root 4569 Nov 10 15:13 /etc/pki/ovirt-engine/ca.pem

# cat 
/var/log/ovirt-engine/ovirt-iso-uploader/ovirt-iso-uploader/20140117112938.log
2014-01-17 11:29:44::ERROR::engine-iso-uploader::512::root:: Problem connecting 
to the REST API.  Is the service available and does the CA certificate exist?

The thread back in March gave a work-around to upload ISO images directly, so I 
am not "blocked" from uploading images, but I would like to get things working 
"right", as I am afraid the problem will "turn around and bite me" down the 
road.

Ted Miller

___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] split-brain + what?

2014-01-13 Thread Ted Miller


From: Ted & Jean Miller 
Sent: Monday, January 13, 2014 11:29 PM
To: Ted Miller
Subject: split-brain + what?

I had a network failure yesterday, and besides some split-brain files, I am
getting the following in my files.  It appears that my two nodes won't stay
connected because they can't agree on something, but I don't understand
what they don't agree on, or what to do about it.

This log segment begins just after one of the nodes was rebooted.

+--+

[2014-01-14 01:20:59.410888] I [rpc-clnt.c:1676:rpc_clnt_reconfig] 
0-VM-client-0: changing port to 49154 (from 0)

[2014-01-14 01:20:59.411008] W [socket.c:514:__socket_rwv] 0-VM-client-0: readv 
failed (No data available)

[2014-01-14 01:20:59.418326] I 
[client-handshake.c:1659:select_server_supported_programs] 0-VM-client-0: Using 
Program GlusterFS 3.3, Num (1298437), Version (330)

[2014-01-14 01:20:59.418462] I [rpc-clnt.c:1676:rpc_clnt_reconfig] 
0-VM-client-1: changing port to 49156 (from 0)

[2014-01-14 01:20:59.418512] W [socket.c:514:__socket_rwv] 0-VM-client-1: readv 
failed (No data available)

[2014-01-14 01:20:59.422397] I [client-handshake.c:1456:client_setvolume_cbk] 
0-VM-client-0: Connected to 10.41.65.4:49154, attached to remote volume 
'/bricks/01/B'.

[2014-01-14 01:20:59.422420] I [client-handshake.c:1468:client_setvolume_cbk] 
0-VM-client-0: Server and Client lk-version numbers are not same, reopening the 
fds

[2014-01-14 01:20:59.422481] I [afr-common.c:3698:afr_notify] 0-VM-replicate-0: 
Subvolume 'VM-client-0' came back up; going online.

[2014-01-14 01:20:59.422617] I 
[client-handshake.c:450:client_set_lk_version_cbk] 0-VM-client-0: Server lk 
version = 1

[2014-01-14 01:20:59.422706] I 
[client-handshake.c:1659:select_server_supported_programs] 0-VM-client-1: Using 
Program GlusterFS 3.3, Num (1298437), Version (330)

[2014-01-14 01:20:59.422899] I [client-handshake.c:1456:client_setvolume_cbk] 
0-VM-client-1: Connected to 10.41.65.2:49156, attached to remote volume 
'/bricks/01/B'.

[2014-01-14 01:20:59.422909] I [client-handshake.c:1468:client_setvolume_cbk] 
0-VM-client-1: Server and Client lk-version numbers are not same, reopening the 
fds

[2014-01-14 01:20:59.429930] I [fuse-bridge.c:4769:fuse_graph_setup] 0-fuse: 
switched to graph 0

[2014-01-14 01:20:59.430077] I 
[client-handshake.c:450:client_set_lk_version_cbk] 0-VM-client-1: Server lk 
version = 1

[2014-01-14 01:20:59.431195] I [fuse-bridge.c:3724:fuse_init] 0-glusterfs-fuse: 
FUSE inited with protocol versions: glusterfs 7.13 kernel 7.13

[2014-01-14 01:20:59.432065] I 
[afr-common.c:2057:afr_set_root_inode_on_first_lookup] 0-VM-replicate-0: added 
root inode

[2014-01-14 01:20:59.432434] I [afr-common.c:2120:afr_discovery_cbk] 
0-VM-replicate-0: selecting local read_child VM-client-0

[2014-01-14 03:00:47.506676] W [socket.c:514:__socket_rwv] 0-glusterfs: readv 
failed (No data available)

[2014-01-14 03:00:47.506693] W [socket.c:1962:__socket_proto_state_machine] 
0-glusterfs: reading from socket failed. Error (No data available), peer 
(10.41.65.4:24007)

[2014-01-14 03:00:57.793050] I [glusterfsd-mgmt.c:1584:mgmt_getspec_cbk] 
0-glusterfs: No change in volfile, continuing

[2014-01-14 03:14:05.976531] W [socket.c:514:__socket_rwv] 0-VM-client-1: readv 
failed (Connection timed out)

[2014-01-14 03:14:05.976563] W [socket.c:1962:__socket_proto_state_machine] 
0-VM-client-1: reading from socket failed. Error (Connection timed out), peer 
(10.41.65.2:49156)

[2014-01-14 03:14:05.976597] I [client.c:2097:client_rpc_notify] 0-VM-client-1: 
disconnected

[2014-01-14 03:14:47.855141] E 
[client-handshake.c:1742:client_query_portmap_cbk] 0-VM-client-1: failed to get 
the port number for remote subvolume. Please run 'gluster volume status' on 
server to see if brick process is running.

[2014-01-14 03:14:47.855192] W [socket.c:514:__socket_rwv] 0-VM-client-1: readv 
failed (No data available)

[2014-01-14 03:14:47.855214] I [client.c:2097:client_rpc_notify] 0-VM-client-1: 
disconnected

[2014-01-14 03:14:49.860364] I [rpc-clnt.c:1676:rpc_clnt_reconfig] 
0-VM-client-1: changing port to 49156 (from 0)

[2014-01-14 03:14:49.860396] W [socket.c:514:__socket_rwv] 0-VM-client-1: readv 
failed (No data available)

[2014-01-14 03:14:49.864065] I 
[client-handshake.c:1659:select_server_supported_programs] 0-VM-client-1: Using 
Program GlusterFS 3.3, Num (1298437), Version (330)

[2014-01-14 03:14:49.864325] I [client-handshake.c:1456:client_setvolume_cbk] 
0-VM-client-1: Connected to 10.41.65.2:49156, attached to remote volume 
'/bricks/01/B'.

[2014-01-14 03:14:49.864349] I [client-handshake.c:1468:client_setvolume_cbk] 
0-VM-client-1: Server and Client lk-version numbers are not same, reopening the 
fds

[2014-01-14 03:14:49.864527] I 
[client-handshake.c:450:client_set_lk_version_

Re: [Gluster-users] standby-server

2013-08-19 Thread Ted Miller

-Ursprüngliche Nachricht-
Von: gluster-users-boun...@gluster.org
[mailto:gluster-users-boun...@gluster.org] Im Auftrag von Ted Miller
Gesendet: Freitag, 16. August 2013 18:29
An: gluster-users@gluster.org
Betreff: [Gluster-users] standby-server



I am looking at glusterfs for a HA application w/o local tech support (don't
ask, third-world country, techs are hard to find).



My current plan is to do a replica-4 + hot spare server.  Of the four in-use
bricks, two will be on "servers" and the other two will be on a "client"
machine and a "hot-backup" client machine.  No striping, all content on each
local machine, each machine using its own disk for all reading.



Part of my plan is to have a cold-spare server in the rack, not powered on.



(This server will also be a cold spare for another server).



I am wondering if this would be a viable way to set up this configuration:



Set up glusterfs as replica-5.



1. server1
2. server2
3. client
4. client-standby
5. server-spare



Initialize and set up glusterfs with all 5 bricks in the system (no file
content).
Install system at client site, and test with all 5 bricks in system.



Shut down spare server.



Once a month, power up spare server, run full heal, shut down.
Power up server-spare for any software updates.



If server1 or server2 dies (or needs maintenance), tell them to power up
server-spare, and let it heal.



It seems to me that this would be easier than setting up a replica-4 system
and then jumping through all the hoops to replace a server from scratch.



Comments, reactions, pot-shots welcome.
Ted Miller



Ted Miller
Design Engineer
HCJB Global Technology Center
Elkhart, IN  46517




On 8/19/2013 2:38 AM, Daniel Müller wrote:

What ist he reason about a "spare server" ?  With gluster just replicate to the 
real status all the time!?


This is a hot, tropical location with terrible power, a lot of lightning 
storms and only part-time air-conditioning.  Hardware failures are common.  I 
am doing what I can to protect against the power problems, but it is still an 
environment that is very hard on equipment.  This system will be 
business-critical--it is a radio station, and their music and other programs 
will be on the glusterfs, with downtime  beyond seconds becoming dead air.


If there is a catastrophic event, I want the spare server to have been off, 
and thus (most likely) not damaged.  It can be brought on-line to restore 
dual server status.



Or just do geo replication to a non gluster machine!?


I plan on doing geo-replication, but restoring a multi-terabyte file system 
over a 4Mb link will take days.  We would probably have to rsync the file 
system to portable drives and fly them over.


Hope that helps the perspective a bit.

Ted Miller

---
EDV Daniel Müller

Leitung EDV
Tropenklinik Paul-Lechler-Krankenhaus
Paul-Lechler-Str. 24
72076 Tübingen

Tel.: 07071/206-463, Fax: 07071/206-499
eMail: muel...@tropenklinik.de
Internet: www.tropenklinik.de
---


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Ted Miller
Design Engineer
HCJB Global Technology Center
Elkhart, IN  46517


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


[Gluster-users] standby-server

2013-08-16 Thread Ted Miller
I am looking at glusterfs for a HA application w/o local tech support (don't 
ask, third-world country, techs are hard to find).


My current plan is to do a replica-4 + hot spare server.  Of the four in-use 
bricks, two will be on "servers" and the other two will be on a "client" 
machine and a "hot-backup" client machine.  No striping, all content on each 
local machine, each machine using its own disk for all reading.


Part of my plan is to have a cold-spare server in the rack, not powered on.  
(This server will also be a cold spare for another server).


I am wondering if this would be a viable way to set up this configuration:

Set up glusterfs as replica-5.

1. server1
2. server2
3. client
4. client-standby
5. server-spare

Initialize and set up glusterfs with all 5 bricks in the system (no file 
content).

Install system at client site, and test with all 5 bricks in system.

Shut down spare server.

Once a month, power up spare server, run full heal, shut down.
Power up server-spare for any software updates.

If server1 or server2 dies (or needs maintenance), tell them to power up 
server-spare, and let it heal.


It seems to me that this would be easier than setting up a replica-4 system 
and then jumping through all the hoops to replace a server from scratch.


Comments, reactions, pot-shots welcome.
Ted Miller

--
"He is no fool who gives what he cannot keep, to gain what he cannot lose." - - 
Jim Elliot
For more information about Jim Elliot and his unusual life, see 
http://www.christianliteratureandliving.com/march2003/carolyn.html.

Ted Miller
Design Engineer
HCJB Global Technology Center
2830 South 17th St
Elkhart, IN  46517
574--970-4272 my desk
574--970-4252 receptionist


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] Some text do not display correctly in GlusterFS_Concepts page

2013-08-06 Thread Ted Miller

The HTML contains the following at that spot:
Error creating thumbnail: Unable to save thumbnail to 
destination

Doesn't look like something we are supposed to see.

Ted Miller

On 8/6/2013 2:22 AM, Asias He wrote:

Hello,

In the Translator section of this link, there are some words in the
background of the text.

http://www.gluster.org/community/documentation/index.php/GlusterFS_Concepts



--
"He is no fool who gives what he cannot keep, to gain what he cannot lose." - - 
Jim Elliot
For more information about Jim Elliot and his unusual life, see 
http://www.christianliteratureandliving.com/march2003/carolyn.html.

Ted Miller
Design Engineer
HCJB Global Technology Center
2830 South 17th St
Elkhart, IN  46517
574--970-4272 my desk
574--970-4252 receptionist


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users


Re: [Gluster-users] nfs test dd cp tar

2013-07-26 Thread Ted Miller


On 7/24/2013 4:10 PM, Heiko L. wrote:

Hallo


I tried glusterfs-3.3.2 dd cp tar.
- dd: ok
- cp: ok
- tar: fail (filesize=0)


Whats going wrong?
I am fairly new to digging this deep into Linux file systems, but I will put 
my understanding into writing.  I believe dd fails because glusterfs is 
neither a file nor a block device.  dd basically interacts with files.  It 
can also interact with disks and partitions because, under the /dev/ 
directory structure, entire file systems can be read and written to as though 
they were one big file.


Gluster is different.  It has no file system of its own--it borrows the file 
system from whatever directory you tell it to use for its brick.  When there 
are redundant copies of a brick, those bricks are not all arranged the same, 
if you dig down far enough.  In fact, if I understand things correctly, on 
one computer the files may be occupying an entire 500GB hard drive formatted 
with XFS.  On another computer, that same brick may occupy a /gluster 
subdirectory on a 2TB drive which contains the entire root file system on an 
ext4 drive.  The way those files are stored on those two drives will be 
different in many, many ways, but gluster is fine with that. However, that 
makes it impossible to use dd to duplicate a "drive" like you can with other 
file systems.


If you want to back up the file system, it will have to be by handling it as 
a bunch of files, not as a block device.  If you want to restore the files to 
glusterfs again, you will need to make sure that your backup includes ALL the 
extended attributes.  I don't know exactly what commands it takes to do that, 
but I do know that without the extended attributes glusterfs won't be able to 
figure out if those files are the same as ones already on the system.


One other approach: if you are using identical partitions on all bricks to 
hold the file system (ext4 or XFS) that glusterfs is using, you can do a dd 
backup of the ext4 or XFS file system, and all the glusterfs information will 
be there.  If you do everything right, you can restore that backup back to a 
hard drive, mount the hard drive in the brick's location, and glusterfs will 
see it (assuming glusterfs configuration has not changed) and glusterfs will 
welcome the new brick back into the volume group.


As I said, I am new to this, so I hope someone will correct any mistakes in 
my explanation.

Ted Miller



regards Heiko



details

- test10 cp
root@z1-gluster:/mnt/gv3/test# sleep 10;date; cp -p /etc/profile .
Wed Jul 24 19:05:14 UTC 2013
root@z1-gluster:/mnt/gv3/test#   sleep 10;date; cp -p /etc/profile test1
Wed Jul 24 19:05:24 UTC 2013
root@z1-gluster:/mnt/gv3/test#   sleep 10;date; cp -p test1 test2
Wed Jul 24 19:05:35 UTC 2013
root@z1-gluster:/mnt/gv3/test# ls -l
total 4
-rw-r--r--   1 root sys 1570 Jul 19 19:39 profile
-rw-r--r--   1 root sys 1570 Jul 19 19:39 test1
-rw-r--r--   1 root sys 1570 Jul 19 19:39 test2
drwxr-xr-x  14 root sys   14 Jul 20 11:05 var
root@z1-gluster:/mnt/gv3/test#

   -> ok

---

- test11 dd

testfile=testfile
testdir=.
sz=1
time dd if=/dev/urandom bs=1024k count=$sz of=$testdir/${testfile}_${sz} 2>&1 | 
grep -v rec

root@z1-gluster:/mnt/gv3/test# time dd if=/dev/urandom bs=1024k count=$sz 
of=$testdir/${testfile}_${sz} 2>&1 | grep -v rec

real0m1.083s
user0m0.002s
sys 0m0.276s
root@z1-gluster:/mnt/gv3/test# ls -l ${testfile}*
-rw-r--r--   1 root root 1048576 Jul 24 19:23 testfile_1
-rw-r--r--   1 root root 10485760 Jul 24 19:24 testfile_10

  -> ok

---
- test12 tar

root@z1-gluster:/mnt/gv3/test# src="profile test1 test2"
root@z1-gluster:/mnt/gv3/test# dst=test

profile
tar: profile: Cannot open: I/O error
test1
tar: test1: Cannot open: I/O error
test2
tar: test2: Cannot open: I/O error
tar: Exiting with failure status due to previous errors

root@z1-gluster:/mnt/gv3/test# ls -l $src $dst
-rw-r--r--   1 root sys 1570 Jul 19 19:39 profile
-rw-r--r--   1 root sys 1570 Jul 19 19:39 test1
-rw-r--r--   1 root sys 1570 Jul 19 19:39 test2

test:
total 0
-rw---   1 root root   0 Jul 24 19:55 profile
-rw---   1 root root   0 Jul 24 19:55 test1
-rw---   1 root root   0 Jul 24 19:55 test2

  -> fail





___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users



--
"He is no fool who gives what he cannot keep, to gain what he cannot lose." - - 
Jim Elliot
For more information about Jim Elliot and his unusual life, see 
http://www.christianliteratureandliving.com/march2003/carolyn.html.

Ted Miller
Design Engineer
HCJB Global Technology Center
2830 South 17th St

[Gluster-users] rogue file

2013-07-08 Thread Ted Miller
A multi-part question relating to a scenario where a rogue program/person 
writes a file into a brick through the underlying file system (e.g. XFS) 
without going through glusterfs?


1. What is gluster supposed to do when a non-gluster file (no gluster xattrs) 
appears on a gluster brick?


2. What does the current implementation actually do?

3. Does it matter whether it is or is not a replicated brick?

4. Does it matter whether it is a striped brick?


Thinking about these scenarios after the file has been written:

A. No one ever looks at the file
B. A self-heal runs
C. A directory read gets done on that sub-directory
D. Someone actually tries to read the file (like if it was called readme.txt 
or some other common name)


Don't have an example at hand, just wondering what would happen as I work on 
putting together a test bed that will be either replica 2 or replica 4, 
unstriped.


Ted Miller
Elkhart, IN, USA


___
Gluster-users mailing list
Gluster-users@gluster.org
http://supercolony.gluster.org/mailman/listinfo/gluster-users