On Mon, Aug 26, 2013 at 10:45 AM, Johannes Klarenbeek < johannes.klarenb...@rigo.nl> wrote:
> Hello ceph-users,**** > > ** ** > > I’m trying to set up a linux cluster but it takes me a little longer then > I hoped for. There are some things that I do not quite understand yet. > Hopefully some of you can help me out.**** > > ** ** > > **1) **When using ceph-deploy, a ceph.conf file is created in the > current directory and in the /etc/ceph directory. Which one is ceph-deploy > using and which one should I edit? > The usual workflow is that ceph-deploy will use the one in the current dir to overwrite the remote one (or the one in /etc/ceph/) but will warn if this is the case and error out specifying it needs the `--overwrite-conf` flag to continue. > **** > > ** ** > > **2) **I have 6 OSD running per machine. I disk zapped them with > ceph-deploy disk zap. I prepared/activated them with a separate journal on > an ssd card and they are all running.**** > > ** ** > > **a) **Ceph-deploy disk list doesn’t show me what file system is in > use (or ‘Linux Filesystem’ as it mentions, is a filesystem in its own > right). Neither does it show you what partition or path it is using for its > journal. > That ceph-deploy command calls `ceph-disk list` in the remote host which in turn will not (as far as I could see) tell you what exact file system is in use. > **** > > **b) **Running parted doesn’t show me what file system is in use > either (except of course that it is a ‘Linux Filesystem’)… I believe parted > should do the trick to show me these settings?? > How are you calling parted? with what flags? Usually something like: `sudo parted /dev/{device} print` would print some output. Can you show what you are getting back? > **** > > **c) **When a GPT partition is corrupt (missing msdos prelude for > example) ceph-deploy disk zap doesn’t work. But after 4 times repeating the > command, it works, showing up as ‘Linux Filesystem’. > So it didn't work 4 times and then it did? This does sound unexpected. > **** > > **d) **How can you set the file system that ceph-deploy disk zap is > formatting the ceph data disk with? I like to zap a disk with XFS for > example. > This is currently not supported but could be added as a feature. > **** > > ** ** > > **3) **Is there a way to set the data partition for ceph-mon with > ceph-deploy or should I do it manually in ceph.conf? How do I format that > partition (what file system should I use)**** > > ** ** > > **4) **When running ceph status the following message is what I get: > root@cephnode1:/root#ceph status > cluster:xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx > health: HEALTH_WARN 37 pgs degraded; 192 pgs stuck unclean > monmap e1: 1 mons at {cephnode1=172.16.1.2:6789/0}, election epoch 1, > quorum 0 cephnode1 > osdmap e38: 6 osds: 6 up, 6 in > pgmap v65: 192 pgs: 155 active+remapped, 37 active+degraded; 0 bytes > data, 213 MB used, 11172GB / 11172GB avail > mdsmap e1: 0/0/1 up**** > > ** ** > > **a) **How do I get rid of the HEALTH_WARN message? Can I run some > tool that initiates a repair?**** > > **b) **I did not put any data in it yet, but it already uses a > whopping 213 MB, why?**** > > ** ** > > **5) **Last but not least, my config file looks like this > root@cephnode1:/root#cat /etc/ceph/ceph.conf > [global]**** > > fsid = xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx > mon_initial_members = cephnode1 > mon_host = 172.16.1.2 > auth_supported = cephx**** > > osd_journal_size = 1024 > filestore_xattr_use_omap = true**** > > ** ** > > This is really strange, since the documentation states that the minimum > requirement for a config with my configuration should at least have the > [mon.1] and [osd.1] [osd.2] [osd.3] [osd.4] [osd.5] [osd.6] directives. I > have set up separate journaling for my OSD’s but it doesn’t show in my conf > file. Also the journaling partitions are 2GB big and not 1024MB (if that is > what it means then).**** > > ** ** > > ** ** > > ** ** > > I can really use your help, since I’m stuck for the moment.**** > > ** ** > > Regards,**** > > Johannes**** > > ** ** > > > > > > **** > > ** ** > > ** ** > > ** ** > > > __________ Informatie van ESET Endpoint Antivirus, versie van database > viruskenmerken 8730 (20130826) __________ > > Het bericht is gecontroleerd door ESET Endpoint Antivirus. > > http://www.eset.com > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com