[ceph-users] multiple active connections to a single LUN

2018-12-28 Thread Никитенко Виталий
hi, install iscsi gateway as write 
http://docs.ceph.com/docs/master/rbd/iscsi-overview/
kernel version 4.16.1
ceph v12.2.10 Luminous
early i use tgt iscsi target i can connect 3 esxi server to one LUN as a shared 
drive and if one esxi was dropped the machines were transferred to others from 
the same storage
Now when I try to connect one LAN to 3 servers, I get an error on the ceph 
server

Dec 29 07:24:29 ceph-node1 tcmu-runner: tcmu_notify_lock_lost: 209 rbd / 
main_pool.store: Async lock drop. Old state 1
Dec 29 07:24:30 ceph-node1 tcmu-runner: alua_implicit_transition: 566 rbd / 
main_pool.store: Starting lock acquisition operation.
Dec 29 07:24:30 ceph-node1 tcmu-runner: tcmu_rbd_lock: 757 rbd / 
main_pool.store: Acquired exclusive lock.
Dec 29 07:24:35 ceph-node1 tcmu-runner: alua_implicit_transition: 566 rbd / 
main_pool.store: Starting lock acquisition operation.

such a scheme is possible in this configuration?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Falls cluster then one node switch off

2016-05-26 Thread Никитенко Виталий
Hello!
>>mon_osd_down_out_subtree_limit = host
Thanks! This realy help me!!!

>>So again, not a full duplication of the data, but a significant amount.
If on the host who was left alone, will step down at this time one OSD. ALL 
data will available? Or part of data, pgs which are marked as 'active + 
undersized + degraded' may be lost?

26.05.2016, 12:33, "Christian Balzer" <ch...@gol.com>:
> Hello,
>
> I've expanded the cache-tier in my test cluster from a single node
> to 2, increased the pool size from 1 to 2, then waited until all the data
> was rebalanced/duplicated and the cluster was healthy again.
>
> Then I stopped all OSDs on one of the 2 nodes and nothing other than
> degraded/undersized PGs happened (even after 10 minutes):
> ---
> pgmap v1529705: 1920 pgs: 1152 active+clean, 768 active+undersized+degraded; 
> 214 GB data, 695 GB used, 4932 GB / 5665 GB avail; 26059/397156 objects 
> degraded (6.561%)
> ---
>
> However I do have this setting in my ceph.conf:
> ---
> mon_osd_down_out_subtree_limit = host
> ---
>
> Which prevents the OSDs to be set out when host goes down.
> The idea here of course is to prevent data movement since it's likely
> faster to get the host back up than all the data moved.
>
> So the 4 OSDs on the "dead" node where still in and had a re-weight of 1.
>
> So next I did set them out and indeed I saw the same thing as you did:
> ---
> pgmap v1529737: 1920 pgs: 75 peering, 44 
> active+recovery_wait+degraded+remapped, 1152 active+clean, 2 
> active+recovering+degraded+remapped, 8 active+remapped, 608 
> active+undersized+degraded, 31 
> active+undersized+degraded+remapped+wait_backfill; 214 GB data, 672 GB used, 
> 3279 GB / 3989 GB avail; 23132/392266 objects degraded (5.897%); 2657/392266 
> objects misplaced (0.677%)
> ---
>
> However that stopped pretty quickly and the data usage on the still
> up/in OSDs only increased about 15%.
> Definitely NOT duplicating all the PGs/objects.
>
> So what I think is happening here (confirmation by the devs appreciated)
> is that due to the OSDs being OUT CRUSH goes and does its thing, thus
> moving some PGs around.
> So if PG 1.1 was primary on osd.0 and now the CRUSH algorithm thinks it
> should be primary on osd.0 what happens is that the data for that PG gets
> copied from osd.0 to osd.1, using up more space on osd.1 and NOT freeing up
> the space on osd.0.
>
> So again, not a full duplication of the data, but a significant amount.
>
> Thus my advice for you (and anybody reading this) would be:
> a) set "mon_osd_down_out_subtree_limit = host" (and monitor your systems
> of course).
> b) keep your space usage on dual node clusters/pools below 70%.
>
> Regards,
>
> Christian
>
> On Wed, 25 May 2016 14:51:24 +0900 Christian Balzer wrote:
>
>>  Hello,
>>
>>  Thanks for the update and I totally agree that it should try to do 2x
>>  replication on the single storage node.
>>
>>  I'll try to reproduce what you're seeing tomorrow on my test cluster,
>>  need to move some data around first.
>>
>>  Christian
>>
>>  On Wed, 25 May 2016 08:58:54 +0700 Никитенко Виталий wrote:
>>
>>  > I'm sorry it was not right map, that map right
>>  >
>>  > # begin crush map
>>  > tunable choose_local_tries 0
>>  > tunable choose_local_fallback_tries 0
>>  > tunable choose_total_tries 50
>>  > tunable chooseleaf_descend_once 1
>>  > tunable straw_calc_version 1
>>  >
>>  > # devices
>>  > device 0 osd.0
>>  > device 1 osd.1
>>  > device 2 osd.2
>>  > device 3 osd.3
>>  > device 4 osd.4
>>  > device 5 osd.5
>>  >
>>  > # types
>>  > type 0 osd
>>  > type 1 host
>>  > type 2 chassis
>>  > type 3 rack
>>  > type 4 row
>>  > type 5 pdu
>>  > type 6 pod
>>  > type 7 room
>>  > type 8 datacenter
>>  > type 9 region
>>  > type 10 root
>>  >
>>  > # buckets
>>  > host ceph1-node {
>>  > id -2 # do not change unnecessarily
>>  > # weight 0.030
>>  > alg straw
>>  > hash 0 # rjenkins1
>>  > item osd.0 weight 0.010
>>  > item osd.1 weight 0.010
>>  > item osd.2 weight 0.010
>>  > }
>>  > host ceph2-node {
>>  > id -3 # do not change unnecessarily
>>  > # weight 0.030
>>  > alg straw
>>  > hash 0 # rjenkins1
>>  > item osd.3 weight 0.010
>>  > item osd.4 weight 0.010
>>  > item osd.5 weight 0.010
>>  > }
>>  > root defau

Re: [ceph-users] Falls cluster then one node switch off

2016-05-24 Thread Никитенко Виталий
1 pgs=0 cs=0 l=1 
c=0x7fe7bc05dfe0).fault
min_size: 1

 root@ceph1-node:~# ceph osd dump  
pool 1 'hdd' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins 
pg_num 512 pgp_num 512 last_change 53 flags hashpspool stripe_width 0


after remapped end

root@ceph1-node:~# ceph -s
2016-05-24 17:23:10.123542 7f2c001cf700  0 -- :/623268863 >> 
192.168.241.12:6789/0 pipe(0x7f2bfc059cd0 sd=3 :0 s=1 pgs=0 cs=0 l=1 
c=0x7f2bfc05dfc0).fault
cluster 808ee682-c121-4867-9fe4-a347d95bf3f0
 health HEALTH_WARN
408 pgs degraded
262 pgs stuck degraded
640 pgs stuck unclean
262 pgs stuck undersized
408 pgs undersized
recovery 315/1098 objects degraded (28.689%)
recovery 234/1098 objects misplaced (21.311%)
1 mons down, quorum 0,2 ceph1-node,ceph-mon2
 monmap e1: 3 mons at 
{ceph-mon2=192.168.241.20:6789/0,ceph1-node=192.168.241.2:6789/0,ceph2-node=192.168.241.12:6789/0}
election epoch 18, quorum 0,2 ceph1-node,ceph-mon2
 osdmap e63: 6 osds: 3 up, 3 in; 232 remapped pgs
  pgmap v209: 640 pgs, 2 pools, 2118 MB data, 549 objects
3149 MB used, 27537 MB / 30686 MB avail
315/1098 objects degraded (28.689%)
234/1098 objects misplaced (21.311%)
 408 active+undersized+degraded
 232 active+remapped


Any idea why ceph making redundancies on local disks host?



24.05.2016, 12:53, "Christian Balzer" <ch...@gol.com>:
> Hello,
>
> On Tue, 24 May 2016 10:28:02 +0700 Никитенко Виталий wrote:
>
>>  Hello!
>>  I have a cluster of 2 nodes with 3 OSD each. The cluster full about 80%.
>
> According to your CRUSH map that's not quite true, namely ceph1-node2
> entry.
>
> And while that again according to your CRUSH map isn't in the default root
> I wonder WHERE it is and if it confuses Ceph into believing that there is
> actually a third node?
>
> "ceph osd tree" output may help, as well as removing ceph1-node2 from the
> picture.
>
>>  df -H
>>  /dev/sdc1 27G 24G 3.9G 86% /var/lib/ceph/osd/ceph-1
>>  /dev/sdd1 27G 20G 6.9G 75% /var/lib/ceph/osd/ceph-2
>>  /dev/sdb1 27G 24G 3.5G 88% /var/lib/ceph/osd/ceph-0
>>
>>  When I switch off one server, then after 10 minutes begins remapped pgs
>
> [snip]
>>  As a result, one disk overflow and the cluster falls. Why ceph remapped
>>  pgs, it was supposed to simply mark all pgs as active+degraded, while
>>  second node down?
>
> Yes, I agree, that shouldn't happen with a properly configured 2 node
> cluster.
>
>>  ceph version 0.80.11
>
> Not aware of any bugs in there and in fact I did test a 2 node cluster
> with Firefly, but be aware that this version is EoL and no longer receiving
> updates.
>
>>  root@ceph1-node:~# cat /etc/ceph/ceph.conf
>>  [global]
>>  fsid = b66c7daa-d6d8-46c7-9e61-15adbb749ed7
>>  mon_initial_members = ceph1-node, ceph2-node, ceph-mon2
>>  mon_host = 192.168.241.97,192.168.241.110,192.168.241.123
>>  auth_cluster_required = cephx
>>  auth_service_required = cephx
>>  auth_client_required = cephx
>>  filestore_xattr_use_omap = true
>>  osd_pool_default_size = 2
>>  osd_pool_default_min_size = 1
>
> Have you verified (ceph osd get  size / min_size) that all your
> pools are actually set like this?
>
> Regards,
>
> Christian
>>  mon_clock_drift_allowed = 2
>>
>>  root@ceph1-node:~#cat crush-map.txt
>>  # begin crush
>>  map tunable choose_local_tries
>>  0 tunable choose_local_fallback_tries
>>  0 tunable choose_total_tries
>>  50 tunable chooseleaf_descend_once
>>  1 tunable straw_calc_version
>>  1
>>  #
>>  devices device 0
>>  osd.0 device 1
>>  osd.1 device 2
>>  osd.2 device 3
>>  osd.3 device 4
>>  osd.4 device 5
>>  osd.5
>>  #
>>  types type 0
>>  osd type 1
>>  host type 2
>>  chassis type 3
>>  rack type 4
>>  row type 5
>>  pdu type 6
>>  pod type 7
>>  room type 8
>>  datacenter type 9
>>  region type 10
>>  root
>>  #
>>  buckets host ceph1-node
>>  { id -2 # do not change
>>  unnecessarily # weight
>>  0.060 alg
>>  straw hash 0 #
>>  rjenkins1 item osd.0 weight
>>  0.020 item osd.1 weight
>>  0.020 item osd.2 weight
>>  0.020 }
>>  host ceph2-node
>>  { id -3 # do not change unnecessarily
>>  # weight 0.060
>>  alg straw
>>  hash 0 # rjenkins1
>>  item osd.3 weight 0.020
>>  item osd.4 weight 0.020
>>  item osd.5 weight 0.020
>>  }
>>  root

[ceph-users] Falls cluster then one node switch off

2016-05-23 Thread Никитенко Виталий
Hello!
I have a cluster of 2 nodes with 3 OSD each. The cluster full about 80%.

df -H
/dev/sdc127G   24G  3.9G  86% /var/lib/ceph/osd/ceph-1
/dev/sdd127G   20G  6.9G  75% /var/lib/ceph/osd/ceph-2
/dev/sdb127G   24G  3.5G  88% /var/lib/ceph/osd/ceph-0

When I switch off one server, then after 10 minutes begins remapped pgs

ceph -w
2016-05-24 09:37:00.065134 mon.0 [INF] pgmap v4850: 704 pgs: 469 
active+degraded, 186 active+remapped, 26 active+recovery_wait+remapped, 23 
active+recovering+remapped; 56038 MB data, 65904 MB used, 10854 MB / 76759 MB 
avail; 11787/28038 objects degraded (42.039%); 6438 kB/s, 1 objects/s recovering
2016-05-24 09:37:03.889165 mon.0 [INF] pgmap v4851: 704 pgs: 469 
active+degraded, 186 active+remapped, 26 active+recovery_wait+remapped, 23 
active+recovering+remapped; 56038 MB data, 65928 MB used, 10830 MB / 76759 MB 
avail; 11786/28038 objects degraded (42.036%); 5753 kB/s, 1 objects/s recovering
2016-05-24 09:37:08.027062 osd.0 [WRN] OSD near full (90%)
2016-05-24 09:37:09.713479 mon.0 [INF] pgmap v4853: 704 pgs: 469 
active+degraded, 186 active+remapped, 26 active+recovery_wait+remapped, 23 
active+recovering+remapped; 56038 MB data, 65556 MB used, 11203 MB / 76759 MB 
avail; 11776/28038 objects degraded (42.000%); 7158 kB/s, 1 objects/s recovering
2016-05-24 09:37:16.232069 mon.0 [INF] pgmap v4854: 704 pgs: 469 
active+degraded, 186 active+remapped, 26 active+recovery_wait+remapped, 23 
active+recovering+remapped; 56038 MB data, 65626 MB used, 11133 MB / 76759 MB 
avail; 11773/28038 objects degraded (41.989%); 5273 kB/s, 1 objects/s recovering

As a result, one disk overflow and the cluster falls. Why ceph remapped pgs, it 
was supposed to simply mark all pgs as active+degraded, while second node down?

ceph version 0.80.11

root@ceph1-node:~# cat /etc/ceph/ceph.conf 
[global]
fsid = b66c7daa-d6d8-46c7-9e61-15adbb749ed7
mon_initial_members = ceph1-node, ceph2-node, ceph-mon2
mon_host = 192.168.241.97,192.168.241.110,192.168.241.123
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
osd_pool_default_size = 2
osd_pool_default_min_size = 1
mon_clock_drift_allowed = 2


root@ceph1-node:~#cat crush-map.txt 
# begin crush map   


   
tunable choose_local_tries 0


   
tunable choose_local_fallback_tries 0   


   
tunable choose_total_tries 50   


   
tunable chooseleaf_descend_once 1   


   
tunable straw_calc_version 1


   



   
# devices   


   
device 0 osd.0  


   
device 1 osd.1  



[ceph-users] active+clean+scrubbing+deep

2015-06-02 Thread Никитенко Виталий
Hi!

I have ceph version 0.94.1.

root@ceph-node1:~# ceph -s
cluster 3e0d58cd-d441-4d44-b49b-6cff08c20abf
 health HEALTH_OK
 monmap e2: 3 mons at 
{ceph-mon=10.10.100.3:6789/0,ceph-node1=10.10.100.1:6789/0,ceph-node2=10.10.100.2:6789/0}
election epoch 428, quorum 0,1,2 ceph-node1,ceph-node2,ceph-mon
 osdmap e978: 16 osds: 16 up, 16 in
  pgmap v6735569: 2012 pgs, 8 pools, 2801 GB data, 703 kobjects
5617 GB used, 33399 GB / 39016 GB avail
2011 active+clean
   1 active+clean+scrubbing+deep
  client io 174 kB/s rd, 30641 kB/s wr, 80 op/s
  
root@ceph-node1:~# ceph pg dump  | grep -i deep | cut -f 1
  dumped all in format plain
  pg_stat
  19.b3  
  
In log file i see 
2015-05-14 03:23:51.556876 7fc708a37700  0 log_channel(cluster) log [INF] : 
19.b3 deep-scrub starts
but no 19.b3 deep-scrub ok

then i do ceph pg deep-scrub 19.b3, nothing happens and in logs file no any 
records about it.

What can i do to pg return in active + clean station?
is there any sense restart OSD or the entirely server where the OSD?

Thanks.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] active+clean+scrubbing+deep

2015-06-02 Thread Никитенко Виталий
Thank Irek, it really worked 02.06.2015, 15:58, "Irek Fasikhov" malm...@gmail.com:Hi. Restart the OSD. :)2015-06-02 11:55 GMT+03:00 Никитенко Виталий v1...@yandex.ru:Hi!  I have ceph version 0.94.1.  root@ceph-node1:~# ceph -s     cluster 3e0d58cd-d441-4d44-b49b-6cff08c20abf      health HEALTH_OK      monmap e2: 3 mons at {ceph-mon=10.10.100.3:6789/0,ceph-node1=10.10.100.1:6789/0,ceph-node2=10.10.100.2:6789/0}             election epoch 428, quorum 0,1,2 ceph-node1,ceph-node2,ceph-mon      osdmap e978: 16 osds: 16 up, 16 in       pgmap v6735569: 2012 pgs, 8 pools, 2801 GB data, 703 kobjects             5617 GB used, 33399 GB / 39016 GB avail                 2011 active+clean                    1 active+clean+scrubbing+deep   client io 174 kB/s rd, 30641 kB/s wr, 80 op/s  root@ceph-node1:~# ceph pg dump  | grep -i deep | cut -f 1   dumped all in format plain   pg_stat   19.b3  In log file i see 2015-05-14 03:23:51.556876 7fc708a37700  0 log_channel(cluster) log [INF] : 19.b3 deep-scrub starts but no "19.b3 deep-scrub ok"  then i do "ceph pg deep-scrub 19.b3", nothing happens and in logs file no any records about it.  What can i do to pg return in "active + clean" station? is there any sense restart OSD or the entirely server where the OSD?  Thanks. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- С уважением, Фасихов Ирек НургаязовичМоб.: +79229045757___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] rbd loaded 100%

2015-01-21 Thread Никитенко Виталий
Hi!
I have a server (ceph version 0.80.7, links 10Gb), there is set: 1 pool is 
write to 5 osd. I'm using the iscsi-target write to this pool (disk rbd3) some 
data from other server. And speed on network is near 150 Mbit / sec. In this 
case, iostat shows the usage rbd3 drive 100%, but drives on which there are 5 
osd (sdc sdd sde sdf sdg) loaded in the region of 20% each. Who knows why this 
could be and what i can run the utility for the diagnosis?

iostat -x 1

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
   1.800.001.460.710.00   96.03

Device: rrqm/s   wrqm/s r/s w/srkB/swkB/s avgrq-sz 
avgqu-sz   await r_await w_await  svctm  %util
sda   0.00 0.000.000.00 0.00 0.00 0.00 
0.000.000.000.00   0.00   0.00
sdb   0.00 9.000.006.00 0.0068.0022.67 
0.000.670.000.67   0.67   0.40
sdc   0.00 2.000.00   33.00 0.00  7756.00   470.06 
2.76   83.760.00   83.76   5.45  18.00
sdd   0.00 0.000.00   59.00 0.00  9236.00   313.08 
0.579.690.009.69   6.58  38.80
sde   0.00 0.000.00   29.00 0.00  5112.00   352.55 
0.43   13.930.00   13.93   7.03  20.40
sdf   0.00 0.000.00   28.00 0.00  4612.00   329.43 
0.269.140.009.14   6.57  18.40
sdg   0.00 0.000.00   24.00 0.00  4032.00   336.00 
0.228.670.008.67   6.67  16.00
rbd0  0.00 0.000.000.00 0.00 0.00 0.00 
0.000.000.000.00   0.00   0.00
rbd1  0.00 0.000.000.00 0.00 0.00 0.00 
0.000.000.000.00   0.00   0.00
rbd2  0.00 0.000.000.00 0.00 0.00 0.00 
0.000.000.000.00   0.00   0.00
rbd3  0.00 0.000.00  318.00 0.00 20045.00   126.07 
7.28   28.290.00   28.29   3.13  99.60
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] iscsi and cache pool

2014-07-01 Thread Никитенко Виталий
Good day!
I have server with Ubunu 14.04 and installed ceph firefly. Configured main_pool 
(2 osd) and ssd_pool (1 ssd osd). I want use ssd_pool as cache pool for 
main_pool
  
  ceph osd tier add main_pool ssd_pool
  ceph osd tier cache-mode ssd_pool writeback
  ceph osd tier set-overlay main_pool ssd_pool
  
  ceph osd pool set ssd_pool hit_set_type bloom
  ceph osd pool set ssd_pool hit_set_count 1
  ceph osd pool set ssd_pool hit_set_period 600
  ceph osd pool set ssd_pool target_max_bytes 1000

 If use tgt as:
 tgtadm --lld iscsi --mode logicalunit --op new --tid 1 --lun 1 --bstype rbd 
--backing-store main_pool/store_main --bsopts conf=/etc/ceph/ceph.conf
 and then connected from iscsi initiator to this Lun1, i see that ssd_pool is 
used as cache (i see through iostat -x 1) but slow speed
 
 If use tgt as (or other sush as scst, iscsitarget):
 tgtadm --lld iscsi --mode logicalunit --op new --tid 1 --lun 1 -b /dev/rbd1 
(where rbd1=main_pool/store_main)
 and then connected from iscsi initiator to this Lun1, i see that ssd_pool is 
not used, that write through to 2 osd
 
 Help me, anyone work this iscsi and cache pool?
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] iscsi and cache pool

2014-07-01 Thread Никитенко Виталий
Thank you. try to do it

02.07.2014, 05:30, Gregory Farnum g...@inktank.com:
 Yeah, the features are new from January or something so you need a
 very new kernel to support it. There are no options to set.
 But in general I wouldn't use krbd if you can use librbd instead; it's
 easier to update and more featureful!
 -Greg
 Software Engineer #42 @ http://inktank.com | http://ceph.com

 On Tue, Jul 1, 2014 at 5:44 PM, Никитенко Виталий v1...@yandex.ru wrote:
  Hi!

  There is some option in the kernel which must be included, or just upgrade
  to the latest version of the kernel? I use 3.13.0-24

  Thanks

  01.07.2014, 20:17, Gregory Farnum g...@inktank.com:

  It looks like you're using a kernel RBD mount in the second case? I imagine
  your kernel doesn't support caching pools and you'd need to upgrade for it
  to work.
  -Greg

  On Tuesday, July 1, 2014, Никитенко Виталий v1...@yandex.ru wrote:

  Good day!
  I have server with Ubunu 14.04 and installed ceph firefly. Configured
  main_pool (2 osd) and ssd_pool (1 ssd osd). I want use ssd_pool as cache
  pool for main_pool

    ceph osd tier add main_pool ssd_pool
    ceph osd tier cache-mode ssd_pool writeback
    ceph osd tier set-overlay main_pool ssd_pool

    ceph osd pool set ssd_pool hit_set_type bloom
    ceph osd pool set ssd_pool hit_set_count 1
    ceph osd pool set ssd_pool hit_set_period 600
    ceph osd pool set ssd_pool target_max_bytes 1000

   If use tgt as:
   tgtadm --lld iscsi --mode logicalunit --op new --tid 1 --lun 1 --bstype rbd
  --backing-store main_pool/store_main --bsopts conf=/etc/ceph/ceph.conf
   and then connected from iscsi initiator to this Lun1, i see that ssd_pool
  is used as cache (i see through iostat -x 1) but slow speed

   If use tgt as (or other sush as scst, iscsitarget):
   tgtadm --lld iscsi --mode logicalunit --op new --tid 1 --lun 1 -b /dev/rbd1
  (where rbd1=main_pool/store_main)
   and then connected from iscsi initiator to this Lun1, i see that ssd_pool
  is not used, that write through to 2 osd

   Help me, anyone work this iscsi and cache pool?
  ___
  ceph-users mailing list
  ceph-users@lists.ceph.com
  http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

  --
  Software Engineer #42 @ http://inktank.com | http://ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Low write speed

2014-01-22 Thread Никитенко Виталий
Hi, Wido den Hollander
 Good day! Please help me solve the problem. There are the following scheme :
 Server ESXi with 1Gb NICs. it has local store store2Tb and two isci storage 
 connected to the second server .
 The second server supermicro: two 1TB hdd (lsi 9261-8i with battery), 8 CPU 
 cores, 32 GB RAM and 2 1Gb NICs . On /dev/sda installed ubuntu 12 and 
 ceph-emperor. /dev/sdb disk placed under osd.0.

How do you do journaling?

When i create osd i see: 
NFO:ceph-disk:Will colocate journal with data on /dev/sdb

Have you tried TGT instead?
I tried tgt (with --bstype rbd) and result same.

Have you also tried to run a rados benchmark? (rados bench)
rados bench -p my_pool 300 write

Total time run: 30.821284
Total writes made:  371
Write size: 4194304
Bandwidth (MB/sec): 48.149 

Stddev Bandwidth:   38.1729
Max bandwidth (MB/sec): 116
Min bandwidth (MB/sec): 0
Average Latency:1.31857
Stddev Latency: 1.6014
Max latency:9.2685
Min latency:0.013897

Also, be aware that Ceph excels in it's parallel performance. You 
shouldn't look at the performance of a single LUN or RBD image that 
much, it's much more interesting to see the aggegrated performance of 10 
or maybe 100 LUNs together.

I dont understand how do it. I must create 10 LUN and 10 iscsi storage on esxi, 
and then test migrate 10 VM from local store to iscsi storage?

Thanks! Vitaliy
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Low write speed

2014-01-19 Thread Никитенко Виталий
Hi, Ирек Нургаязович.
Number of PGS is default
pg_num 64 pgp_num 64


17.01.2014, 14:39, Ирек Фасихов malm...@gmail.com:
 Hi, Виталий.
 Whether a sufficient number of PGS?

 2014/1/17 Никитенко Виталий v1...@yandex.ru
 Good day! Please help me solve the problem. There are the following scheme :
 Server ESXi with 1Gb NICs. it has local store store2Tb and two isci storage 
 connected to the second server .
 The second server supermicro: two 1TB hdd (lsi 9261-8i with battery), 8 CPU 
 cores, 32 GB RAM and 2 1Gb NICs . On /dev/sda installed ubuntu 12 and 
 ceph-emperor. /dev/sdb disk placed under osd.0.
 What i do next:
   # rbd create esxi
   # rbd map esxi

 Get /dev/rbd1 which shared using iscsitarget

   # cat ietd.conf
   Target iqn.2014-01.ru.ceph: rados.iscsi.001
     Lun 0 Path = / dev/rbd1, Type = blockio, ScsiId = f817ab
   Target iqn.2014-01.ru.ceph: rados.iscsi.002
     Lun 1 Path = / opt/storlun0.bin, Type = fileio, ScsiId = lun1, ScsiSN = 
 lun1

 For test I also create iscsi storage on /dev/sda (Lun1).
 When migrating a virtual machine from store2Tb to Lun0 (ceph) the rate of 
 migration of 400-450 Mbit/second.
 When migrating a VM from store2Tb to Lun1 (ubuntu file) then the rate of 
 migration of 800-900 Mbit / second.
 From this I conclude that the rate is not limited by disk(controller) and 
 not to the network.
 Tried osd format to ext4 and xfs and btrfs but same speed. For me, speed is 
 very important , especially since the plan
 translate 10Gb network links.
 Thanks.
 Vitaliy
 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 --
 С уважением, Фасихов Ирек НургаязовичМоб.: +79229045757
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com