Why not use S3? The RADOS Gateway speaks S3 just fine. I've been using CloudStack with Ceph S3 without any problems.
Wido On 21-01-16 23:06, ilya wrote: > You might find this relevant > > https://issues.apache.org/jira/browse/CLOUDSTACK-9248 > > > > On 1/21/16 12:32 AM, ilya wrote: >> ++ Wido - perhaps he seen it.. >> ---- >> Hi Yuriy, >> >> I'm going to switch to english as i'm posting this thread to "users" >> mailing list >> >> Sorry, i dont have enough experience with ceph + cloudstack. >> >> When you ask on mailing lists, please clearly mention the problem - as >> its not clear. >> >> Id think its related to your configuration, since error states >>> There is no secondary storage VM for secondary storage host Images >> >> Do you have Secondary Storage VM running? >> >> Regards >> ilya >> >> On 1/19/16 10:47 PM, Юрий Карпель wrote: >>> Приветствую! >>> >>> >>> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7) >>> >>> Кластер пока только для тестов, если коротко то сделал 2-а rgw на >>> civetweb и сервер nfs + haproxy для s3,swift : >>> >>> [client.rgw.srv-rgw01] >>> rgw print continue = false >>> host = srv-rgw01 >>> rgw frontends = civetweb port=8080 >>> rgw_socket_path = /tmp/radosgw.sock >>> >>> [client.rgw.srv-rgw02] >>> rgw print continue = false >>> host = srv-rgw01 >>> rgw frontends = civetweb port=8080 >>> rgw_socket_path = /tmp/radosgw.sock >>> >>> Проверяем: >>> [ceph@ceph-adm ~]$ swift -A http://ceph-rgw.test.bst.ru:8080/auth/v1.0/ >>> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list >>> images >>> [ceph@ceph-adm ~]$ >>> [ceph@ceph-adm ~]$ swift upload images ceph.log >>> ceph.log >>> [ceph@ceph-adm ~]$ swift stat images >>> Account: v1 >>> Container: images >>> Objects: 1 >>> Bytes: 96355 >>> Read ACL: >>> Write ACL: >>> Sync To: >>> Sync Key: >>> Accept-Ranges: bytes >>> X-Storage-Policy: default-placement >>> X-Container-Bytes-Used-Actual: 98304 >>> X-Timestamp: 1453193667.00000 >>> X-Trans-Id: >>> tx000000000000000000208-00569f2d4c-4395-default >>> Content-Type: text/plain; charset=utf-8 >>> [ceph@ceph-adm ~]$ >>> >>> >>> Добавляю в cloudstack: >>> Name: Images >>> Provider: Swift >>> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/ >>> Account:cloudstack >>> Username: swift >>> Key: >>> >>> Лог: >>> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet] >>> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END=== >>> 192.168.7.29 -- GET >>> >>> command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%2Fceph-rgw.test.bst.ru >>> <http://2Fceph-rgw.test.bst.ru>%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545 >>> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting >>> hosts suitable for reconnect >>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed >>> resetting hosts suitable for reconnect >>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring >>> hosts for clusters already owned by this management server >>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed >>> acquiring hosts for clusters already owned by this management server >>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring >>> hosts for clusters not owned by any management server >>> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl] >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed >>> acquiring hosts for clusters not owned by any management server >>> 2016-01-20 09:35:47,073 INFO [o.a.c.f.j.i.AsyncJobManagerImpl] >>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup >>> expired async-jobs >>> 2016-01-20 09:35:47,100 INFO [o.a.c.f.j.i.AsyncJobManagerImpl] >>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup >>> expired async-jobs >>> 2016-01-20 09:35:47,906 INFO [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host >>> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru> >>> corresponds to IP 10.30.15.2 >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with >>> nfs-style path of 10.30.15.2:/nfs >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available >>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on >>> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs> >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount >>> will be /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 >>> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount >>> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is successful. >>> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already >>> mounted at >>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no >>> need to mount nfs://rgw-lb01test.bst.ru/nfs <http://rgw-lb01test.bst.ru/nfs> >>> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource] >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get >>> url:http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2, >>> due to java.io.IOException: access denied >>> >>> >>> Как итог не загружены шаблоны systemvm: >>> [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c) (logid:0523bff9) >>> There is no secondary storage VM for secondary storage host Images >>> >>> >>> >>> >>> >>> >>>