Why not use S3? The RADOS Gateway speaks S3 just fine.

I've been using CloudStack with Ceph S3 without any problems.

Wido

On 21-01-16 23:06, ilya wrote:
> You might find this relevant
> 
> https://issues.apache.org/jira/browse/CLOUDSTACK-9248
> 
> 
> 
> On 1/21/16 12:32 AM, ilya wrote:
>> ++ Wido - perhaps he seen it..
>> ----
>> Hi Yuriy,
>>
>> I'm going to switch to english as i'm posting this thread to "users"
>> mailing list
>>
>> Sorry, i dont have enough experience with ceph + cloudstack.
>>
>> When you ask on mailing lists, please clearly mention the problem - as
>> its not clear.
>>
>> Id think its related to your configuration, since error states
>>> There is no secondary storage VM for secondary storage host Images
>>
>> Do you have Secondary Storage VM running?
>>
>> Regards
>> ilya
>>
>> On 1/19/16 10:47 PM, Юрий Карпель wrote:
>>> Приветствую!
>>>
>>>
>>> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7)
>>>
>>> Кластер пока только для тестов, если коротко то сделал 2-а rgw на
>>> civetweb и сервер nfs + haproxy для s3,swift :
>>>
>>> [client.rgw.srv-rgw01]
>>> rgw print continue = false
>>> host = srv-rgw01
>>> rgw frontends = civetweb port=8080
>>> rgw_socket_path = /tmp/radosgw.sock
>>>
>>> [client.rgw.srv-rgw02]
>>> rgw print continue = false
>>> host = srv-rgw01
>>> rgw frontends = civetweb port=8080
>>> rgw_socket_path = /tmp/radosgw.sock
>>>
>>> Проверяем:
>>> [ceph@ceph-adm ~]$ swift -A http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>>> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list
>>> images
>>> [ceph@ceph-adm ~]$
>>> [ceph@ceph-adm ~]$ swift upload images ceph.log
>>> ceph.log
>>> [ceph@ceph-adm ~]$ swift stat images
>>>                       Account: v1
>>>                     Container: images
>>>                       Objects: 1
>>>                         Bytes: 96355
>>>                      Read ACL:
>>>                     Write ACL:
>>>                       Sync To:
>>>                      Sync Key:
>>>                 Accept-Ranges: bytes
>>>              X-Storage-Policy: default-placement
>>> X-Container-Bytes-Used-Actual: 98304
>>>                   X-Timestamp: 1453193667.00000
>>>                    X-Trans-Id:
>>> tx000000000000000000208-00569f2d4c-4395-default
>>>                  Content-Type: text/plain; charset=utf-8
>>> [ceph@ceph-adm ~]$
>>>
>>>
>>> Добавляю в cloudstack:
>>> Name: Images
>>> Provider: Swift
>>> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>>> Account:cloudstack
>>> Username: swift
>>> Key: 
>>>
>>> Лог:
>>> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet]
>>> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END===
>>>  192.168.7.29 -- GET
>>>  
>>> command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%2Fceph-rgw.test.bst.ru
>>> <http://2Fceph-rgw.test.bst.ru>%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545
>>> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting
>>> hosts suitable for reconnect
>>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> resetting hosts suitable for reconnect
>>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>>> hosts for clusters already owned by this management server
>>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> acquiring hosts for clusters already owned by this management server
>>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>>> hosts for clusters not owned by any management server
>>> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> acquiring hosts for clusters not owned by any management server
>>> 2016-01-20 09:35:47,073 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup
>>> expired async-jobs
>>> 2016-01-20 09:35:47,100 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup
>>> expired async-jobs
>>> 2016-01-20 09:35:47,906 INFO  [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host
>>> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru>
>>> corresponds to IP 10.30.15.2
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with
>>> nfs-style path of 10.30.15.2:/nfs
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available
>>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on
>>> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs>
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount
>>> will be /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84
>>> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount
>>> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is successful.
>>> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already
>>> mounted at
>>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no
>>> need to mount nfs://rgw-lb01test.bst.ru/nfs <http://rgw-lb01test.bst.ru/nfs>
>>> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get
>>> url:http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2,
>>> due to java.io.IOException: access denied
>>>
>>>
>>> Как итог не загружены шаблоны systemvm:
>>>  [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c) (logid:0523bff9)
>>> There is no secondary storage VM for secondary storage host Images
>>>
>>>
>>>
>>>
>>>  
>>>
>>>

Reply via email to