Hello Chenrui, thank you for answering, I've done some changes and they are like you said,
cinder quota-update --gigabytes <#number> $tenant-id cinder quota-show $tenant-id I'll try to create the storages again tomorrow and I'll post the results here. Regards. Guilherme. 2013/11/19 Chenrui (A) <kiwik.chen...@huawei.com> > Please check your tenant quota and calculate how much free space in > Grizzly > > cinder quota-show ${tenant-id} > > cinder list > > > > If you are using Havana, there is a convenient command “cinder quota-usage” > > J > > > > *发件人:* Guilherme Russi [mailto:luisguilherme...@gmail.com] > *发送时间:* 2013年11月18日 21:12 > *收件人:* Razique Mahroua > *抄送:* openstack > *主题:* Re: [Openstack] Multiples storages > > > > Hello again guys, I'm trying to overcome another issue with cinder, I'm > here trying to create e 500 GB volume, this is my disk: > > > > pvscan > > PV /dev/sdb1 VG cinder-volumes-2 lvm2 [931,51 GiB / 531,51 GiB free] > > Total: 1 [931,51 GiB] / in use: 1 [931,51 GiB] / in no VG: 0 [0 ] > > > > But when I try: > > cinder create --volume_type lvm_one --display_name v2st3-500 500 > > > > I get: > > ERROR: VolumeSizeExceedsAvailableQuota: Requested volume or snapshot > exceeds allowed Gigabytes quota > > > > Does anybody know where I begin to fix it? > > > > The outputs on cinder-api.log are: > > > > ERROR [cinder.api.middleware.fault] Caught error: Requested volume or > snapshot exceeds allowed Gigabytes quota > > Traceback (most recent call last): > > File "/usr/lib/python2.7/dist-packages/cinder/api/middleware/fault.py", > line 73, in __call__ > > return req.get_response(self.application) > > File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1296, in > send > > application, catch_exc_info=False) > > File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1260, in > call_application > > app_iter = application(self.environ, start_response) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in > __call__ > > return resp(environ, start_response) > > File > "/usr/lib/python2.7/dist-packages/keystoneclient/middleware/auth_token.py", > line 450, in __call__ > > return self.app(env, start_response) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in > __call__ > > return resp(environ, start_response) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in > __call__ > > return resp(environ, start_response) > > File "/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131, > in __call__ > > response = self.app(environ, start_response) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in > __call__ > > return resp(environ, start_response) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 130, in > __call__ > > resp = self.call_func(req, *args, **self.kwargs) > > File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 195, in > call_func > > return self.func(req, *args, **kwargs) > > File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", > line 803, in __call__ > > content_type, body, accept) > > File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", > line 851, in _process_stack > > action_result = self.dispatch(meth, request, action_args) > > File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", > line 927, in dispatch > > return method(req=request, **action_args) > > File "/usr/lib/python2.7/dist-packages/cinder/api/v1/volumes.py", line > 358, in create > > **kwargs) > > File "/usr/lib/python2.7/dist-packages/cinder/volume/api.py", line 165, > in create > > raise exception.VolumeSizeExceedsAvailableQuota() > > VolumeSizeExceedsAvailableQuota: Requested volume or snapshot exceeds > allowed Gigabytes quota > > > > Thank you all. > > > > > > 2013/11/15 Razique Mahroua <razique.mahr...@gmail.com> > > Awesome :) > > Razique > > > On 14 Nov 2013, at 15:27, Guilherme Russi wrote: > > That's right, I've stopped the open-iscsi and tgt process, and the > lvremove > worked. Thank you all. > > Regards. > > > 2013/11/13 Razique Mahroua <razique.mahr...@gmail.com> > > Hey :) > that means the volume is still in use. (lvopen : 1) make sure it's not by > checking the process, qemu-nbd, etc... > > On 13 Nov 2013, at 4:50, Guilherme Russi wrote: > > Hello Razique, I'm here opening this thread again, I've done some cinder > delete but when I try to create another storeges it returns there's no > space to create a new volume. > > Here is part of my lvdisplay output: > > Alloc PE / Size 52224 / 204,00 GiB > Free PE / Size 19350 / 75,59 GiB > > And here is my lvdisplay: > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b > VG Name cinder-volumes > LV UUID wdqxVd-GgUQ-21O4-OWlR-sRT3-HvUA-Q8j9kL > LV Write Access read/write > LV snapshot status source of > > /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09 [active] > LV Status available > open 0 > > LV Size 10,00 GiB > Current LE 2560 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:1 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09 > VG Name cinder-volumes > LV UUID EZz1lC-a8H2-1PlN-pJTN-XAIm-wW0q-qtUQOc > LV Write Access read/write > LV snapshot status active destination for > /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b > LV Status available > open 0 > > LV Size 10,00 GiB > Current LE 2560 > COW-table size 10,00 GiB > COW-table LE 2560 > Allocated to snapshot 0,00% > Snapshot chunk size 4,00 KiB > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:3 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-ca36920e-938e-4ad1-b9c4-74c1e28abd31 > VG Name cinder-volumes > LV UUID b40kQV-P8N4-R6jt-k97Z-I2a1-9TXm-5GXqfz > LV Write Access read/write > LV Status available > open 1 > > LV Size 60,00 GiB > Current LE 15360 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:4 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-70be4f36-10bd-4877-b841-80333ccfe985 > VG Name cinder-volumes > LV UUID 2YDrMs-BrYo-aQcZ-8AlX-A4La-HET1-9UQ0gV > LV Write Access read/write > LV Status available > open 1 > > LV Size 1,00 GiB > Current LE 256 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:5 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-00c532bd-91fb-4a38-b340-4389fb7f0ed5 > VG Name cinder-volumes > LV UUID MfVOuB-5x5A-jne3-H4Ul-4NP8-eI7b-UYSYE7 > LV Write Access read/write > LV Status available > open 0 > > LV Size 1,00 GiB > Current LE 256 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:6 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-ae133dbc-6141-48cf-beeb-9d6576e57a45 > VG Name cinder-volumes > LV UUID 53w8j3-WT4V-8m52-r6LK-ZYd3-mMHA-FtuyXV > LV Write Access read/write > LV Status available > open 0 > > LV Size 1,00 GiB > Current LE 256 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:7 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-954d2f1b-837b-4ba5-abfd-b3610597be5e > VG Name cinder-volumes > LV UUID belquE-WxQ2-gt6Y-WlPE-Hmq3-B9Am-zcYD3P > LV Write Access read/write > LV Status available > open 0 > > LV Size 60,00 GiB > Current LE 15360 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:8 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-05d037d1-4e61-4419-929a-fe340e00e1af > VG Name cinder-volumes > LV UUID Pt61e7-l3Nu-1IdX-T2sb-0GQD-PhS6-XtIIUj > LV Write Access read/write > LV Status available > open 1 > > LV Size 1,00 GiB > Current LE 256 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:9 > > --- Logical volume --- > LV Name > /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f > VG Name cinder-volumes > LV UUID e46mBx-CRps-HYKk-aJsc-XFRd-B1Rv-UVk8gT > LV Write Access read/write > LV Status available > open 1 > > LV Size 60,00 GiB > Current LE 15360 > Segments 1 > Allocation inherit > Read ahead sectors auto > - currently set to 256 > Block device 252:10 > > Do you know how I can remove all of them? When I try lvremove > /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f, for > example, I get: > > Can't remove open logical volume > "volume-316f77c6-bf13-4ea4-9b98-028198f3922f" > > Thank you again. > > 2013/11/8 Guilherme Russi luisguilherme...@gmail.com > > Very thanks again. > > Best regards. > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > Oh yah true! > not sure “conductors” exist yet for Cinder, meaning meanwhile, every node > needs a direct access to the database > glad to hear it’s working :) > > On 08 Nov 2013, at 08:53, Guilherme Russi luisguilherme...@gmail.com > wrote: > > Hello again Razique, I've found the problem, I need to add the grants on > the mysql to my another IP. Now it's working really good :D > I've found this link too if someone needs: > http://docs.openstack.org/admin-guide-cloud/content//managing-volumes.html<http://docs.openstack.org/admin-guide-cloud/content/managing-volumes.html> > > Thank you so much, and if you need me just let me know. > > Best regards. > > Guilherme. > > 2013/11/8 Guilherme Russi luisguilherme...@gmail.com > > Hello Razique, I got a couple of doubts, do you know if I need to do > something else that's is not on the link you sent me? I'm asking because I > followed the configuration but it's not working, here is what I get: I've > installed the cinder-volume at the second computer that have the HD, and > I've changed it's cinder.conf. I've changed too the master's cinder.conf > like is following: > > [DEFAULT] > rootwrap > > *config = /etc/cinder/rootwrap.conf sql*connection = > mysql://cinder:password@localhost/cinder > api*paste*confg = /etc/cinder/api-paste.ini > > > iscsi_helper=iscsiadm iscsi_helper = ietadm > > iscsi > > *helper = tgtadm volume*name_template = volume-%s > > > volume_group = cinder-volumes > > verbose = True > auth > > *strategy = keystone iscsi*ip > *address = 192.168.3.1 scheduler* > driver=cinder.scheduler.filter_scheduler.FilterScheduler > Rabbit authorization > > rabbit > *host = localhost rabbit*port = 5672 > rabbit*hosts = $rabbit*host:$rabbit > *port rabbit*use > *ssl = false rabbit*userid = guest > rabbit_password = password > rabbit*virtual*host = /nova > > state > *path = /var/lib/cinder lock*path = /var/lock/cinder > volumes_dir = /var/lib/cinder/volumes > rpc*backend = cinder.rpc.impl*kombu > > enabled > > *backends=orion-1,orion-4 [orion-1] volume*group=cinder-volumes > volume > *driver=cinder.volume.drivers.lvm.LVMISCSIDriver volume*backend*name=LVM* > iSCSI > [orion-4] > volume > *group=cinder-volumes-2 volume* > driver=cinder.volume.drivers.lvm.LVMISCSIDriver > volume*backend*name=LVM_iSCSI > > > > The cinder.conf on the second computer is like this but the IPs are > changed with the controller IP (It has the cinder-api), and when I run > service cinder-volume restart at the second computer it's status is > stop/waiting. > > Any ideas? > > Thanks :) > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > sure :) > > On 08 Nov 2013, at 05:39, Guilherme Russi luisguilherme...@gmail.com > wrote: > > Oh great! I'll try here and send you the results. > > Very thanks :) > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > If I’m not mistaken, you only need to install the “cinder-volume’ > service that will update its status to your main node > :) > > On 08 Nov 2013, at 05:34, Guilherme Russi luisguilherme...@gmail.com > wrote: > > Great! I was reading the link and I have one question, do I need to > install cinder at the other computer too? > > Thanks :) > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > Ok in that case, with Grizzly you can use the “multi-backends” > feature: > https://wiki.openstack.org/wiki/Cinder-multi-backend > > and that should do it :) > > On 08 Nov 2013, at 05:29, Guilherme Russi luisguilherme...@gmail.com > wrote: > > It is a hard disk, my scenario is one Controller (where I have my > storage cinder and my network quantum) and four compute nodes. > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > ok ! > what is your actual Cinder backend? Is it a hard disk, a SAN, a > network volume, etc… > > On 08 Nov 2013, at 05:20, Guilherme Russi < > luisguilherme...@gmail.com> wrote: > > Hi Razique, thank you for answering, I want to expand my cinder > storage, is it the block storage? I'll use the storage to allow VMs to have > more hard disk space. > > Regards. > > Guilherme. > > 2013/11/8 Razique Mahroua razique.mahr...@gmail.com > > Hi Guilherme ! > Which storage do you precisely want to expand? > > Regards, > Razique > > On 08 Nov 2013, at 04:52, Guilherme Russi < > luisguilherme...@gmail.com> wrote: > > Hello guys, I have a Grizzly deployment running fine with 5 > > nodes, and I want to add more storage on it. My question is, can I install > a new HD on another computer thats not the controller and link this HD with > my cinder that it can be a storage too? > > The computer I will install my new HD is at the same network as > > my cloud is. I'm asking because I haven't seen a question like that here. > Does anybody knows how to do that? Have a clue? Any help is welcome. > > Thank you all. > > Best regards. > > ------------------------------ > > > > Mailing list: > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > Post to : openstack@lists.openstack.org > Unsubscribe : > > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > >
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack