Any ideas? Pretty please..... dk
On Fri, Sep 21, 2012 at 11:51 AM, Dominic Kaiser <domi...@bostonvineyard.org > wrote: > I noticed something. If I am trying to mount the gluster share from > another computer and do not include mounproto=tcp it times out. vers=3 or > 4 does not matter. Could this be why I can not add it from the engine gui? > > dk > > > On Fri, Sep 21, 2012 at 11:12 AM, Dominic Kaiser < > domi...@bostonvineyard.org> wrote: > >> Here is the engine.log info: >> >> [root@ovirt ovirt-engine]# tail engine.log >> 2012-09-21 11:10:00,007 INFO >> [org.ovirt.engine.core.bll.AutoRecoveryManager] >> (QuartzScheduler_Worker-49) Autorecovering 0 hosts >> 2012-09-21 11:10:00,007 INFO >> [org.ovirt.engine.core.bll.AutoRecoveryManager] >> (QuartzScheduler_Worker-49) Checking autorecoverable hosts done >> 2012-09-21 11:10:00,008 INFO >> [org.ovirt.engine.core.bll.AutoRecoveryManager] >> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains >> 2012-09-21 11:10:00,009 INFO >> [org.ovirt.engine.core.bll.AutoRecoveryManager] >> (QuartzScheduler_Worker-49) Autorecovering 0 storage domains >> 2012-09-21 11:10:00,010 INFO >> [org.ovirt.engine.core.bll.AutoRecoveryManager] >> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains done >> 2012-09-21 11:10:22,710 ERROR >> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils] >> (QuartzScheduler_Worker-84) Failed to decryptData must not be longer than >> 256 bytes >> 2012-09-21 11:10:22,726 ERROR >> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils] >> (QuartzScheduler_Worker-12) Failed to decryptData must start with zero >> 2012-09-21 11:10:54,519 INFO >> [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand] >> (ajp--0.0.0.0-8009-11) [3769be9c] Running command: >> RemoveStorageServerConnectionCommand internal: false. Entities affected : >> ID: aaa00000-0000-0000-0000-123456789aaa Type: System >> 2012-09-21 11:10:54,537 INFO >> >> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand] >> (ajp--0.0.0.0-8009-11) [3769be9c] START, >> DisconnectStorageServerVDSCommand(vdsId = >> 3822e6c0-0295-11e2-86e6-d74ad5358c03, storagePoolId = >> 00000000-0000-0000-0000-000000000000, storageType = NFS, connectionList = >> [{ id: null, connection: gfs1.bostonvineyard.org:/data };]), log id: >> 16dd4a1b >> 2012-09-21 11:10:56,417 INFO >> >> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand] >> (ajp--0.0.0.0-8009-11) [3769be9c] FINISH, >> DisconnectStorageServerVDSCommand, return: >> {00000000-0000-0000-0000-000000000000=477}, log id: 16dd4a1b >> >> Thanks, >> >> dk >> >> On Fri, Sep 21, 2012 at 11:09 AM, Dominic Kaiser < >> domi...@bostonvineyard.org> wrote: >> >>> I can mount to another computer with this command: >>> >>> mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data >>> /home/administrator/test >>> >>> So volumes work but I get a 500 error timeout when trying to add as a >>> storage domain in ovirt. weird? >>> >>> dk >>> >>> On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser < >>> domi...@bostonvineyard.org> wrote: >>> >>>> Hey All, >>>> >>>> So I finally found the problem. Cheap NIC's. Installed Intel NIC's no >>>> problems creating gluster volumes and distributed replicated ones. >>>> Broadcom and Realtek yuk! So now I am trying to mount the gluster volume >>>> as a nfs mount and am having a problem. It is timing out like it is >>>> blocked by a firewall. >>>> >>>> I am trying to: mount -t nfs gfs1.bostonvineyard.org:/export >>>> /home/administrator/test >>>> >>>> Here is gfs1 tail vdsm.log >>>> >>>> [root@gfs1 vdsm]# tail vdsm.log >>>> Thread-88731::DEBUG::2012-09-21 >>>> 10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll) >>>> Owner.cancelAll requests {} >>>> Thread-88731::DEBUG::2012-09-21 >>>> 10:35:56,567::task::978::TaskManager.Task::(_decref) >>>> Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,890::task::588::TaskManager.Task::(_updateState) >>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init -> >>>> state preparing >>>> Thread-88737::INFO::2012-09-21 >>>> 10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect: >>>> repoStats(options=None) >>>> Thread-88737::INFO::2012-09-21 >>>> 10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect: >>>> repoStats, Return response: {} >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,891::task::1172::TaskManager.Task::(prepare) >>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {} >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,892::task::588::TaskManager.Task::(_updateState) >>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state preparing -> >>>> state finished >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll) >>>> Owner.releaseAll requests {} resources {} >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll) >>>> Owner.cancelAll requests {} >>>> Thread-88737::DEBUG::2012-09-21 >>>> 10:36:06,893::task::978::TaskManager.Task::(_decref) >>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False >>>> >>>> Do you know why I can not connect via NFS? Using an older kernel not >>>> 3.5 and iptables are off. >>>> >>>> Dominic >>>> >>>> >>>> On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hat...@redhat.com> wrote: >>>> >>>>> On 09/10/2012 06:27 PM, Dominic Kaiser wrote: >>>>> >>>>>> Here is the message and the logs again except zipped I failed the >>>>>> first delivery: >>>>>> >>>>>> Ok here are the logs 4 node and 1 engine log. Tried making /data >>>>>> folder owned by root and then tried by 36:36 neither worked. Name of >>>>>> volume is data to match folders on nodes also. >>>>>> >>>>>> Let me know what you think, >>>>>> >>>>>> Dominic >>>>>> >>>>> >>>>> this is the actual failure (taken from gfs2vdsm.log). >>>>> >>>>> Thread-332442::DEBUG::2012-09-**10 >>>>> 10:28:05,788::BindingXMLRPC::**859::vds::(wrapper) >>>>> client [10.3.0.241]::call volumeCreate with ('data', ['10.4.0.97:/data', >>>>> '10.4.0.98:/data', '10.4.0.99:/data', '10.4.0.100:/data'], >>>>> 2, 0, ['TCP']) {} flowID [406f2c8e] >>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10 >>>>> 10:28:05,792::__init__::1249::**Storage.Misc.excCmd::(_log) >>>>> '/usr/sbin/gluster --mode=script volume create data replica 2 transport >>>>> TCP >>>>> 10.4.0.97:/data 10.4.0.98:/data 10 >>>>> .4.0.99:/data 10.4.0.100:/data' (cwd None) >>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10 >>>>> 10:28:05,900::__init__::1249::**Storage.Misc.excCmd::(_log) FAILED: >>>>> <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255 >>>>> MainProcess|Thread-332442::**ERROR::2012-09-10 >>>>> 10:28:05,900::supervdsmServer:**:76::SuperVdsm.ServerCallback:**:(wrapper) >>>>> Error in wrapper >>>>> Traceback (most recent call last): >>>>> File "/usr/share/vdsm/**supervdsmServer.py", line 74, in wrapper >>>>> return func(*args, **kwargs) >>>>> File "/usr/share/vdsm/**supervdsmServer.py", line 286, in wrapper >>>>> return func(*args, **kwargs) >>>>> File "/usr/share/vdsm/gluster/cli.**py", line 46, in wrapper >>>>> return func(*args, **kwargs) >>>>> File "/usr/share/vdsm/gluster/cli.**py", line 176, in volumeCreate >>>>> raise ge.**GlusterVolumeCreateFailedExcep**tion(rc, out, err) >>>>> GlusterVolumeCreateFailedExcep**tion: Volume create failed >>>>> error: Host 10.4.0.99 not a friend >>>>> return code: 255 >>>>> Thread-332442::ERROR::2012-09-**10 >>>>> 10:28:05,901::BindingXMLRPC::**877::vds::(wrapper) >>>>> unexpected error >>>>> Traceback (most recent call last): >>>>> File "/usr/share/vdsm/**BindingXMLRPC.py", line 864, in wrapper >>>>> res = f(*args, **kwargs) >>>>> File "/usr/share/vdsm/gluster/api.**py", line 32, in wrapper >>>>> rv = func(*args, **kwargs) >>>>> File "/usr/share/vdsm/gluster/api.**py", line 87, in volumeCreate >>>>> transportList) >>>>> File "/usr/share/vdsm/supervdsm.py"**, line 67, in __call__ >>>>> return callMethod() >>>>> File "/usr/share/vdsm/supervdsm.py"**, line 65, in <lambda> >>>>> **kwargs) >>>>> File "<string>", line 2, in glusterVolumeCreate >>>>> File "/usr/lib64/python2.7/**multiprocessing/managers.py", line >>>>> 759, in _callmethod >>>>> kind, result = conn.recv() >>>>> TypeError: ('__init__() takes exactly 4 arguments (1 given)', <class >>>>> 'gluster.exception.**GlusterVolumeCreateFailedExcep**tion'>, ()) >>>>> >>>>> can you please run gluster peer status on all your nodes ? also, it >>>>> appears that '10.4.0.99' is problematic, can you try create the volume >>>>> without it ? >>>>> >>>>> >>>>> >>>>>> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser < >>>>>> domi...@bostonvineyard.org >>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org>>> >>>>>> wrote: >>>>>> >>>>>> Here are the other two logs forgot them. >>>>>> >>>>>> dk >>>>>> >>>>>> >>>>>> On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser >>>>>> <domi...@bostonvineyard.org >>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org> >>>>>> >> >>>>>> >>>>>> wrote: >>>>>> >>>>>> Ok here are the logs 4 node and 1 engine log. Tried making >>>>>> /data folder owned by root and then tried by 36:36 neither >>>>>> worked. Name of volume is data to match folders on nodes >>>>>> also. >>>>>> >>>>>> Let me know what you think, >>>>>> >>>>>> Dominic >>>>>> >>>>>> >>>>>> On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout >>>>>> <ma...@wzzrd.com <mailto:ma...@wzzrd.com>> wrote: >>>>>> >>>>>> I just ran into this as well, and it seems that you have >>>>>> to either reformat previously used gluster bricks or >>>>>> manually tweak some extended attributes. >>>>>> >>>>>> Maybe this helps you in setting up your gluster volume, >>>>>> Dominic? >>>>>> >>>>>> More info here: >>>>>> >>>>>> https://bugzilla.redhat.com/**show_bug.cgi?id=812214<https://bugzilla.redhat.com/show_bug.cgi?id=812214> >>>>>> >>>>>> >>>>>> Maxim Burgerhout >>>>>> ma...@wzzrd.com <mailto:ma...@wzzrd.com> >>>>>> >>>>>> ---------------- >>>>>> EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal >>>>>> <san...@redhat.com <mailto:san...@redhat.com>> wrote: >>>>>> >>>>>> Hi Dominic, >>>>>> >>>>>> Looking at the engine log immediately after trying to >>>>>> create the volume should tell us on which node the >>>>>> gluster volume creation was attempted. Then looking at >>>>>> the vdsm log on that node should help us identifying >>>>>> the exact reason for failure. >>>>>> >>>>>> In case this doesn't help you, can you please simulate >>>>>> the issue again and send back all the 5 log files? >>>>>> (engine.log from engine server and vdsm.log from the 4 >>>>>> nodes) >>>>>> >>>>>> Regards, >>>>>> Shireesh >>>>>> >>>>>> >>>>>> On Wednesday 05 September 2012 11:50 PM, Dominic >>>>>> Kaiser wrote: >>>>>> >>>>>>> So I have a problem creating glusterfs volumes. Here >>>>>>> is the install: >>>>>>> >>>>>>> 1. Ovirt 3.1 >>>>>>> 2. 4 Nodes are Fedora 17 with kernel 3.3.4 - >>>>>>> 5.fc17.x86_64 >>>>>>> 3. 4 nodes peer joined and running >>>>>>> 4. 4 nodes added as hosts to ovirt >>>>>>> 5. created a directory on each node this path /data >>>>>>> 6. chmod 36.36 -R /data all nodes >>>>>>> 7. went back to ovirt and created a >>>>>>> >>>>>>> distributed/replicated volume and added the 4 >>>>>>> nodes with brick path of /data >>>>>>> >>>>>>> I received this error: >>>>>>> >>>>>>> Creation of Gluster Volume maingfs1 failed. >>>>>>> >>>>>>> I went and looked at the vdsm logs on the nodes and >>>>>>> the ovirt server which did not say much. Where else >>>>>>> should I look? Also this error is vague what does it >>>>>>> mean? >>>>>>> >>>>>>> >>>>>>> -- Dominic Kaiser >>>>>>> Greater Boston Vineyard >>>>>>> Director of Operations >>>>>>> >>>>>>> cell: 617-230-1412 <tel:617-230-1412> >>>>>>> fax: 617-252-0238 <tel:617-252-0238> >>>>>>> email: domi...@bostonvineyard.org >>>>>>> >>>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org> >>>>>>> > >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> ______________________________**_________________ >>>>>>> Users mailing list >>>>>>> Users@ovirt.org <mailto:Users@ovirt.org> >>>>>>> >>>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users> >>>>>>> >>>>>> >>>>>> >>>>>> ______________________________**_________________ >>>>>> Users mailing list >>>>>> Users@ovirt.org <mailto:Users@ovirt.org> >>>>>> >>>>>> >>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- Dominic Kaiser >>>>>> Greater Boston Vineyard >>>>>> Director of Operations >>>>>> >>>>>> cell: 617-230-1412 <tel:617-230-1412> >>>>>> fax: 617-252-0238 <tel:617-252-0238> >>>>>> email: domi...@bostonvineyard.org >>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org> >>>>>> > >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- Dominic Kaiser >>>>>> Greater Boston Vineyard >>>>>> Director of Operations >>>>>> >>>>>> cell: 617-230-1412 <tel:617-230-1412> >>>>>> fax: 617-252-0238 <tel:617-252-0238> >>>>>> email: domi...@bostonvineyard.org <mailto:dominic@** >>>>>> bostonvineyard.org <domi...@bostonvineyard.org>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Dominic Kaiser >>>>>> Greater Boston Vineyard >>>>>> Director of Operations >>>>>> >>>>>> cell: 617-230-1412 >>>>>> fax: 617-252-0238 >>>>>> email: domi...@bostonvineyard.org <mailto:dominic@** >>>>>> bostonvineyard.org <domi...@bostonvineyard.org>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> ______________________________**_________________ >>>>>> Users mailing list >>>>>> Users@ovirt.org >>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users> >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Dominic Kaiser >>>> Greater Boston Vineyard >>>> Director of Operations >>>> >>>> cell: 617-230-1412 >>>> fax: 617-252-0238 >>>> email: domi...@bostonvineyard.org >>>> >>>> >>>> >>> >>> >>> -- >>> Dominic Kaiser >>> Greater Boston Vineyard >>> Director of Operations >>> >>> cell: 617-230-1412 >>> fax: 617-252-0238 >>> email: domi...@bostonvineyard.org >>> >>> >>> >> >> >> -- >> Dominic Kaiser >> Greater Boston Vineyard >> Director of Operations >> >> cell: 617-230-1412 >> fax: 617-252-0238 >> email: domi...@bostonvineyard.org >> >> >> > > > -- > Dominic Kaiser > Greater Boston Vineyard > Director of Operations > > cell: 617-230-1412 > fax: 617-252-0238 > email: domi...@bostonvineyard.org > > > -- Dominic Kaiser Greater Boston Vineyard Director of Operations cell: 617-230-1412 fax: 617-252-0238 email: domi...@bostonvineyard.org
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users