Any ideas?  Pretty please.....

dk

On Fri, Sep 21, 2012 at 11:51 AM, Dominic Kaiser <domi...@bostonvineyard.org
> wrote:

> I noticed something.  If I am trying to mount the gluster share from
> another computer and do not include mounproto=tcp it times out.  vers=3 or
> 4 does not matter.  Could this be why I can not add it from the engine gui?
>
> dk
>
>
> On Fri, Sep 21, 2012 at 11:12 AM, Dominic Kaiser <
> domi...@bostonvineyard.org> wrote:
>
>> Here is the engine.log info:
>>
>> [root@ovirt ovirt-engine]# tail engine.log
>> 2012-09-21 11:10:00,007 INFO
>>  [org.ovirt.engine.core.bll.AutoRecoveryManager]
>> (QuartzScheduler_Worker-49) Autorecovering 0 hosts
>> 2012-09-21 11:10:00,007 INFO
>>  [org.ovirt.engine.core.bll.AutoRecoveryManager]
>> (QuartzScheduler_Worker-49) Checking autorecoverable hosts done
>> 2012-09-21 11:10:00,008 INFO
>>  [org.ovirt.engine.core.bll.AutoRecoveryManager]
>> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains
>> 2012-09-21 11:10:00,009 INFO
>>  [org.ovirt.engine.core.bll.AutoRecoveryManager]
>> (QuartzScheduler_Worker-49) Autorecovering 0 storage domains
>> 2012-09-21 11:10:00,010 INFO
>>  [org.ovirt.engine.core.bll.AutoRecoveryManager]
>> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains done
>> 2012-09-21 11:10:22,710 ERROR
>> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils]
>> (QuartzScheduler_Worker-84) Failed to decryptData must not be longer than
>> 256 bytes
>> 2012-09-21 11:10:22,726 ERROR
>> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils]
>> (QuartzScheduler_Worker-12) Failed to decryptData must start with zero
>> 2012-09-21 11:10:54,519 INFO
>>  [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand]
>> (ajp--0.0.0.0-8009-11) [3769be9c] Running command:
>> RemoveStorageServerConnectionCommand internal: false. Entities affected :
>>  ID: aaa00000-0000-0000-0000-123456789aaa Type: System
>> 2012-09-21 11:10:54,537 INFO
>>  
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand]
>> (ajp--0.0.0.0-8009-11) [3769be9c] START,
>> DisconnectStorageServerVDSCommand(vdsId =
>> 3822e6c0-0295-11e2-86e6-d74ad5358c03, storagePoolId =
>> 00000000-0000-0000-0000-000000000000, storageType = NFS, connectionList =
>> [{ id: null, connection: gfs1.bostonvineyard.org:/data };]), log id:
>> 16dd4a1b
>> 2012-09-21 11:10:56,417 INFO
>>  
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand]
>> (ajp--0.0.0.0-8009-11) [3769be9c] FINISH,
>> DisconnectStorageServerVDSCommand, return:
>> {00000000-0000-0000-0000-000000000000=477}, log id: 16dd4a1b
>>
>> Thanks,
>>
>> dk
>>
>> On Fri, Sep 21, 2012 at 11:09 AM, Dominic Kaiser <
>> domi...@bostonvineyard.org> wrote:
>>
>>> I can mount to another computer with this command:
>>>
>>> mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data
>>> /home/administrator/test
>>>
>>> So volumes work but I get a 500 error timeout when trying to add as a
>>> storage domain in ovirt.  weird?
>>>
>>> dk
>>>
>>> On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser <
>>> domi...@bostonvineyard.org> wrote:
>>>
>>>> Hey All,
>>>>
>>>> So I finally found the problem.  Cheap NIC's.  Installed Intel NIC's no
>>>> problems creating gluster volumes and distributed replicated ones.
>>>>  Broadcom and Realtek yuk!  So now I am trying to mount the gluster volume
>>>> as a nfs mount and am having a problem.  It is timing out like it is
>>>> blocked by a firewall.
>>>>
>>>> I am trying to:  mount -t nfs gfs1.bostonvineyard.org:/export
>>>> /home/administrator/test
>>>>
>>>> Here is gfs1 tail vdsm.log
>>>>
>>>> [root@gfs1 vdsm]# tail vdsm.log
>>>> Thread-88731::DEBUG::2012-09-21
>>>> 10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
>>>> Owner.cancelAll requests {}
>>>> Thread-88731::DEBUG::2012-09-21
>>>> 10:35:56,567::task::978::TaskManager.Task::(_decref)
>>>> Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,890::task::588::TaskManager.Task::(_updateState)
>>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init ->
>>>> state preparing
>>>> Thread-88737::INFO::2012-09-21
>>>> 10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
>>>> repoStats(options=None)
>>>> Thread-88737::INFO::2012-09-21
>>>> 10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
>>>> repoStats, Return response: {}
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,891::task::1172::TaskManager.Task::(prepare)
>>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,892::task::588::TaskManager.Task::(_updateState)
>>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state preparing ->
>>>> state finished
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
>>>> Owner.releaseAll requests {} resources {}
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
>>>> Owner.cancelAll requests {}
>>>> Thread-88737::DEBUG::2012-09-21
>>>> 10:36:06,893::task::978::TaskManager.Task::(_decref)
>>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False
>>>>
>>>> Do you know why I can not connect via NFS?  Using an older kernel not
>>>> 3.5 and iptables are off.
>>>>
>>>> Dominic
>>>>
>>>>
>>>> On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hat...@redhat.com> wrote:
>>>>
>>>>> On 09/10/2012 06:27 PM, Dominic Kaiser wrote:
>>>>>
>>>>>> Here is the message and the logs again except zipped I failed the
>>>>>> first delivery:
>>>>>>
>>>>>> Ok here are the logs 4 node and 1 engine log.  Tried making /data
>>>>>> folder owned by root and then tried by 36:36 neither worked.  Name of
>>>>>> volume is data to match folders on nodes also.
>>>>>>
>>>>>> Let me know what you think,
>>>>>>
>>>>>> Dominic
>>>>>>
>>>>>
>>>>> this is the actual failure (taken from gfs2vdsm.log).
>>>>>
>>>>> Thread-332442::DEBUG::2012-09-**10 
>>>>> 10:28:05,788::BindingXMLRPC::**859::vds::(wrapper)
>>>>> client [10.3.0.241]::call volumeCreate with ('data', ['10.4.0.97:/data',
>>>>> '10.4.0.98:/data', '10.4.0.99:/data', '10.4.0.100:/data'],
>>>>>  2, 0, ['TCP']) {} flowID [406f2c8e]
>>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>>>>> 10:28:05,792::__init__::1249::**Storage.Misc.excCmd::(_log)
>>>>> '/usr/sbin/gluster --mode=script volume create data replica 2 transport 
>>>>> TCP
>>>>> 10.4.0.97:/data 10.4.0.98:/data 10
>>>>> .4.0.99:/data 10.4.0.100:/data' (cwd None)
>>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>>>>> 10:28:05,900::__init__::1249::**Storage.Misc.excCmd::(_log) FAILED:
>>>>> <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
>>>>> MainProcess|Thread-332442::**ERROR::2012-09-10
>>>>> 10:28:05,900::supervdsmServer:**:76::SuperVdsm.ServerCallback:**:(wrapper)
>>>>> Error in wrapper
>>>>> Traceback (most recent call last):
>>>>>   File "/usr/share/vdsm/**supervdsmServer.py", line 74, in wrapper
>>>>>     return func(*args, **kwargs)
>>>>>   File "/usr/share/vdsm/**supervdsmServer.py", line 286, in wrapper
>>>>>     return func(*args, **kwargs)
>>>>>   File "/usr/share/vdsm/gluster/cli.**py", line 46, in wrapper
>>>>>     return func(*args, **kwargs)
>>>>>   File "/usr/share/vdsm/gluster/cli.**py", line 176, in volumeCreate
>>>>>     raise ge.**GlusterVolumeCreateFailedExcep**tion(rc, out, err)
>>>>> GlusterVolumeCreateFailedExcep**tion: Volume create failed
>>>>> error: Host 10.4.0.99 not a friend
>>>>> return code: 255
>>>>> Thread-332442::ERROR::2012-09-**10 
>>>>> 10:28:05,901::BindingXMLRPC::**877::vds::(wrapper)
>>>>> unexpected error
>>>>> Traceback (most recent call last):
>>>>>   File "/usr/share/vdsm/**BindingXMLRPC.py", line 864, in wrapper
>>>>>     res = f(*args, **kwargs)
>>>>>   File "/usr/share/vdsm/gluster/api.**py", line 32, in wrapper
>>>>>     rv = func(*args, **kwargs)
>>>>>   File "/usr/share/vdsm/gluster/api.**py", line 87, in volumeCreate
>>>>>     transportList)
>>>>>   File "/usr/share/vdsm/supervdsm.py"**, line 67, in __call__
>>>>>     return callMethod()
>>>>>   File "/usr/share/vdsm/supervdsm.py"**, line 65, in <lambda>
>>>>>     **kwargs)
>>>>>   File "<string>", line 2, in glusterVolumeCreate
>>>>>   File "/usr/lib64/python2.7/**multiprocessing/managers.py", line
>>>>> 759, in _callmethod
>>>>>     kind, result = conn.recv()
>>>>> TypeError: ('__init__() takes exactly 4 arguments (1 given)', <class
>>>>> 'gluster.exception.**GlusterVolumeCreateFailedExcep**tion'>, ())
>>>>>
>>>>> can you please run  gluster peer status on all your nodes ? also, it
>>>>> appears that '10.4.0.99' is problematic, can you try create the volume
>>>>> without it ?
>>>>>
>>>>>
>>>>>
>>>>>> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser <
>>>>>> domi...@bostonvineyard.org 
>>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org>>>
>>>>>> wrote:
>>>>>>
>>>>>>     Here are the other two logs forgot them.
>>>>>>
>>>>>>     dk
>>>>>>
>>>>>>
>>>>>>     On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
>>>>>>     <domi...@bostonvineyard.org 
>>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org>
>>>>>> >>
>>>>>>
>>>>>>     wrote:
>>>>>>
>>>>>>         Ok here are the logs 4 node and 1 engine log.  Tried making
>>>>>>         /data folder owned by root and then tried by 36:36 neither
>>>>>>         worked.  Name of volume is data to match folders on nodes
>>>>>> also.
>>>>>>
>>>>>>         Let me know what you think,
>>>>>>
>>>>>>         Dominic
>>>>>>
>>>>>>
>>>>>>         On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
>>>>>>         <ma...@wzzrd.com <mailto:ma...@wzzrd.com>> wrote:
>>>>>>
>>>>>>             I just ran into this as well, and it seems that you have
>>>>>>             to either reformat previously used gluster bricks or
>>>>>>             manually tweak some extended attributes.
>>>>>>
>>>>>>             Maybe this helps you in setting up your gluster volume,
>>>>>>             Dominic?
>>>>>>
>>>>>>             More info here:
>>>>>>             
>>>>>> https://bugzilla.redhat.com/**show_bug.cgi?id=812214<https://bugzilla.redhat.com/show_bug.cgi?id=812214>
>>>>>>
>>>>>>
>>>>>>             Maxim Burgerhout
>>>>>>             ma...@wzzrd.com <mailto:ma...@wzzrd.com>
>>>>>>
>>>>>>             ----------------
>>>>>>             EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>             On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
>>>>>>             <san...@redhat.com <mailto:san...@redhat.com>> wrote:
>>>>>>
>>>>>>                 Hi Dominic,
>>>>>>
>>>>>>                 Looking at the engine log immediately after trying to
>>>>>>                 create the volume should tell us on which node the
>>>>>>                 gluster volume creation was attempted. Then looking at
>>>>>>                 the vdsm log on that node should help us identifying
>>>>>>                 the exact reason for failure.
>>>>>>
>>>>>>                 In case this doesn't help you, can you please simulate
>>>>>>                 the issue again and send back all the 5 log files?
>>>>>>                 (engine.log from engine server and vdsm.log from the 4
>>>>>>                 nodes)
>>>>>>
>>>>>>                 Regards,
>>>>>>                 Shireesh
>>>>>>
>>>>>>
>>>>>>                 On Wednesday 05 September 2012 11:50 PM, Dominic
>>>>>>                 Kaiser wrote:
>>>>>>
>>>>>>>                 So I have a problem creating glusterfs volumes.  Here
>>>>>>>                 is the install:
>>>>>>>
>>>>>>>                  1. Ovirt 3.1
>>>>>>>                  2. 4 Nodes are Fedora 17 with kernel 3.3.4 -
>>>>>>>                     5.fc17.x86_64
>>>>>>>                  3. 4 nodes peer joined and running
>>>>>>>                  4. 4 nodes added as hosts to ovirt
>>>>>>>                  5. created a directory on each node this path /data
>>>>>>>                  6. chmod 36.36 -R /data all nodes
>>>>>>>                  7. went back to ovirt and created a
>>>>>>>
>>>>>>>                     distributed/replicated volume and added the 4
>>>>>>>                     nodes with brick path of /data
>>>>>>>
>>>>>>>                 I received this error:
>>>>>>>
>>>>>>>                 Creation of Gluster Volume maingfs1 failed.
>>>>>>>
>>>>>>>                 I went and looked at the vdsm logs on the nodes and
>>>>>>>                 the ovirt server which did not say much.  Where else
>>>>>>>                 should I look?  Also this error is vague what does it
>>>>>>>                 mean?
>>>>>>>
>>>>>>>
>>>>>>>                 --                 Dominic Kaiser
>>>>>>>                 Greater Boston Vineyard
>>>>>>>                 Director of Operations
>>>>>>>
>>>>>>>                 cell: 617-230-1412 <tel:617-230-1412>
>>>>>>>                 fax: 617-252-0238 <tel:617-252-0238>
>>>>>>>                 email: domi...@bostonvineyard.org
>>>>>>>                 
>>>>>>> <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org>
>>>>>>> >
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>                 ______________________________**_________________
>>>>>>>                 Users mailing list
>>>>>>>                 Users@ovirt.org  <mailto:Users@ovirt.org>
>>>>>>>                 
>>>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>>>
>>>>>>
>>>>>>
>>>>>>                 ______________________________**_________________
>>>>>>                 Users mailing list
>>>>>>                 Users@ovirt.org <mailto:Users@ovirt.org>
>>>>>>
>>>>>>                 
>>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>         --         Dominic Kaiser
>>>>>>         Greater Boston Vineyard
>>>>>>         Director of Operations
>>>>>>
>>>>>>         cell: 617-230-1412 <tel:617-230-1412>
>>>>>>         fax: 617-252-0238 <tel:617-252-0238>
>>>>>>         email: domi...@bostonvineyard.org
>>>>>>         <mailto:dominic@**bostonvineyard.org<domi...@bostonvineyard.org>
>>>>>> >
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>     --     Dominic Kaiser
>>>>>>     Greater Boston Vineyard
>>>>>>     Director of Operations
>>>>>>
>>>>>>     cell: 617-230-1412 <tel:617-230-1412>
>>>>>>     fax: 617-252-0238 <tel:617-252-0238>
>>>>>>     email: domi...@bostonvineyard.org <mailto:dominic@**
>>>>>> bostonvineyard.org <domi...@bostonvineyard.org>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Dominic Kaiser
>>>>>> Greater Boston Vineyard
>>>>>> Director of Operations
>>>>>>
>>>>>> cell: 617-230-1412
>>>>>> fax: 617-252-0238
>>>>>> email: domi...@bostonvineyard.org <mailto:dominic@**
>>>>>> bostonvineyard.org <domi...@bostonvineyard.org>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> ______________________________**_________________
>>>>>> Users mailing list
>>>>>> Users@ovirt.org
>>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Dominic Kaiser
>>>> Greater Boston Vineyard
>>>> Director of Operations
>>>>
>>>> cell: 617-230-1412
>>>> fax: 617-252-0238
>>>> email: domi...@bostonvineyard.org
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Dominic Kaiser
>>> Greater Boston Vineyard
>>> Director of Operations
>>>
>>> cell: 617-230-1412
>>> fax: 617-252-0238
>>> email: domi...@bostonvineyard.org
>>>
>>>
>>>
>>
>>
>> --
>> Dominic Kaiser
>> Greater Boston Vineyard
>> Director of Operations
>>
>> cell: 617-230-1412
>> fax: 617-252-0238
>> email: domi...@bostonvineyard.org
>>
>>
>>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412
> fax: 617-252-0238
> email: domi...@bostonvineyard.org
>
>
>


-- 
Dominic Kaiser
Greater Boston Vineyard
Director of Operations

cell: 617-230-1412
fax: 617-252-0238
email: domi...@bostonvineyard.org
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to