Hmmm...nope, not a peep (no extra output at all). Just segfaulted like before.

On Sep 20, 2013, at 4:06 PM, Brice Goglin <brice.gog...@inria.fr> wrote:

> Try adding HWLOC_DEBUG_CHECK=1 in your environment, it will enable many 
> assertions at the end of hwloc_topology_load()
> 
> Brice
> 
> 
> 
> Le 21/09/2013 01:03, Ralph Castain a écrit :
>> I didn't try loading it with lstopo - just tried the OMPI trunk. It loads 
>> okay, but segfaults when you try to find an object by depth
>> 
>> #0  0x00000001005fe5dc in opal_hwloc172_hwloc_get_obj_by_depth 
>> (topology=Cannot access memory at address 0xfffffffffffffff7
>> ) at traversal.c:623
>> #1  0x0000000100b6dfaa in opal_hwloc172_hwloc_get_root_obj (topology=Cannot 
>> access memory at address 0xfffffffffffffff7
>> ) at rmaps_rr_mappers.c:747
>> #2  0x0000000100b6e139 in orte_rmaps_rr_byslot (jdata=Cannot access memory 
>> at address 0xffffffffffffff77
>> ) at rmaps_rr_mappers.c:774
>> #3  0x0000000100b6d6da in orte_rmaps_rr_map (jdata=Cannot access memory at 
>> address 0xffffffffffffff17
>> ) at rmaps_rr.c:211
>> #4  0x0000000100353098 in orte_rmaps_base_map_job (fd=Cannot access memory 
>> at address 0xfffffffffffffe7b
>> ) at base/rmaps_base_map_job.c:320
>> #5  0x00000001005ce28c in event_process_active_single_queue (base=Cannot 
>> access memory at address 0xffffffffffffffe7
>> ) at event.c:1367
>> #6  0x00000001005ce500 in event_process_active (base=Cannot access memory at 
>> address 0xffffffffffffffe7
>> ) at event.c:1437
>> #7  0x00000001005ceb71 in opal_libevent2021_event_base_loop (base=Cannot 
>> access memory at address 0xffffffffffffffb7
>> ) at event.c:1645
>> #8  0x00000001002c5158 in orterun (argc=Cannot access memory at address 
>> 0xfffffffffffffd1b
>> ) at orterun.c:3039
>> #9  0x00000001002c32a4 in main (argc=Cannot access memory at address 
>> 0xfffffffffffffffb
>> ) at main.c:14
>> 
>> Looks to me like memory may be getting hosed
>> 
>> 
>> On Sep 20, 2013, at 2:59 PM, Brice Goglin <brice.gog...@inria.fr> wrote:
>> 
>>> I can't see any segfault. Where does the segfault occurs for you? In OMPI 
>>> only (or lstopo too)? When loading or when using the topology?
>>> 
>>> I tried lstopo on that file with and without HWLOC_NO_LIBXML_IMPORT=1 (in 
>>> case the bug is in one of XML backends), looks ok.
>>> 
>>> Brice
>>> 
>>> 
>>> 
>>> 
>>> 
>>> Le 20/09/2013 23:53, Ralph Castain a écrit :
>>>> Here are the two files I tried - not from the same machine. The foo.xml 
>>>> works, the topo.xml segfaults
>>>> 
>>>> 
>>>> 
>>>> 
>>>> One of our users reported it from their machine, but I don't have their 
>>>> topo file.
>>>> 
>>>> On Sep 20, 2013, at 2:41 PM, Brice Goglin <brice.gog...@inria.fr> wrote:
>>>> 
>>>>> Hello,
>>>>> I don't see anything reason for such an incompatibility. But there are
>>>>> many combinations, we can't test everything.
>>>>> I can't reproduce that on my machines. Can you send the XML output of
>>>>> both versions on one of your machines?
>>>>> Brice
>>>>> 
>>>>> 
>>>>> 
>>>>> Le 20/09/2013 23:32, Ralph Castain a écrit :
>>>>>> Hi folks
>>>>>> 
>>>>>> I've run across a rather strange behavior. We have two branches in OMPI 
>>>>>> - the devel trunk (using hwloc v1.7.2) and our feature release series 
>>>>>> (using hwloc 1.5.2). I have found the following:
>>>>>> 
>>>>>> *the feature series can correctly load an xml file generated by lstopo 
>>>>>> of versions 1.5 or greater
>>>>>> 
>>>>>> * the devel series can correctly load an xml file generated by lstopo of 
>>>>>> versions 1.7 or greater, but not files generated by prior versions. In 
>>>>>> the latter case, I segfault as soon as I try to use the loaded topology.
>>>>>> 
>>>>>> Any ideas why the discrepancy? Can I at least detect the version used to 
>>>>>> create a file when loading it so I can error out instead of segfaulting?
>>>>>> 
>>>>>> Ralph
>>>>>> 
>>>>>> _______________________________________________
>>>>>> hwloc-devel mailing list
>>>>>> hwloc-de...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
>>>>> _______________________________________________
>>>>> hwloc-devel mailing list
>>>>> hwloc-de...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
>>>> 
>>>> 
>>>> _______________________________________________
>>>> hwloc-devel mailing list
>>>> hwloc-de...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
>>> 
>>> _______________________________________________
>>> hwloc-devel mailing list
>>> hwloc-de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
>> 
>> 
>> 
>> _______________________________________________
>> hwloc-devel mailing list
>> hwloc-de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
> 
> _______________________________________________
> hwloc-devel mailing list
> hwloc-de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel

Reply via email to