On Tue, Nov 25, 2008 at 8:03 AM, David Cournapeau <[EMAIL PROTECTED]>wrote:
> On Tue, Nov 25, 2008 at 10:55 PM, David Cournapeau > <[EMAIL PROTECTED]> wrote: > > > > > I used the path of least resistance: instead of using the > > WORDS_BIGENDIAN macro, I added a numpy header which gives the endianness > > every time it is included. IOW, instead of the endianness to be fixed at > > numpy build time (which would fail for universal builds), it is set > > everytime the numpy headers are included (which is the only way to make > > it work). A better solution IMO would be to avoid any endianness > > dependency at all in the headers, but that does not seem possible > > without breaking the API (because the endianness-related macro > > PyArray_NBO and co would need to be set as functions instead). > > Hm, for reference, I came across this: > > http://www.mail-archive.com/[EMAIL PROTECTED]/msg14382.html > > So some people thought about the same problem. > Apart from the Mac, the ppc can be configured to run either bigendian or littleendian, so the hardware encompasses more than just the cpu, it's the whole darn board. Chuck
_______________________________________________ Numpy-discussion mailing list [email protected] http://projects.scipy.org/mailman/listinfo/numpy-discussion
