Hi,

On Thu, Apr 24, 2014 at 5:26 PM,  <josef.p...@gmail.com> wrote:
>
>
>
> On Thu, Apr 24, 2014 at 7:29 PM, <josef.p...@gmail.com> wrote:
>>
>>
>>
>>
>> On Thu, Apr 24, 2014 at 7:20 PM, Charles R Harris
>> <charlesr.har...@gmail.com> wrote:
>>>
>>>
>>>
>>>
>>> On Thu, Apr 24, 2014 at 5:08 PM, <josef.p...@gmail.com> wrote:
>>>>
>>>>
>>>>
>>>>
>>>> On Thu, Apr 24, 2014 at 7:00 PM, Charles R Harris
>>>> <charlesr.har...@gmail.com> wrote:
>>>>>
>>>>>
>>>>> Hi Matthew,
>>>>>
>>>>> On Thu, Apr 24, 2014 at 3:56 PM, Matthew Brett
>>>>> <matthew.br...@gmail.com> wrote:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Thanks to Cark Kleffner's toolchain and some help from Clint Whaley
>>>>>> (main author of ATLAS), I've built 64-bit windows numpy and scipy
>>>>>> wheels for testing.
>>>>>>
>>>>>> The build uses Carl's custom mingw-w64 build with static linking.
>>>>>>
>>>>>> There are two harmless test failures on scipy (being discussed on the
>>>>>> list at the moment) - tests otherwise clean.
>>>>>>
>>>>>> Wheels are here:
>>>>>>
>>>>>>
>>>>>> https://nipy.bic.berkeley.edu/scipy_installers/numpy-1.8.1-cp27-none-win_amd64.whl
>>>>>>
>>>>>> https://nipy.bic.berkeley.edu/scipy_installers/scipy-0.13.3-cp27-none-win_amd64.whl
>>>>>>
>>>>>> You can test with:
>>>>>>
>>>>>> pip install -U pip # to upgrade pip to latest
>>>>>> pip install -f https://nipy.bic.berkeley.edu/scipy_installers numpy
>>>>>> scipy
>>>>>>
>>>>>> Please do send feedback.
>>>>>>
>>>>>> ATLAS binary here:
>>>>>>
>>>>>>
>>>>>> https://nipy.bic.berkeley.edu/scipy_installers/atlas_builds/atlas-64-full-sse2.tar.bz2
>>>>>>
>>>>>> Many thanks for Carl in particular for doing all the hard work,
>>>>>>
>>>>>
>>>>> Cool. After all these long years... Now all we need is a box running
>>>>> tests for CI.
>>>>>
>>>>> Chuck
>>>>>
>>>>> _______________________________________________
>>>>> NumPy-Discussion mailing list
>>>>> NumPy-Discussion@scipy.org
>>>>> http://mail.scipy.org/mailman/listinfo/numpy-discussion
>>>>>
>>>>
>>>> I get two test failures with numpy
>>>>
>>>> Josef
>>>>
>>>> >>> np.test()
>>>> Running unit tests for numpy
>>>> NumPy version 1.8.1
>>>> NumPy is installed in C:\Python27\lib\site-packages\numpy
>>>> Python version 2.7.3 (default, Apr 10 2012, 23:24:47) [MSC v.1500 64 bit
>>>> (AMD64)]
>>>> nose version 1.1.2
>>>>
>>>> ======================================================================
>>>> FAIL: test_iterator.test_iter_broadcasting_errors
>>>> ----------------------------------------------------------------------
>>>> Traceback (most recent call last):
>>>>   File "C:\Python27\lib\site-packages\nose\case.py", line 197, in
>>>> runTest
>>>>     self.test(*self.arg)
>>>>   File
>>>> "C:\Python27\lib\site-packages\numpy\core\tests\test_iterator.py", line 
>>>> 657,
>>>> in test_iter_broadcasting_errors
>>>>     '(2)->(2,newaxis)') % msg)
>>>>   File "C:\Python27\lib\site-packages\numpy\testing\utils.py", line 44,
>>>> in assert_
>>>>     raise AssertionError(msg)
>>>> AssertionError: Message "operands could not be broadcast together with
>>>> remapped shapes [original->remapped]: (2,3)->(2,3) (2,)->(2,newaxis) and
>>>> requested shape (4,3)" doesn't contain remapped operand
>>>> shape(2)->(2,newaxis)
>>>>
>>>> ======================================================================
>>>> FAIL: test_iterator.test_iter_array_cast
>>>> ----------------------------------------------------------------------
>>>> Traceback (most recent call last):
>>>>   File "C:\Python27\lib\site-packages\nose\case.py", line 197, in
>>>> runTest
>>>>     self.test(*self.arg)
>>>>   File
>>>> "C:\Python27\lib\site-packages\numpy\core\tests\test_iterator.py", line 
>>>> 836,
>>>> in test_iter_array_cast
>>>>     assert_equal(i.operands[0].strides, (-96,8,-32))
>>>>   File "C:\Python27\lib\site-packages\numpy\testing\utils.py", line 255,
>>>> in assert_equal
>>>>     assert_equal(actual[k], desired[k], 'item=%r\n%s' % (k, err_msg),
>>>> verbose)
>>>>   File "C:\Python27\lib\site-packages\numpy\testing\utils.py", line 317,
>>>> in assert_equal
>>>>     raise AssertionError(msg)
>>>> AssertionError:
>>>> Items are not equal:
>>>> item=0
>>>>
>>>>  ACTUAL: 96L
>>>>  DESIRED: -96
>>>>
>>>> ----------------------------------------------------------------------
>>>> Ran 4828 tests in 46.306s
>>>>
>>>> FAILED (KNOWNFAIL=10, SKIP=8, failures=2)
>>>> <nose.result.TextTestResult run=4828 errors=0 failures=2>
>>>>
>>>
>>> Strange. That second one looks familiar, at least the "-96" part. Wonder
>>> why this doesn't show up with the MKL builds.
>>
>>
>> ok tried again, this time deleting the old numpy directories before
>> installing
>>
>> Ran 4760 tests in 42.124s
>>
>> OK (KNOWNFAIL=10, SKIP=8)
>> <nose.result.TextTestResult run=4760 errors=0 failures=0>
>>
>>
>> so pip also seems to be reusing leftover files.
>>
>> all clear.
>
>
> Running the statsmodels test suite, I get a failure in
> test_discrete.TestProbitCG where fmin_cg converges to something that differs
> in the 3rd decimal.
>
> I usually only test the 32-bit version, so I don't know if this is specific
> to this scipy version, but we haven't seen this in a long time.
> I used our nightly binaries http://statsmodels.sourceforge.net/binaries/

That's interesting, you saw also we're getting failures on the tests
for powell optimization because of small unit-at-last-place
differences in the exp function in mingw-w64.  Is there any chance you
can track down where the optimization path is diverging and why?
It's just that - if this is also the exp function maybe we can see if
the error is exceeding reasonable bounds and then feed back to
mingw-w64 and fall back to the numpy default implementation in the
meantime.

Cheers,

Matthew
_______________________________________________
NumPy-Discussion mailing list
NumPy-Discussion@scipy.org
http://mail.scipy.org/mailman/listinfo/numpy-discussion

Reply via email to