On Thu, May 5, 2011 at 9:18 PM, Benjamin Root <ben.r...@ou.edu> wrote:

>
>
> On Thu, May 5, 2011 at 1:08 PM, Paul Anton Letnes <
> paul.anton.let...@gmail.com> wrote:
>
>>
>> On 5. mai 2011, at 08.49, Benjamin Root wrote:
>>
>> >
>> >
>> > On Wed, May 4, 2011 at 11:08 PM, Paul Anton Letnes <
>> paul.anton.let...@gmail.com> wrote:
>> >
>> > On 4. mai 2011, at 20.33, Benjamin Root wrote:
>> >
>> > > On Wed, May 4, 2011 at 7:54 PM, Derek Homeier <
>> de...@astro.physik.uni-goettingen.de> wrote:
>> > > On 05.05.2011, at 2:40AM, Paul Anton Letnes wrote:
>> > >
>> > > > But: Isn't the numpy.atleast_2d and numpy.atleast_1d functions
>> written for this? Shouldn't we reuse them? Perhaps it's overkill, and
>> perhaps it will reintroduce the 'transposed' problem?
>> > >
>> > > Yes, good point, one could replace the
>> > > X.shape = (X.size, ) with X = np.atleast_1d(X),
>> > > but for the ndmin=2 case, we'd need to replace
>> > > X.shape = (X.size, 1) with X = np.atleast_2d(X).T -
>> > > not sure which solution is more efficient in terms of memory access
>> etc...
>> > >
>> > > Cheers,
>> > >                                                Derek
>> > >
>> > >
>> > > I can confirm that the current behavior is not sufficient for all of
>> the original corner cases that ndmin was supposed to address.  Keep in mind
>> that np.loadtxt takes a one-column data file and a one-row data file down to
>> the same shape.  I don't see how the current code is able to produce the
>> correct array shape when ndmin=2.  Do we have some sort of counter in
>> loadtxt for counting the number of rows and columns read?  Could we use
>> those to help guide the ndmin=2 case?
>> > >
>> > > I think that using atleast_1d(X) might be a bit overkill, but it would
>> be very clear as to the code's intent.  I don't think we have to worry about
>> memory usage if we limit its use to only situations where ndmin is greater
>> than the number of dimensions of the array.  In those cases, the array is
>> either an empty result, a scalar value (in which memory access is trivial),
>> or 1-d (in which a transpose is cheap).
>> >
>> > What if one does things the other way around - avoid calling squeeze
>> until _after_ doing the atleast_Nd() magic? That way the row/column
>> information should be conserved, right? Also, we avoid transposing, memory
>> use, ...
>> >
>> > Oh, and someone could conceivably have a _looong_ 1D file, but would
>> want it read as a 2D array.
>> >
>> > Paul
>> >
>> >
>> >
>> > @Derek, good catch with noticing the error in the tests. We do still
>> need to handle the case I mentioned, however.  I have attached an example
>> script to demonstrate the issue.  In this script, I would expect the
>> second-to-last array to be a shape of (1, 5).  I believe that the
>> single-row, multi-column case would actually be the more common type of
>> edge-case encountered by users than the others.  Therefore, I believe that
>> this ndmin fix is not adequate until this is addressed.
>> >
>> > @Paul, we can't call squeeze after doing the atleast_Nd() magic.  That
>> would just undo whatever we had just done.  Also, wrt the transpose, a (1,
>> 100000) array looks the same in memory as a (100000, 1) array, right?
>> Agree. I thought more along the lines of (pseudocode-ish)
>> if ndmin == 0:
>>        squeeze()
>> if ndmin == 1:
>>        atleast_1D()
>> elif ndmin == 2:
>>        atleast_2D()
>> else:
>>        I don't rightly know what would go here, maybe raise ValueError?
>>
>> That would avoid the squeeze call before the atleast_Nd magic. But the
>> code was changed, so I think my comment doesn't make sense anymore. It's
>> probably fine the way it is!
>>
>> Paul
>>
>>
> I have thought of that too, but the problem with that approach is that
> after reading the file, X will have 2 or 3 dimensions, regardless of how
> many singleton dims were in the file.  A squeeze will always be needed.
> Also, the purpose of squeeze is opposite that of the atleast_*d()
> functions:  squeeze reduces dimensions, while atleast_*d will add
> dimensions.
>
> Therefore, I re-iterate... the patch by Derek gets the job done.  I have
> tested it for a wide variety of inputs for both regular arrays and record
> arrays.  Is there room for improvements?  Yes, but I think that can wait for
> later.  Derek's patch however fixes an important bug in the ndmin
> implementation and should be included for the release.
>
> Two questions: can you point me to the patch/ticket, and is this a
regression?

Thanks,
Ralf
_______________________________________________
NumPy-Discussion mailing list
NumPy-Discussion@scipy.org
http://mail.scipy.org/mailman/listinfo/numpy-discussion

Reply via email to