Just to clarify my position, performance is a secondary concern, primarily I'm concerned about intuitiveness and convenience. I suspect an immutable vector type will cause a regular amount of mail traffic asking how to mutate them and why that is not possible. I could be wrong, but I think many folks will be surprised by an immutable vector class.

I'm +0 on immutable vectors, they're certainly much better than nothing, but I personally would prefer mutable.

To make immutable types more performant, you should look at the implementation of tuples in Python. tuple instances are pooled so they can be recycled, saving a lot of memory management overhead. I did a similar thing for vectors in Lepton with good results.

-Casey

On May 4, 2009, at 12:09 PM, Lorenz Quack wrote:

Hi all,

I think this discussion got a little bit out of hand.
I probably shouldn't have posted those numbers.
I would like to get this thread back on track.
But after that I shortly want to answer the questions from Gregor's last mail and append the relevant sctions of my code.

so here we go "mutable or immutable":

So if I may try to summarize:

mutable:
========
* Stuart Axon: mainly consistency with rects
* René Dudfield: personally uses list more often than tuples
* Casey Duncan: consistency with rects and performance concerns

immutable:
==========
* Brian Fisher: immutable prevent subtle bugs
* Marcus von Appen: no reason given
* Gregor Lingl: should behave more like numbers than like list
* Lorenz Quack: personally thinks the presented argument for immutable are stronger

So would anyone have strong objections if we go with immutable?
Vectors would then behave more like a mix of floats and tuples :)



what follows is the response to the Gregor:

Gregor Lingl wrote:
Lorenz Quack schrieb:
>>> a = 2
>>> a += a
I believe the interpreter internally takes the two operands (in this case a and a) adds them and then rebinds the result to a (the id changes) so effectively doing
>>> a = a + a
exactly because a is immutable
if a were mutable the two expressions would be indeed different the += version would not create a new instance and rebind the name a to it but modify the object a is referring to, while a = a + a would again create a new object and rebind it.
So therefore I believe that this test does make sense.
Tell me if I'm wrong somewhere.


here are the calls with the results:


[snip]


which has more than 30000 digits. Which result did you get after 10000000 executions of the statement x = x + x?

And which implementation of the long integer type did you use that is that much faster than Python's ?

Regards,
Gregor


indeed those are valid objections. well first of all I used a self- written C extension with double as the underlying type. but the result after 1023 iterations turns into (inf, inf). this could of course invalidate the results so I modified the test:

>>> timeit.repeat("x = Vector2d(2,3); x += x", "from vector import Vector2d", repeat=5, number=10000000)
[5.1832518577575684,
5.1106431484222412,
5.1510121822357178,
5.0923140048980713,
5.0608019828796387]

>>> timeit.repeat("x = Vector2d(2,3); x = x + x", "from vector import Vector2d", repeat=5, number=10000000)
[6.5348029136657715,
6.3499071598052979,
6.4433431625366211,
6.412431001663208,
6.4398849010467529]

>>> timeit.repeat("x = Vector2d(2,3)", "from vector import Vector2d", repeat=5, number=10000000)
[3.7264928817749023,
3.6346859931945801,
3.6241021156311035,
3.7733709812164307,
3.6264529228210449]

Did you use two different Vector2d classes here, one mutable and one immutable? Why do they have the same name then? Or did you merely implement the operations x+=x and x=x+x differently?

The latter. Same class just once use the "nb_add" callback (or from a python persprctive: "__add__") and once the "nb_inplace_add" callback (or again from python: "__iadd__")

If x = x + y creates a new object x or changes x is also a matter of how it is implemented.

not really. when the "nb_add" C callback (or "__add__" for that matter) is called you have no way of knowing what the caller is going to do with the result. so from inside that callback you cannot distinguish between
>>> x = x + y
and
>>> z = x + y
so you really don't have a choice but to return a new object.

Moreover it is my conviction that one must not decide about which data type to use on
the basis of a +- 50 percent difference in performance.

ok.

One more remark: At least on module of the standard library of Python has a (rather simple) 2d-Vector class implemented in pure Python, which of course has a considerably worse performance,
by a factor of 4 approximately:
>>> timeit.repeat("x = Vec2D(2,3); x = x + x", "from turtle import Vec2D", repeat=1, number=10000000)
[25.274672320512536]
Nevertheless one would expect a class implemented in C to run *much* faster than a pure Python solution. So I suspect that your implementation may not be sufficiently significant to serve as a criterion to
decide that issue.
Best regards,
Gregor
.

so here comes the boild down version of my code:

#define PyVector2d_Check(v)  PyObject_TypeCheck(v, &PyVector2d_Type)
#define PyVector3d_Check(v)  PyObject_TypeCheck(v, &PyVector3d_Type)
#define PyVector4d_Check(v)  PyObject_TypeCheck(v, &PyVector4d_Type)
#define PyVectorNd_Check(v) (PyVector4d_Check(v) || PyVector3d_Check(v) || PyVector2d_Check(v))

static PyObject *
PyVectorNd_add(PyObject *o1, PyObject *o2)
{
   int i;
   if (PyVectorNd_Check(o1)) {
       int dim = ((PyVectorNd*)o1)->dim;
       if (checkPyVectorNdCompatible(o2, dim)) {
           PyVectorNd *ret = (PyVectorNd*)PyVectorNd_NEW(dim);
           for (i = 0; i < dim; i++) {
ret->data[i] = ((PyVectorNd*)o1)->data[i] + PySequence_GetItem_AsDouble(o2, i);
           }
           return (PyObject*)ret;
       }
   }
   else {
       int dim = ((PyVectorNd*)o2)->dim;
       if (checkPyVectorNdCompatible(o1, dim)) {
           PyVectorNd *ret = (PyVectorNd*)PyVectorNd_NEW(dim);
           for (i = 0; i < dim; i++) {
ret->data[i] = PySequence_GetItem_AsDouble(o1, i) + ((PyVectorNd*)o2)->data[i];
           }
           return (PyObject*)ret;
       }
   }

   Py_INCREF(Py_NotImplemented);
   return Py_NotImplemented;
}

static PyObject *
PyVectorNd_inplace_add(PyVectorNd *self, PyObject *other)
{
   int i;
   if (checkPyVectorNdCompatible(other, self->dim)) {
       for (i = 0; i < self->dim; i++) {
           self->data[i] += PySequence_GetItem_AsDouble(other, i);
       }
       Py_INCREF(self);
       return (PyObject*)self;
   }
   Py_INCREF(Py_NotImplemented);
   return Py_NotImplemented;
}


static PyObject *
PyVectorNd_NEW(int dim)
{
   PyVectorNd *object;
   switch (dim) {
   case 2:
       object = PyObject_New(PyVectorNd, &PyVector2d_Type);
       break;
   case 3:
       object = PyObject_New(PyVectorNd, &PyVector3d_Type);
       break;
   case 4:
       object = PyObject_New(PyVectorNd, &PyVector4d_Type);
       break;
   default:
fprintf(stderr, "Error: wrong internal call to PyVectorNd_NEW. \n");
       exit(1);
   }

   if (object != NULL) {
       object->dim = dim;
       object->epsilon = FLT_EPSILON;
       object->data = PyMem_New(double, dim);
       if (object->data == NULL) {
           return PyErr_NoMemory();
       }
   }
   else {
fprintf(stderr, "FAILURE: could not create new PyVectorNd object!\n");
   }
   return (PyObject *)object;
}


Note that the main difference between the two (PyVectorNd_add and PyVectorNd_inplace_add) in this case is mainly a call to PyVectorNd_Check and PyVectorNd_NEW.

And again: I'm not really here to discuss this particular code or look for optimizations.


regards,
//Lorenz

Reply via email to