Received from Keith Brown on Tue, Nov 17, 2015 at 08:17:28PM EST: > Hi, > > I have been using scikit-cudablas > (https://github.com/lebedov/scikit-cuda). It rocks! > > Does anyone have a 2d matrix multiplication example with DgemmBatched? > > Similar to, > > >>> a=np.random.randint(0,3,(16,2)); b=np.random.randint(0,4,(2,16)) > >>> np.dot(a,b)
Not sure I follow what you want to do - batched GEMM is intended for concurrent matrix multiplication of collections of matrices (effectively 3rd-order tensors). Do you want to obtain the products of the individual submatrices within the two matrices above, i.e., something like [np.dot(a[0:2,:], b[:,0:2]), np.dot(a[2:4,:], b[:,2:4]), ...]? > I have been using, > https://github.com/lebedov/scikit-cuda/blob/7e7300474286019c917a6c8a4bca59405c64fbce/tests/test_cublas.py#L531 > but it has too many dimensions and I keep getting confused by too many > dimensions for DgemmBatched -- Lev Givon Bionet Group | Neurokernel Project http://lebedov.github.io/ http://neurokernel.github.io/ _______________________________________________ PyCUDA mailing list PyCUDA@tiker.net http://lists.tiker.net/listinfo/pycuda