Received from Keith Brown on Tue, Nov 17, 2015 at 08:17:28PM EST:
> Hi,
> 
> I have been using scikit-cudablas
> (https://github.com/lebedov/scikit-cuda). It rocks!
> 
> Does anyone have a 2d matrix multiplication example with DgemmBatched?
> 
> Similar to,
> 
> >>> a=np.random.randint(0,3,(16,2)); b=np.random.randint(0,4,(2,16))
> >>> np.dot(a,b)

Not sure I follow what you want to do - batched GEMM is intended for concurrent
matrix multiplication of collections of matrices (effectively 3rd-order
tensors). Do you want to obtain the products of the individual submatrices 
within
the two matrices above, i.e., something like [np.dot(a[0:2,:], b[:,0:2]), 
np.dot(a[2:4,:],
b[:,2:4]), ...]?

> I have been using,
> https://github.com/lebedov/scikit-cuda/blob/7e7300474286019c917a6c8a4bca59405c64fbce/tests/test_cublas.py#L531
> but it has too many dimensions and I keep getting confused by too many
> dimensions for DgemmBatched
-- 
Lev Givon
Bionet Group | Neurokernel Project
http://lebedov.github.io/
http://neurokernel.github.io/


_______________________________________________
PyCUDA mailing list
PyCUDA@tiker.net
http://lists.tiker.net/listinfo/pycuda

Reply via email to