Barry, Yes. It improves the performance dramatically, but the execution time for KSPSolve stays the same.
MatMult 5.2 Gflops Thanks, ================================ ?Keita Teranishi ?Scientific Library Group ?Cray, Inc. ?keita at cray.com ================================ -----Original Message----- From: petsc-dev-bounces at mcs.anl.gov [mailto:petsc-dev-boun...@mcs.anl.gov] On Behalf Of Barry Smith Sent: Friday, August 27, 2010 2:15 PM To: For users of the development version of PETSc Subject: [petsc-dev] [GPU] Performance on Fermi PETSc-dev folks, Please prepend all messages to petsc-dev that involve GPUs with [GPU] so they can be easily filtered. Keita, To run src/ksp/ksp/examples/tutorials/ex2.c with CUDA you need the flag -vec_type cuda Note also that this example is fine for simple ONE processor tests but should not be used for parallel testing because it does not do a proper parallel partitioning for performance Barry On Aug 27, 2010, at 2:04 PM, Keita Teranishi wrote: > Hi, > > I ran ex2.c with a matrix from 512x512 grid. > I set CG and Jacobi for the solver and preconditioner. > GCC-4.4.4 and CUDA-3.1 are used to compile the code. > BLAS and LAPAKCK are not optimized. > > MatMult > Fermi: 1142 MFlops > 1 core Istanbul: 420 MFlops > > KSPSolve: > Fermi: 1.5 Sec > 1 core Istanbul: 1.7 Sec > > > ================================ > Keita Teranishi > Scientific Library Group > Cray, Inc. > keita at cray.com > ================================ > > > -----Original Message----- > From: petsc-dev-bounces at mcs.anl.gov [mailto:petsc-dev-bounces at > mcs.anl.gov] On Behalf Of Satish Balay > Sent: Friday, August 27, 2010 1:49 PM > To: For users of the development version of PETSc > Subject: Re: [petsc-dev] Problem with petsc-dev > > On Fri, 27 Aug 2010, Satish Balay wrote: > >> There was a problem with tarball creation for the past few days. Will >> try to respin manually today - and update you. > > the petsc-dev tarball is now updated on the website.. > > Satish -------------- next part -------------- A non-text attachment was scrubbed... Name: out_fermi Type: application/octet-stream Size: 10179 bytes Desc: out_fermi URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20100827/b6b29cf3/attachment.obj>