Hi Oleg I will answer your questions one by one.
1) file size There is no exactly number of file size that will definitely works well for GPGPU+Hadoop. You need to do your project POC to get the number. I think the GPU+Hadoop is very suitable for computation-intensive and data-intensive applications. However, be aware of the bottleneck between the GPU memory and CPU memory. I mean the benefit you obtained from using GPGPU should be larger than the performance that you sacrificed by shipping data between GPU memory and CPU memory. If you only have computation-intensive applications and can be parallelized by GPGPU, CUDA+Hadoop can also provide a parallel framework for you to distribute your work among the cluster nodes with fault-tolerance. 2) Is it good Idea to process data as locally as possble (I mean process a data like one file per one map) Local Map tasks are shorter than non-local tasks in the Hadoop MapReduce framework. 3) During your project did you face with limitations , problems? During my project, the video card was not fancy, it only allowed one CUDA program using the card in anytime. Then, we only configured one map slot and one reduce slot in a cluster node. Now, nvidia has some powerful products that support multiple program run on the same card simultaneously. 4) By the way I didn't fine code Jcuda example with Hadoop. :-) Your MapReduce code is written in Java, right? Integrate your Jcude code to either map() or reduce() method of your MapReduce code (you can also do this in the combiner, partitioner or whatever you need). Jcuda example only helps you know how Jcuda works. Chen On Mon, Sep 24, 2012 at 11:22 AM, Oleg Ruchovets <oruchov...@gmail.com>wrote: > Great , > Can you give some tips or best practices like: > 1) file size > 2) Is it good Idea to process data as locally as possble (I mean process a > data like one file per one map) > 3) During your project did you face with limitations , problems? > > > Can you point me on which hartware is better to use( I understand in > order to use GPU I need NVIDIA) . > I mean using CPU only arthitecture I have 8-12 core per one computer(for > example). > What should I do in orger to use CPU+GPU arthitecture? What kind of NVIDIA > do I need for this. > > By the way I didn't fine code Jcuda example with Hadoop. :-) > > Thanks in advane > Oleg. > > On Mon, Sep 24, 2012 at 6:07 PM, Chen He <airb...@gmail.com> wrote: > > > Please see the Jcuda example. I do refer from there. BTW, you can also > > compile your cuda code in advance and let your hadoop code call those > > compiled code through Jcuda. That is what I did in my program. > > > > On Mon, Sep 24, 2012 at 10:45 AM, Oleg Ruchovets <oruchov...@gmail.com > > >wrote: > > > > > Thank you very much. I saw this link !!! . Do you have any code , > > example > > > shared in the network (github for example). > > > > > > On Mon, Sep 24, 2012 at 5:33 PM, Chen He <airb...@gmail.com> wrote: > > > > > > > http://wiki.apache.org/hadoop/CUDA%20On%20Hadoop > > > > > > > > On Mon, Sep 24, 2012 at 10:30 AM, Oleg Ruchovets < > oruchov...@gmail.com > > > > >wrote: > > > > > > > > > Hi > > > > > > > > > > I am going to process video analytics using hadoop > > > > > I am very interested about CPU+GPU architercute espessially using > > CUDA > > > ( > > > > > http://www.nvidia.com/object/cuda_home_new.html) and JCUDA ( > > > > > http://jcuda.org/) > > > > > Does using HADOOP and CPU+GPU architecture bring significant > > > performance > > > > > improvement and does someone succeeded to implement it in > production > > > > > quality? > > > > > > > > > > I didn't fine any projects / examples to use such technology. > > > > > If someone could give me a link to best practices and example using > > > > > CUDA/JCUDA + hadoop that would be great. > > > > > Thanks in advane > > > > > Oleg. > > > > > > > > > > > > > > >