Dr. Rajen Bhatt, I don't now of any example code that will split a video file for Hadoop. If you write one then please start a git-hub project or something like it and let us know because I can see it being extremely useful for others wanting to do multimedia processing.
You might want to look at ffmpeg(libavformat/libavcodec) I am not sure how this would work completely, but in the splitter you could open up the file and get out the file's length in total time or frames and divided it up per mapper, then in each mapper have libavformat/libavcodec open up the file again and seek to their start time or frame. I am not sure how efficient libavformat/libavcodec are in seeking. It would be bad if they actually read the entire file to get to a given point. It would also be good if the in the splitter you could try and get what the offset into the file is for a given time or frame so that you can give hits to Hadoop to know where to place the mapper. This also assumes that you can set up something to bridge HDFS to libavformat probably using JNI. --Bobby Evans On 9/20/11 11:12 PM, "Rajen Bhatt (RBEI/EST1)" <rajen.bh...@in.bosch.com> wrote: Dear Bobby: Thanks for your reply. My problem is that I have just one file, which is very large typically in terabytes or petabytes and then I want to split them, feed to mappers, and process on worker nodes. Is some example class is available with somebody, it shall help the student. Thanks and Regards, ~~ Dr. Rajen Bhatt (Corporate Research @ Robert Bosch, India) Off: +91-80-4191-6699 Mob: +91-9901241005 [cid:3399453662_6132054] From: Robert Evans [mailto:ev...@yahoo-inc.com] Sent: Tuesday, 20. September 2011 11:32 PM To: mapreduce-user@hadoop.apache.org Subject: Re: Using HADOOP for Processing Videos Another thing to think about is that you may not need to split the videos at all. If you have lots of video files instead of a few big ones, and each can more or less be processed independently then you can use something like nline input format, not nline itself necessarily but something like it, to process each video separately. You would have to write the code to read in the video file, but there are APIs to do that, like OpenCV. This is what I did in the past to train and score machine learned classifiers on image and video files using Hadoop. --Bobby Evans On 9/19/11 11:54 PM, "Swathi V" <swat...@zinniasystems.com> wrote: This link might help you... example <http://musicmachinery.com/2011/09/04/how-to-process-a-million-songs-in-20-minutes/> On Tue, Sep 20, 2011 at 9:52 AM, Rajen Bhatt (RBEI/EST1) <rajen.bh...@in.bosch.com> wrote: Dear MapReduce User Groups: We want to process large amount of videos (typically 30 days old storage with size around 1TB) using Hadoop. Can somebody point me to code samples or classes which can take video files in its original compressed format (H.264, MPEG-4) and then process using Mappers? Thanks and Regards, ~~ Dr. Rajen Bhatt (Corporate Research @ Robert Bosch, India) Off: +91-80-4191-6699 Mob: +91-9901241005 [cid:3399453662_6130082]
<<inline: image.jpg>>
<<inline: image.jpg>>