Is there another option? or a efficient workload balancing algorithm for this case? If so, please, let me know.
Thanks, Ed On Wed, Aug 13, 2008 at 9:55 PM, Edward J. Yoon <[EMAIL PROTECTED]> wrote: > Yes, but then, as the i grows, the task to workload ratio gets larger > and larger. Is It Right? > > -Edward > > On Wed, Aug 13, 2008 at 9:23 PM, Amar Kamat <[EMAIL PROTECTED]> wrote: >> Edward J. Yoon wrote: >>> >>> Hi communities, >>> >>> Do you have any idea how to get the pairs of all row key combinations >>> w/o repetition on Map/Reduce as describe below? >>> >>> Input : (MapFile or Hbase Table) >>> >>> <Key1, Value or RowResult> >>> <Key2, Value or RowResult> >>> <Key3, Value or RowResult> >>> <Key4, Value or RowResult> >>> >>> Output : >>> >>> <Key1, Key2> >>> <Key1, Key3> >>> <Key1, Key4> >>> <Key2, Key3> >>> <Key2, Key4> >>> <Key3, Key4> >>> >> >> One way to do it would be as follows >> For every key with index i, >> for (k=0; k < i; k++) { >> emit(i,key_i) >> } >> So the above input becomes >> 1,key1 >> 1,key1 >> 1,key1 >>> >>> It would be nice if someone can review my pseudo code of traditional >>> CF using cosine similarity. >>> http://wiki.apache.org/hama/TraditionalCollaborativeFiltering >>> >>> Thanks. >>> >> >> > > > > -- > Best regards, Edward J. Yoon > [EMAIL PROTECTED] > http://blog.udanax.org > -- Best regards, Edward J. Yoon [EMAIL PROTECTED] http://blog.udanax.org