Is there another option? or a efficient workload balancing algorithm
for this case? If so, please, let me know.

Thanks, Ed

On Wed, Aug 13, 2008 at 9:55 PM, Edward J. Yoon <[EMAIL PROTECTED]> wrote:
> Yes, but then, as the i grows, the task to workload ratio gets larger
> and larger. Is It Right?
>
> -Edward
>
> On Wed, Aug 13, 2008 at 9:23 PM, Amar Kamat <[EMAIL PROTECTED]> wrote:
>> Edward J. Yoon wrote:
>>>
>>> Hi communities,
>>>
>>> Do you have any idea how to get the pairs of all row key combinations
>>> w/o repetition on Map/Reduce as describe below?
>>>
>>> Input : (MapFile or Hbase Table)
>>>
>>> <Key1, Value or RowResult>
>>> <Key2, Value or RowResult>
>>> <Key3, Value or RowResult>
>>> <Key4, Value or RowResult>
>>>
>>> Output :
>>>
>>> <Key1, Key2>
>>> <Key1, Key3>
>>> <Key1, Key4>
>>> <Key2, Key3>
>>> <Key2, Key4>
>>> <Key3, Key4>
>>>
>>
>> One way to do it would be as follows
>> For every key with index i,
>> for (k=0; k < i; k++) {
>> emit(i,key_i)
>> }
>> So the above input becomes
>> 1,key1
>> 1,key1
>> 1,key1
>>>
>>> It would be nice if someone can review my pseudo code of traditional
>>> CF using cosine similarity.
>>> http://wiki.apache.org/hama/TraditionalCollaborativeFiltering
>>>
>>> Thanks.
>>>
>>
>>
>
>
>
> --
> Best regards, Edward J. Yoon
> [EMAIL PROTECTED]
> http://blog.udanax.org
>



-- 
Best regards, Edward J. Yoon
[EMAIL PROTECTED]
http://blog.udanax.org

Reply via email to