On 01/14/2014 01:05 PM, Minchan Kim wrote: > On Tue, Jan 14, 2014 at 01:50:22PM +0900, Minchan Kim wrote: >> Hello Bob, >> >> On Tue, Jan 14, 2014 at 09:19:23AM +0800, Bob Liu wrote: >>> >>> On 01/14/2014 07:35 AM, Minchan Kim wrote: >>>> Hello, >>>> >>>> On Sat, Jan 11, 2014 at 03:43:07PM +0800, Cai Liu wrote: >>>>> zswap can support multiple swapfiles. So we need to check >>>>> all zbud pool pages in zswap. >>>> >>>> True but this patch is rather costly that we should iterate >>>> zswap_tree[MAX_SWAPFILES] to check it. SIGH. >>>> >>>> How about defining zswap_tress as linked list instead of static >>>> array? Then, we could reduce unnecessary iteration too much. >>>> >>> >>> But if use linked list, it might not easy to access the tree like this: >>> struct zswap_tree *tree = zswap_trees[type]; >> >> struct zswap_tree { >> .. >> .. >> struct list_head list; >> } >> >> zswap_frontswap_init() >> { >> .. >> .. >> zswap_trees[type] = tree; >> list_add(&tree->list, &zswap_list); >> } >> >> get_zswap_pool_pages(void) >> { >> struct zswap_tree *cur; >> list_for_each_entry(cur, &zswap_list, list) { >> pool_pages += zbud_get_pool_size(cur->pool); >> } >> return pool_pages; >> }
Okay, I see your point. Yes, it's much better. Cai, Please make an new patch. Thanks, -Bob >> >> >>> >>> BTW: I'm still prefer to use dynamic pool size, instead of use >>> zswap_is_full(). AFAIR, Seth has a plan to replace the rbtree with radix >>> which will be more flexible to support this feature and page migration >>> as well. >>> >>>> Other question: >>>> Why do we need to update zswap_pool_pages too frequently? >>>> As I read the code, I think it's okay to update it only when user >>>> want to see it by debugfs and zswap_is_full is called. >>>> So could we optimize it out? >>>> >>>>> >>>>> Signed-off-by: Cai Liu <cai....@samsung.com> >>> >>> Reviewed-by: Bob Liu <bob....@oracle.com> >> >> Hmm, I really suprised you are okay in this code piece where we have >> unnecessary cost most of case(ie, most system has a swap device) in >> *mm* part. >> >> Anyway, I don't want to merge this patchset. >> If Andrew merge it and anybody doesn't do right work, I will send a patch. >> Cai, Could you redo a patch? >> I don't want to intercept your credit. >> >> Even, we could optimize to reduce the the number of call as I said in >> previous reply. > > You did it already. Please write it out in description. > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/