[PATCH 03/16] zcache: Provide accessory functions for counter decrease.
This way we can have all wrapped with these functions and can disable/enable this with CONFIG_DEBUG_FS. [v2: Rebase on top of staging/zcache: Fix/improve zcache writeback code, tie to a config option] Signed-off-by: Konrad Rzeszutek Wilk --- drivers/staging/zcache/zcache-main.c | 96 +--- 1 file changed, 57 insertions(+), 39 deletions(-) diff --git a/drivers/staging/zcache/zcache-main.c b/drivers/staging/zcache/zcache-main.c index 5ad915a..50a408a 100644 --- a/drivers/staging/zcache/zcache-main.c +++ b/drivers/staging/zcache/zcache-main.c @@ -144,8 +144,12 @@ static inline void inc_zcache_obj_count(void) if (zcache_obj_count > zcache_obj_count_max) zcache_obj_count_max = zcache_obj_count; } - static long zcache_objnode_count; +static inline void dec_zcache_obj_count(void) +{ + zcache_obj_count = atomic_dec_return(_obj_atomic); + BUG_ON(zcache_obj_count < 0); +}; static atomic_t zcache_objnode_atomic = ATOMIC_INIT(0); static long zcache_objnode_count_max; static inline void inc_zcache_objnode_count(void) @@ -154,6 +158,11 @@ static inline void inc_zcache_objnode_count(void) if (zcache_objnode_count > zcache_objnode_count_max) zcache_objnode_count_max = zcache_objnode_count; }; +static inline void dec_zcache_objnode_count(void) +{ + zcache_objnode_count = atomic_dec_return(_objnode_atomic); + BUG_ON(zcache_objnode_count < 0); +}; static u64 zcache_eph_zbytes; static atomic_long_t zcache_eph_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_eph_zbytes_max; @@ -163,6 +172,10 @@ static inline void inc_zcache_eph_zbytes(unsigned clen) if (zcache_eph_zbytes > zcache_eph_zbytes_max) zcache_eph_zbytes_max = zcache_eph_zbytes; }; +static inline void dec_zcache_eph_zbytes(unsigned zsize) +{ + zcache_eph_zbytes = atomic_long_sub_return(zsize, _eph_zbytes_atomic); +}; static u64 zcache_pers_zbytes; static atomic_long_t zcache_pers_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_pers_zbytes_max; @@ -173,6 +186,10 @@ static inline void inc_zcache_pers_zbytes(unsigned clen) zcache_pers_zbytes_max = zcache_pers_zbytes; } static long zcache_eph_pageframes; +static inline void dec_zcache_pers_zbytes(unsigned zsize) +{ + zcache_pers_zbytes = atomic_long_sub_return(zsize, _pers_zbytes_atomic); +} static atomic_t zcache_eph_pageframes_atomic = ATOMIC_INIT(0); static long zcache_eph_pageframes_max; static inline void inc_zcache_eph_pageframes(void) @@ -182,6 +199,10 @@ static inline void inc_zcache_eph_pageframes(void) zcache_eph_pageframes_max = zcache_eph_pageframes; }; static long zcache_pers_pageframes; +static inline void dec_zcache_eph_pageframes(void) +{ + zcache_eph_pageframes = atomic_dec_return(_eph_pageframes_atomic); +}; static atomic_t zcache_pers_pageframes_atomic = ATOMIC_INIT(0); static long zcache_pers_pageframes_max; static inline void inc_zcache_pers_pageframes(void) @@ -191,6 +212,10 @@ static inline void inc_zcache_pers_pageframes(void) zcache_pers_pageframes_max = zcache_pers_pageframes; } static long zcache_pageframes_alloced; +static inline void dec_zcache_pers_pageframes(void) +{ + zcache_pers_pageframes = atomic_dec_return(_pers_pageframes_atomic); +} static atomic_t zcache_pageframes_alloced_atomic = ATOMIC_INIT(0); static inline void inc_zcache_pageframes_alloced(void) { @@ -212,6 +237,10 @@ static inline void inc_zcache_eph_zpages(void) zcache_eph_zpages_max = zcache_eph_zpages; } static long zcache_pers_zpages; +static inline void dec_zcache_eph_zpages(unsigned zpages) +{ + zcache_eph_zpages = atomic_sub_return(zpages, _eph_zpages_atomic); +} static atomic_t zcache_pers_zpages_atomic = ATOMIC_INIT(0); static long zcache_pers_zpages_max; static inline void inc_zcache_pers_zpages(void) @@ -220,6 +249,10 @@ static inline void inc_zcache_pers_zpages(void) if (zcache_pers_zpages > zcache_pers_zpages_max) zcache_pers_zpages_max = zcache_pers_zpages; } +static inline void dec_zcache_pers_zpages(unsigned zpages) +{ + zcache_pers_zpages = atomic_sub_return(zpages, _pers_zpages_atomic); +} /* but for the rest of these, counting races are ok */ static unsigned long zcache_flush_total; static unsigned long zcache_flush_found; @@ -484,9 +517,7 @@ static struct tmem_objnode *zcache_objnode_alloc(struct tmem_pool *pool) static void zcache_objnode_free(struct tmem_objnode *objnode, struct tmem_pool *pool) { - zcache_objnode_count = - atomic_dec_return(_objnode_atomic); - BUG_ON(zcache_objnode_count < 0); + dec_zcache_objnode_count(); kmem_cache_free(zcache_objnode_cache, objnode); } @@ -505,9 +536,7 @@ static struct tmem_obj *zcache_obj_alloc(struct tmem_pool *pool) static void zcache_obj_free(struct tmem_obj *obj, struct
[PATCH 03/16] zcache: Provide accessory functions for counter decrease.
This way we can have all wrapped with these functions and can disable/enable this with CONFIG_DEBUG_FS. [v2: Rebase on top of staging/zcache: Fix/improve zcache writeback code, tie to a config option] Signed-off-by: Konrad Rzeszutek Wilk konrad.w...@oracle.com --- drivers/staging/zcache/zcache-main.c | 96 +--- 1 file changed, 57 insertions(+), 39 deletions(-) diff --git a/drivers/staging/zcache/zcache-main.c b/drivers/staging/zcache/zcache-main.c index 5ad915a..50a408a 100644 --- a/drivers/staging/zcache/zcache-main.c +++ b/drivers/staging/zcache/zcache-main.c @@ -144,8 +144,12 @@ static inline void inc_zcache_obj_count(void) if (zcache_obj_count zcache_obj_count_max) zcache_obj_count_max = zcache_obj_count; } - static long zcache_objnode_count; +static inline void dec_zcache_obj_count(void) +{ + zcache_obj_count = atomic_dec_return(zcache_obj_atomic); + BUG_ON(zcache_obj_count 0); +}; static atomic_t zcache_objnode_atomic = ATOMIC_INIT(0); static long zcache_objnode_count_max; static inline void inc_zcache_objnode_count(void) @@ -154,6 +158,11 @@ static inline void inc_zcache_objnode_count(void) if (zcache_objnode_count zcache_objnode_count_max) zcache_objnode_count_max = zcache_objnode_count; }; +static inline void dec_zcache_objnode_count(void) +{ + zcache_objnode_count = atomic_dec_return(zcache_objnode_atomic); + BUG_ON(zcache_objnode_count 0); +}; static u64 zcache_eph_zbytes; static atomic_long_t zcache_eph_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_eph_zbytes_max; @@ -163,6 +172,10 @@ static inline void inc_zcache_eph_zbytes(unsigned clen) if (zcache_eph_zbytes zcache_eph_zbytes_max) zcache_eph_zbytes_max = zcache_eph_zbytes; }; +static inline void dec_zcache_eph_zbytes(unsigned zsize) +{ + zcache_eph_zbytes = atomic_long_sub_return(zsize, zcache_eph_zbytes_atomic); +}; static u64 zcache_pers_zbytes; static atomic_long_t zcache_pers_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_pers_zbytes_max; @@ -173,6 +186,10 @@ static inline void inc_zcache_pers_zbytes(unsigned clen) zcache_pers_zbytes_max = zcache_pers_zbytes; } static long zcache_eph_pageframes; +static inline void dec_zcache_pers_zbytes(unsigned zsize) +{ + zcache_pers_zbytes = atomic_long_sub_return(zsize, zcache_pers_zbytes_atomic); +} static atomic_t zcache_eph_pageframes_atomic = ATOMIC_INIT(0); static long zcache_eph_pageframes_max; static inline void inc_zcache_eph_pageframes(void) @@ -182,6 +199,10 @@ static inline void inc_zcache_eph_pageframes(void) zcache_eph_pageframes_max = zcache_eph_pageframes; }; static long zcache_pers_pageframes; +static inline void dec_zcache_eph_pageframes(void) +{ + zcache_eph_pageframes = atomic_dec_return(zcache_eph_pageframes_atomic); +}; static atomic_t zcache_pers_pageframes_atomic = ATOMIC_INIT(0); static long zcache_pers_pageframes_max; static inline void inc_zcache_pers_pageframes(void) @@ -191,6 +212,10 @@ static inline void inc_zcache_pers_pageframes(void) zcache_pers_pageframes_max = zcache_pers_pageframes; } static long zcache_pageframes_alloced; +static inline void dec_zcache_pers_pageframes(void) +{ + zcache_pers_pageframes = atomic_dec_return(zcache_pers_pageframes_atomic); +} static atomic_t zcache_pageframes_alloced_atomic = ATOMIC_INIT(0); static inline void inc_zcache_pageframes_alloced(void) { @@ -212,6 +237,10 @@ static inline void inc_zcache_eph_zpages(void) zcache_eph_zpages_max = zcache_eph_zpages; } static long zcache_pers_zpages; +static inline void dec_zcache_eph_zpages(unsigned zpages) +{ + zcache_eph_zpages = atomic_sub_return(zpages, zcache_eph_zpages_atomic); +} static atomic_t zcache_pers_zpages_atomic = ATOMIC_INIT(0); static long zcache_pers_zpages_max; static inline void inc_zcache_pers_zpages(void) @@ -220,6 +249,10 @@ static inline void inc_zcache_pers_zpages(void) if (zcache_pers_zpages zcache_pers_zpages_max) zcache_pers_zpages_max = zcache_pers_zpages; } +static inline void dec_zcache_pers_zpages(unsigned zpages) +{ + zcache_pers_zpages = atomic_sub_return(zpages, zcache_pers_zpages_atomic); +} /* but for the rest of these, counting races are ok */ static unsigned long zcache_flush_total; static unsigned long zcache_flush_found; @@ -484,9 +517,7 @@ static struct tmem_objnode *zcache_objnode_alloc(struct tmem_pool *pool) static void zcache_objnode_free(struct tmem_objnode *objnode, struct tmem_pool *pool) { - zcache_objnode_count = - atomic_dec_return(zcache_objnode_atomic); - BUG_ON(zcache_objnode_count 0); + dec_zcache_objnode_count(); kmem_cache_free(zcache_objnode_cache, objnode); } @@ -505,9 +536,7 @@ static struct tmem_obj *zcache_obj_alloc(struct tmem_pool
[PATCH 03/16] zcache: Provide accessory functions for counter decrease.
This way we can have all wrapped with these functions and can disable/enable this with CONFIG_DEBUG_FS. Signed-off-by: Konrad Rzeszutek Wilk --- drivers/staging/zcache/zcache-main.c | 88 +--- 1 file changed, 51 insertions(+), 37 deletions(-) diff --git a/drivers/staging/zcache/zcache-main.c b/drivers/staging/zcache/zcache-main.c index abc9722..a55a95b 100644 --- a/drivers/staging/zcache/zcache-main.c +++ b/drivers/staging/zcache/zcache-main.c @@ -143,8 +143,12 @@ static inline void inc_zcache_obj_count(void) if (zcache_obj_count > zcache_obj_count_max) zcache_obj_count_max = zcache_obj_count; } - static long zcache_objnode_count; +static inline void dec_zcache_obj_count(void) +{ + zcache_obj_count = atomic_dec_return(_obj_atomic); + BUG_ON(zcache_obj_count < 0); +}; static atomic_t zcache_objnode_atomic = ATOMIC_INIT(0); static long zcache_objnode_count_max; static inline void inc_zcache_objnode_count(void) @@ -153,6 +157,11 @@ static inline void inc_zcache_objnode_count(void) if (zcache_objnode_count > zcache_objnode_count_max) zcache_objnode_count_max = zcache_objnode_count; }; +static inline void dec_zcache_objnode_count(void) +{ + zcache_objnode_count = atomic_dec_return(_objnode_atomic); + BUG_ON(zcache_objnode_count < 0); +}; static u64 zcache_eph_zbytes; static atomic_long_t zcache_eph_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_eph_zbytes_max; @@ -162,6 +171,10 @@ static inline void inc_zcache_eph_zbytes(unsigned clen) if (zcache_eph_zbytes > zcache_eph_zbytes_max) zcache_eph_zbytes_max = zcache_eph_zbytes; }; +static inline void dec_zcache_eph_zbytes(unsigned zsize) +{ + zcache_eph_zbytes = atomic_long_sub_return(zsize, _eph_zbytes_atomic); +}; static u64 zcache_pers_zbytes; static atomic_long_t zcache_pers_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_pers_zbytes_max; @@ -172,6 +185,10 @@ static inline void inc_zcache_pers_zbytes(unsigned clen) zcache_pers_zbytes_max = zcache_pers_zbytes; } static long zcache_eph_pageframes; +static inline void dec_zcache_pers_zbytes(unsigned zsize) +{ + zcache_pers_zbytes = atomic_long_sub_return(zsize, _pers_zbytes_atomic); +} static atomic_t zcache_eph_pageframes_atomic = ATOMIC_INIT(0); static long zcache_eph_pageframes_max; static inline void inc_zcache_eph_pageframes(void) @@ -181,6 +198,10 @@ static inline void inc_zcache_eph_pageframes(void) zcache_eph_pageframes_max = zcache_eph_pageframes; }; static long zcache_pers_pageframes; +static inline void dec_zcache_eph_pageframes(void) +{ + zcache_eph_pageframes = atomic_dec_return(_eph_pageframes_atomic); +}; static atomic_t zcache_pers_pageframes_atomic = ATOMIC_INIT(0); static long zcache_pers_pageframes_max; static inline void inc_zcache_pers_pageframes(void) @@ -190,6 +211,10 @@ static inline void inc_zcache_pers_pageframes(void) zcache_pers_pageframes_max = zcache_pers_pageframes; } static long zcache_pageframes_alloced; +static inline void dec_zcache_pers_pageframes(void) +{ + zcache_pers_pageframes = atomic_dec_return(_pers_pageframes_atomic); +} static atomic_t zcache_pageframes_alloced_atomic = ATOMIC_INIT(0); static inline void inc_zcache_pageframes_alloced(void) { @@ -211,6 +236,10 @@ static inline void inc_zcache_eph_zpages(void) zcache_eph_zpages_max = zcache_eph_zpages; } static long zcache_pers_zpages; +static inline void dec_zcache_eph_zpages(unsigned zpages) +{ + zcache_eph_zpages = atomic_sub_return(zpages, _eph_zpages_atomic); +} static atomic_t zcache_pers_zpages_atomic = ATOMIC_INIT(0); static long zcache_pers_zpages_max; static inline void inc_zcache_pers_zpages(void) @@ -219,6 +248,10 @@ static inline void inc_zcache_pers_zpages(void) if (zcache_pers_zpages > zcache_pers_zpages_max) zcache_pers_zpages_max = zcache_pers_zpages; } +static inline void dec_zcache_pers_zpages(unsigned zpages) +{ + zcache_pers_zpages = atomic_sub_return(zpages, _pers_zpages_atomic); +} /* but for the rest of these, counting races are ok */ static unsigned long zcache_flush_total; static unsigned long zcache_flush_found; @@ -463,9 +496,7 @@ static struct tmem_objnode *zcache_objnode_alloc(struct tmem_pool *pool) static void zcache_objnode_free(struct tmem_objnode *objnode, struct tmem_pool *pool) { - zcache_objnode_count = - atomic_dec_return(_objnode_atomic); - BUG_ON(zcache_objnode_count < 0); + dec_zcache_objnode_count(); kmem_cache_free(zcache_objnode_cache, objnode); } @@ -484,9 +515,7 @@ static struct tmem_obj *zcache_obj_alloc(struct tmem_pool *pool) static void zcache_obj_free(struct tmem_obj *obj, struct tmem_pool *pool) { - zcache_obj_count = - atomic_dec_return(_obj_atomic); -
[PATCH 03/16] zcache: Provide accessory functions for counter decrease.
This way we can have all wrapped with these functions and can disable/enable this with CONFIG_DEBUG_FS. Signed-off-by: Konrad Rzeszutek Wilk konrad.w...@oracle.com --- drivers/staging/zcache/zcache-main.c | 88 +--- 1 file changed, 51 insertions(+), 37 deletions(-) diff --git a/drivers/staging/zcache/zcache-main.c b/drivers/staging/zcache/zcache-main.c index abc9722..a55a95b 100644 --- a/drivers/staging/zcache/zcache-main.c +++ b/drivers/staging/zcache/zcache-main.c @@ -143,8 +143,12 @@ static inline void inc_zcache_obj_count(void) if (zcache_obj_count zcache_obj_count_max) zcache_obj_count_max = zcache_obj_count; } - static long zcache_objnode_count; +static inline void dec_zcache_obj_count(void) +{ + zcache_obj_count = atomic_dec_return(zcache_obj_atomic); + BUG_ON(zcache_obj_count 0); +}; static atomic_t zcache_objnode_atomic = ATOMIC_INIT(0); static long zcache_objnode_count_max; static inline void inc_zcache_objnode_count(void) @@ -153,6 +157,11 @@ static inline void inc_zcache_objnode_count(void) if (zcache_objnode_count zcache_objnode_count_max) zcache_objnode_count_max = zcache_objnode_count; }; +static inline void dec_zcache_objnode_count(void) +{ + zcache_objnode_count = atomic_dec_return(zcache_objnode_atomic); + BUG_ON(zcache_objnode_count 0); +}; static u64 zcache_eph_zbytes; static atomic_long_t zcache_eph_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_eph_zbytes_max; @@ -162,6 +171,10 @@ static inline void inc_zcache_eph_zbytes(unsigned clen) if (zcache_eph_zbytes zcache_eph_zbytes_max) zcache_eph_zbytes_max = zcache_eph_zbytes; }; +static inline void dec_zcache_eph_zbytes(unsigned zsize) +{ + zcache_eph_zbytes = atomic_long_sub_return(zsize, zcache_eph_zbytes_atomic); +}; static u64 zcache_pers_zbytes; static atomic_long_t zcache_pers_zbytes_atomic = ATOMIC_INIT(0); static u64 zcache_pers_zbytes_max; @@ -172,6 +185,10 @@ static inline void inc_zcache_pers_zbytes(unsigned clen) zcache_pers_zbytes_max = zcache_pers_zbytes; } static long zcache_eph_pageframes; +static inline void dec_zcache_pers_zbytes(unsigned zsize) +{ + zcache_pers_zbytes = atomic_long_sub_return(zsize, zcache_pers_zbytes_atomic); +} static atomic_t zcache_eph_pageframes_atomic = ATOMIC_INIT(0); static long zcache_eph_pageframes_max; static inline void inc_zcache_eph_pageframes(void) @@ -181,6 +198,10 @@ static inline void inc_zcache_eph_pageframes(void) zcache_eph_pageframes_max = zcache_eph_pageframes; }; static long zcache_pers_pageframes; +static inline void dec_zcache_eph_pageframes(void) +{ + zcache_eph_pageframes = atomic_dec_return(zcache_eph_pageframes_atomic); +}; static atomic_t zcache_pers_pageframes_atomic = ATOMIC_INIT(0); static long zcache_pers_pageframes_max; static inline void inc_zcache_pers_pageframes(void) @@ -190,6 +211,10 @@ static inline void inc_zcache_pers_pageframes(void) zcache_pers_pageframes_max = zcache_pers_pageframes; } static long zcache_pageframes_alloced; +static inline void dec_zcache_pers_pageframes(void) +{ + zcache_pers_pageframes = atomic_dec_return(zcache_pers_pageframes_atomic); +} static atomic_t zcache_pageframes_alloced_atomic = ATOMIC_INIT(0); static inline void inc_zcache_pageframes_alloced(void) { @@ -211,6 +236,10 @@ static inline void inc_zcache_eph_zpages(void) zcache_eph_zpages_max = zcache_eph_zpages; } static long zcache_pers_zpages; +static inline void dec_zcache_eph_zpages(unsigned zpages) +{ + zcache_eph_zpages = atomic_sub_return(zpages, zcache_eph_zpages_atomic); +} static atomic_t zcache_pers_zpages_atomic = ATOMIC_INIT(0); static long zcache_pers_zpages_max; static inline void inc_zcache_pers_zpages(void) @@ -219,6 +248,10 @@ static inline void inc_zcache_pers_zpages(void) if (zcache_pers_zpages zcache_pers_zpages_max) zcache_pers_zpages_max = zcache_pers_zpages; } +static inline void dec_zcache_pers_zpages(unsigned zpages) +{ + zcache_pers_zpages = atomic_sub_return(zpages, zcache_pers_zpages_atomic); +} /* but for the rest of these, counting races are ok */ static unsigned long zcache_flush_total; static unsigned long zcache_flush_found; @@ -463,9 +496,7 @@ static struct tmem_objnode *zcache_objnode_alloc(struct tmem_pool *pool) static void zcache_objnode_free(struct tmem_objnode *objnode, struct tmem_pool *pool) { - zcache_objnode_count = - atomic_dec_return(zcache_objnode_atomic); - BUG_ON(zcache_objnode_count 0); + dec_zcache_objnode_count(); kmem_cache_free(zcache_objnode_cache, objnode); } @@ -484,9 +515,7 @@ static struct tmem_obj *zcache_obj_alloc(struct tmem_pool *pool) static void zcache_obj_free(struct tmem_obj *obj, struct tmem_pool *pool) { -