/*
* putback_zspage - add @zspage into right class's fullness list
- * @pool: target pool
* @class: destination class
* @zspage: target page
*
* Return @zspage's fullness_group
*/
-static enum fullness_group putback_zspage(struct zs_pool *pool,
- struct size_class *class,
+static enum fullness_group putback_zspage(struct size_class *class,
struct zspage *zspage)
{
enum fullness_group fullness;
insert_zspage(class, zspage, fullness);
set_zspage_mapping(zspage, class->index, fullness);
- if (fullness == ZS_EMPTY) {
- zs_stat_dec(class, OBJ_ALLOCATED, get_maxobj_per_zspage(
- class->size, class->pages_per_zspage));
- atomic_long_sub(class->pages_per_zspage,
- &pool->pages_allocated);
-
- free_zspage(pool, zspage);
- }
-
return fullness;
}
if (!migrate_zspage(pool, class, &cc))
break;
- putback_zspage(pool, class, dst_zspage);
+ putback_zspage(class, dst_zspage);
}
/* Stop if we couldn't find slot */
if (dst_zspage == NULL)
break;
- putback_zspage(pool, class, dst_zspage);
- if (putback_zspage(pool, class, src_zspage) == ZS_EMPTY)
+ putback_zspage(class, dst_zspage);
+ if (putback_zspage(class, src_zspage) == ZS_EMPTY) {
+ zs_stat_dec(class, OBJ_ALLOCATED, get_maxobj_per_zspage(
+ class->size, class->pages_per_zspage));
+ atomic_long_sub(class->pages_per_zspage,
+ &pool->pages_allocated);
+ free_zspage(pool, src_zspage);
pool->stats.pages_compacted += class->pages_per_zspage;
+ }
spin_unlock(&class->lock);
cond_resched();
spin_lock(&class->lock);
}
if (src_zspage)
- putback_zspage(pool, class, src_zspage);
+ putback_zspage(class, src_zspage);
spin_unlock(&class->lock);
}