return memcg ? memcg->kmemcg_id : -1;
}
-struct mem_cgroup *mem_cgroup_from_slab_obj(void *p);
+struct mem_cgroup *mem_cgroup_from_virt(void *p);
static inline void count_objcg_events(struct obj_cgroup *objcg,
enum vm_event_item idx,
return -1;
}
-static inline struct mem_cgroup *mem_cgroup_from_slab_obj(void *p)
+static inline struct mem_cgroup *mem_cgroup_from_virt(void *p)
{
return NULL;
}
if (list_lru_memcg_aware(lru)) {
rcu_read_lock();
- ret = list_lru_add(lru, item, nid, mem_cgroup_from_slab_obj(item));
+ ret = list_lru_add(lru, item, nid, mem_cgroup_from_virt(item));
rcu_read_unlock();
} else {
ret = list_lru_add(lru, item, nid, NULL);
if (list_lru_memcg_aware(lru)) {
rcu_read_lock();
- ret = list_lru_del(lru, item, nid, mem_cgroup_from_slab_obj(item));
+ ret = list_lru_del(lru, item, nid, mem_cgroup_from_virt(item));
rcu_read_unlock();
} else {
ret = list_lru_del(lru, item, nid, NULL);
struct lruvec *lruvec;
rcu_read_lock();
- memcg = mem_cgroup_from_slab_obj(p);
+ memcg = mem_cgroup_from_virt(p);
/*
* Untracked pages have no memcg, no lruvec. Update only the
* The caller must ensure the memcg lifetime, e.g. by taking rcu_read_lock(),
* cgroup_mutex, etc.
*/
-struct mem_cgroup *mem_cgroup_from_slab_obj(void *p)
+struct mem_cgroup *mem_cgroup_from_virt(void *p)
{
struct slab *slab;