Searched refs:memcg (Results 1 - 25 of 48) sorted by relevance

12

/linux-master/mm/
H A Dmemcontrol.c24 * Per memcg lru locking
138 * memcg which the event belongs to.
140 struct mem_cgroup *memcg; member in struct:mem_cgroup_event
154 int (*register_event)(struct mem_cgroup *memcg,
161 void (*unregister_event)(struct mem_cgroup *memcg,
173 static void mem_cgroup_threshold(struct mem_cgroup *memcg);
174 static void mem_cgroup_oom_notify(struct mem_cgroup *memcg);
242 struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg) argument
244 if (!memcg)
245 memcg
329 memcg_reparent_objcgs(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument
377 struct mem_cgroup *memcg = folio_memcg(folio); local
400 struct mem_cgroup *memcg; local
473 soft_limit_excess(struct mem_cgroup *memcg) argument
485 mem_cgroup_update_tree(struct mem_cgroup *memcg, int nid) argument
528 mem_cgroup_remove_from_trees(struct mem_cgroup *memcg) argument
846 memcg_rstat_updated(struct mem_cgroup *memcg, int val) argument
874 do_flush_stats(struct mem_cgroup *memcg) argument
891 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument
903 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument
920 memcg_page_state(struct mem_cgroup *memcg, int idx) argument
958 __mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, int val) argument
974 memcg_page_state_local(struct mem_cgroup *memcg, int idx) argument
995 struct mem_cgroup *memcg; local
1057 struct mem_cgroup *memcg; local
1079 struct mem_cgroup *memcg; local
1106 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1123 memcg_events(struct mem_cgroup *memcg, int event) argument
1133 memcg_events_local(struct mem_cgroup *memcg, int event) argument
1143 mem_cgroup_charge_statistics(struct mem_cgroup *memcg, int nr_pages) argument
1157 mem_cgroup_event_ratelimit(struct mem_cgroup *memcg, enum mem_cgroup_events_target target) argument
1186 memcg_check_events(struct mem_cgroup *memcg, int nid) argument
1239 struct mem_cgroup *memcg; local
1281 struct mem_cgroup *memcg; local
1320 struct mem_cgroup *memcg = NULL; local
1445 struct mem_cgroup *memcg = dead_memcg; local
1477 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument
1503 struct mem_cgroup *memcg; local
1632 mem_cgroup_margin(struct mem_cgroup *memcg) argument
1662 mem_cgroup_under_move(struct mem_cgroup *memcg) argument
1684 mem_cgroup_wait_acct_move(struct mem_cgroup *memcg) argument
1786 memcg_page_state_output(struct mem_cgroup *memcg, int item) argument
1793 memcg_page_state_local_output( struct mem_cgroup *memcg, int item) argument
1800 memcg_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
1855 memory_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
1873 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument
1894 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument
1929 mem_cgroup_get_max(struct mem_cgroup *memcg) argument
1948 mem_cgroup_size(struct mem_cgroup *memcg) argument
1953 mem_cgroup_out_of_memory(struct mem_cgroup *memcg, gfp_t gfp_mask, int order) argument
2044 mem_cgroup_oom_trylock(struct mem_cgroup *memcg) argument
2083 mem_cgroup_oom_unlock(struct mem_cgroup *memcg) argument
2094 mem_cgroup_mark_under_oom(struct mem_cgroup *memcg) argument
2104 mem_cgroup_unmark_under_oom(struct mem_cgroup *memcg) argument
2122 struct mem_cgroup *memcg; member in struct:oom_wait_info
2142 memcg_oom_recover(struct mem_cgroup *memcg) argument
2160 mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) argument
2230 struct mem_cgroup *memcg = current->memcg_in_oom; local
2281 struct mem_cgroup *memcg; local
2324 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument
2343 struct mem_cgroup *memcg; local
2385 __folio_memcg_unlock(struct mem_cgroup *memcg) argument
2450 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument
2466 consume_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2539 __refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2557 refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2587 struct mem_cgroup *memcg; local
2621 reclaim_high(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument
2649 struct mem_cgroup *memcg; local
2726 mem_find_max_overage(struct mem_cgroup *memcg) argument
2740 swap_find_max_overage(struct mem_cgroup *memcg) argument
2760 calculate_high_delay(struct mem_cgroup *memcg, unsigned int nr_pages, u64 max_overage) argument
2804 struct mem_cgroup *memcg; local
2890 try_charge_memcg(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument
3086 try_charge(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument
3100 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument
3110 commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
3130 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
3147 struct mem_cgroup *memcg; local
3239 __get_obj_cgroup_from_memcg(struct mem_cgroup *memcg) argument
3254 struct mem_cgroup *memcg; local
3305 struct mem_cgroup *memcg; local
3357 struct mem_cgroup *memcg; local
3370 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument
3390 struct mem_cgroup *memcg; local
3411 struct mem_cgroup *memcg; local
3569 struct mem_cgroup *memcg; local
3624 struct mem_cgroup *memcg; local
3756 struct mem_cgroup *memcg; local
3815 struct mem_cgroup *memcg = folio_memcg(folio); local
3872 mem_cgroup_resize_max(struct mem_cgroup *memcg, unsigned long max, bool memsw) argument
4013 mem_cgroup_force_empty(struct mem_cgroup *memcg) argument
4039 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
4065 mem_cgroup_usage(struct mem_cgroup *memcg, bool swap) argument
4098 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4149 memcg_online_kmem(struct mem_cgroup *memcg) argument
4175 memcg_offline_kmem(struct mem_cgroup *memcg) argument
4200 memcg_online_kmem(struct mem_cgroup *memcg) argument
4204 memcg_offline_kmem(struct mem_cgroup *memcg) argument
4209 memcg_update_tcp_max(struct mem_cgroup *memcg, unsigned long max) argument
4251 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
4300 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
4344 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4376 mem_cgroup_node_nr_lru_pages(struct mem_cgroup *memcg, int nid, unsigned int lru_mask, bool tree) argument
4396 mem_cgroup_nr_lru_pages(struct mem_cgroup *memcg, unsigned int lru_mask, bool tree) argument
4429 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
4504 memcg1_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
4581 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4589 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4602 __mem_cgroup_threshold(struct mem_cgroup *memcg, bool swap) argument
4653 mem_cgroup_threshold(struct mem_cgroup *memcg) argument
4678 mem_cgroup_oom_notify_cb(struct mem_cgroup *memcg) argument
4691 mem_cgroup_oom_notify(struct mem_cgroup *memcg) argument
4699 __mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args, enum res_type type) argument
4779 mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4785 memsw_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4791 __mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, enum res_type type) argument
4876 mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4882 memsw_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4888 mem_cgroup_oom_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4910 mem_cgroup_oom_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4929 struct mem_cgroup *memcg = mem_cgroup_from_seq(sf); local
4941 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4958 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument
4963 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument
4968 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument
4975 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
5005 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
5073 struct mem_cgroup *memcg = folio_memcg(folio); local
5125 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
5152 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument
5157 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument
5161 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument
5189 struct mem_cgroup *memcg = event->memcg; local
5213 struct mem_cgroup *memcg = event->memcg; local
5263 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5572 mem_cgroup_id_remove(struct mem_cgroup *memcg) argument
5580 mem_cgroup_id_get_many(struct mem_cgroup *memcg, unsigned int n) argument
5586 mem_cgroup_id_put_many(struct mem_cgroup *memcg, unsigned int n) argument
5596 mem_cgroup_id_put(struct mem_cgroup *memcg) argument
5618 struct mem_cgroup *memcg; local
5636 alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument
5665 free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument
5677 __mem_cgroup_free(struct mem_cgroup *memcg) argument
5690 mem_cgroup_free(struct mem_cgroup *memcg) argument
5700 struct mem_cgroup *memcg; local
5776 struct mem_cgroup *memcg, *old_memcg; local
5825 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5869 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5901 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5909 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5949 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5965 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6570 struct mem_cgroup *memcg = NULL; /* unneeded init to make gcc happy */ local
6905 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6913 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6927 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6950 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6973 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
7022 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
7084 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
7092 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
7100 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
7124 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
7153 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
7163 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
7185 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
7443 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument
7490 charge_memcg(struct folio *folio, struct mem_cgroup *memcg, gfp_t gfp) argument
7506 struct mem_cgroup *memcg; local
7533 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument
7566 struct mem_cgroup *memcg; local
7620 struct mem_cgroup *memcg; member in struct:uncharge_gather
7658 struct mem_cgroup *memcg; local
7755 struct mem_cgroup *memcg; local
7805 struct mem_cgroup *memcg; local
7846 struct mem_cgroup *memcg; local
7882 mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument
7913 mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) argument
7986 mem_cgroup_id_get_online(struct mem_cgroup *memcg) argument
8013 struct mem_cgroup *memcg, *swap_memcg; local
8085 struct mem_cgroup *memcg; local
8129 struct mem_cgroup *memcg; local
8148 mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument
8163 struct mem_cgroup *memcg; local
8203 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
8211 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
8225 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8248 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8264 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
8350 struct mem_cgroup *memcg, *original_memcg; local
8394 struct mem_cgroup *memcg; local
8421 struct mem_cgroup *memcg; local
8435 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument
8444 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
8459 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8475 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
8484 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
[all...]
H A Dshrinker_debug.c19 struct mem_cgroup *memcg,
30 .memcg = memcg,
51 struct mem_cgroup *memcg; local
64 memcg = mem_cgroup_iter(NULL, NULL, NULL);
66 if (memcg && !mem_cgroup_online(memcg))
70 memcg_aware ? memcg : NULL,
73 seq_printf(m, "%lu", mem_cgroup_ino(memcg));
80 mem_cgroup_iter_break(NULL, memcg);
18 shrinker_count_objects(struct shrinker *shrinker, struct mem_cgroup *memcg, unsigned long *count_per_node) argument
113 struct mem_cgroup *memcg = NULL; local
[all...]
H A Dshrinker.c62 void free_shrinker_info(struct mem_cgroup *memcg) argument
69 pn = memcg->nodeinfo[nid];
77 int alloc_shrinker_info(struct mem_cgroup *memcg) argument
92 rcu_assign_pointer(memcg->nodeinfo[nid]->shrinker_info, info);
100 free_shrinker_info(memcg);
104 static struct shrinker_info *shrinker_info_protected(struct mem_cgroup *memcg, argument
107 return rcu_dereference_protected(memcg->nodeinfo[nid]->shrinker_info,
111 static int expand_one_shrinker_info(struct mem_cgroup *memcg, int new_size, argument
119 pn = memcg->nodeinfo[nid];
120 old = shrinker_info_protected(memcg, ni
153 struct mem_cgroup *memcg; local
194 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument
250 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
266 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
283 reparent_shrinker_deferred(struct mem_cgroup *memcg) argument
321 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
327 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
467 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
585 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
612 shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
[all...]
H A Dvmpressure.c77 struct mem_cgroup *memcg = vmpressure_to_memcg(vmpr); local
79 memcg = parent_mem_cgroup(memcg);
80 if (!memcg)
82 return memcg_to_vmpressure(memcg);
221 * @memcg: cgroup memory controller handle
231 * mode: @memcg is considered the pressure root and userspace is
234 * If @tree is not set, reclaim efficiency is recorded for @memcg, and
239 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument
249 * for this @memcg rathe
335 vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) argument
374 vmpressure_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
435 vmpressure_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
[all...]
H A Dlist_lru.c89 struct mem_cgroup *memcg)
96 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
100 set_shrinker_bit(memcg, nid, lru_shrinker_id(lru));
113 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local
116 return list_lru_add(lru, item, nid, memcg);
121 struct mem_cgroup *memcg)
128 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
143 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local
146 return list_lru_del(lru, item, nid, memcg);
166 int nid, struct mem_cgroup *memcg)
88 list_lru_add(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument
120 list_lru_del(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument
165 list_lru_count_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg) argument
258 list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument
274 list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument
420 memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument
454 memcg_list_lru_allocated(struct mem_cgroup *memcg, struct list_lru *lru) argument
462 memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, gfp_t gfp) argument
469 struct mem_cgroup *memcg; member in struct:list_lru_memcg_table
[all...]
H A Dworkingset.c243 struct mem_cgroup *memcg = folio_memcg(folio); local
248 lruvec = mem_cgroup_lruvec(memcg, pgdat);
256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs);
268 struct mem_cgroup *memcg; local
273 memcg = mem_cgroup_from_id(memcg_id);
274 *lruvec = mem_cgroup_lruvec(memcg, pgdat);
444 * Look up the memcg associated with the stored ID. It might
470 * XXX: With per-memcg flushing and thresholding, is ratelimiting
527 * evicted folio in the context of the node and the memcg whose memory
534 struct mem_cgroup *memcg; local
586 struct mem_cgroup *memcg; local
[all...]
H A Dmmap_lock.c32 * Size of the buffer for memcg path names. Ignoring stack trace support,
188 * Write the given mm_struct's memcg path to a percpu buffer, and return a
202 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local
204 if (memcg == NULL)
206 if (unlikely(memcg->css.cgroup == NULL))
213 cgroup_path(memcg->css.cgroup, buf, MEMCG_PATH_BUF_SIZE);
216 css_put(&memcg->css);
H A Dzswap.c716 struct mem_cgroup *memcg; local
721 * concurrent memcg offlining. Thanks to the memcg->kmemcg_id indirection
724 * 1. list_lru_add() is called before memcg->kmemcg_id is updated. The
725 * new entry will be reparented to memcg's parent's list_lru.
726 * 2. list_lru_add() is called after memcg->kmemcg_id is updated. The
727 * new entry will be added directly to memcg's parent's list_lru.
732 memcg = mem_cgroup_from_entry(entry);
734 list_lru_add(list_lru, &entry->lru, nid, memcg);
737 lru_size = list_lru_count_one(list_lru, nid, memcg);
754 struct mem_cgroup *memcg; local
778 zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) argument
1216 struct mem_cgroup *memcg = sc->memcg; local
1290 shrink_memcg(struct mem_cgroup *memcg) argument
1315 struct mem_cgroup *memcg; local
1419 struct mem_cgroup *memcg = NULL; local
[all...]
H A Dvmscan.c218 * completely broken with the legacy memcg and direct stalling in
279 * single memcg. For example, a memcg-aware shrinker can free one object
280 * charged to the target memcg, causing an entire page to be freed.
281 * If we count the entire page as reclaimed from the memcg, we end up
285 * from the target memcg; preventing unnecessary retries during memcg
289 * charged to the target memcg, we end up underestimating the reclaimed
296 * memcg reclaim, to make reporting more accurate and reduce
317 static inline bool can_reclaim_anon_pages(struct mem_cgroup *memcg, argument
389 struct mem_cgroup *memcg = NULL; local
2354 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2605 get_lruvec(struct mem_cgroup *memcg, int nid) argument
2627 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2753 get_mm_list(struct mem_cgroup *memcg) argument
2795 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local
2823 struct mem_cgroup *memcg = NULL; local
2861 struct mem_cgroup *memcg; local
2890 get_mm_list(struct mem_cgroup *memcg) argument
2938 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2999 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3315 get_pfn_folio(unsigned long pfn, struct mem_cgroup *memcg, struct pglist_data *pgdat, bool can_swap) argument
3356 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local
3424 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local
3630 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3909 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3933 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3956 struct mem_cgroup *memcg; local
4016 struct mem_cgroup *memcg = folio_memcg(folio); local
4167 lru_gen_online_memcg(struct mem_cgroup *memcg) argument
4192 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument
4203 lru_gen_release_memcg(struct mem_cgroup *memcg) argument
4229 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument
4359 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4536 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4685 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4772 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4817 struct mem_cgroup *memcg; local
5055 struct mem_cgroup *memcg; local
5190 struct mem_cgroup *memcg; local
5222 struct mem_cgroup *memcg = lruvec_memcg(v); local
5301 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
5410 struct mem_cgroup *memcg = NULL; local
5583 lru_gen_init_memcg(struct mem_cgroup *memcg) argument
5594 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument
5848 struct mem_cgroup *memcg; local
6476 mem_cgroup_shrink_node(struct mem_cgroup *memcg, gfp_t gfp_mask, bool noswap, pg_data_t *pgdat, unsigned long *nr_scanned) argument
6515 try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, unsigned long nr_pages, gfp_t gfp_mask, unsigned int reclaim_options) argument
6557 struct mem_cgroup *memcg; local
[all...]
H A Doom_kill.c74 return oc->memcg != NULL;
261 oc->totalpages = mem_cgroup_get_max(oc->memcg) ?: 1;
370 mem_cgroup_scan_tasks(oc->memcg, oom_evaluate_task, oc);
420 * memcg, not in the same cpuset, or bound to a disjoint set of mempolicy nodes
431 mem_cgroup_scan_tasks(oc->memcg, dump_task, oc);
449 mem_cgroup_print_oom_context(oc->memcg, victim);
464 mem_cgroup_print_oom_meminfo(oc->memcg);
761 /* OOM killer might race with memcg OOM */
1044 oom_group = mem_cgroup_get_oom_group(victim, oc->memcg);
1070 * does not panic for cpuset, mempolicy, or memcg allocatio
[all...]
/linux-master/include/linux/
H A Dmemcontrol.h73 * Per memcg event counter is incremented at every pagein/pageout. With THP,
76 * than using jiffies etc. to handle periodic memcg event.
114 struct mem_cgroup *memcg; /* Back pointer, we cannot */ member in struct:mem_cgroup_per_node
156 int memcg_id; /* memcg->css.id of foreign inode */
169 struct mem_cgroup *memcg; member in struct:obj_cgroup
186 /* Private memcg ID. Used to ID objects that outlive the cgroup */
208 * Prevent pages from this memcg from being written back from zswap to
283 * memcg->objcg is wiped out as a part of the objcg repaprenting
284 * process. memcg->orig_objcg preserves a pointer (and a reference)
285 * to the original objcg until the end of live of memcg
531 struct mem_cgroup *memcg; local
573 mem_cgroup_is_root(struct mem_cgroup *memcg) argument
583 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument
636 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
648 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
658 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
737 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument
772 struct mem_cgroup *memcg = folio_memcg(folio); local
825 mem_cgroup_tryget(struct mem_cgroup *memcg) argument
830 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument
835 mem_cgroup_put(struct mem_cgroup *memcg) argument
851 mem_cgroup_id(struct mem_cgroup *memcg) argument
861 mem_cgroup_ino(struct mem_cgroup *memcg) argument
891 parent_mem_cgroup(struct mem_cgroup *memcg) argument
896 mem_cgroup_is_descendant(struct mem_cgroup *memcg, struct mem_cgroup *root) argument
904 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument
921 mem_cgroup_online(struct mem_cgroup *memcg) argument
981 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument
998 mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, int val) argument
1011 struct mem_cgroup *memcg; local
1046 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1060 struct mem_cgroup *memcg = folio_memcg(folio); local
1069 struct mem_cgroup *memcg; local
1081 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument
1109 struct mem_cgroup *memcg; local
1172 mem_cgroup_is_root(struct mem_cgroup *memcg) argument
1182 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument
1192 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument
1200 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument
1205 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1210 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1216 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1222 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
1233 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument
1257 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument
1271 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument
1288 parent_mem_cgroup(struct mem_cgroup *memcg) argument
1293 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument
1319 mem_cgroup_tryget(struct mem_cgroup *memcg) argument
1324 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument
1329 mem_cgroup_put(struct mem_cgroup *memcg) argument
1371 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument
1376 mem_cgroup_id(struct mem_cgroup *memcg) argument
1389 mem_cgroup_ino(struct mem_cgroup *memcg) argument
1410 mem_cgroup_online(struct mem_cgroup *memcg) argument
1422 mem_cgroup_get_max(struct mem_cgroup *memcg) argument
1427 mem_cgroup_size(struct mem_cgroup *memcg) argument
1433 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument
1438 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument
1450 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument
1490 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument
1494 __mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, int nr) argument
1500 mod_memcg_state(struct mem_cgroup *memcg, enum memcg_stat_item idx, int nr) argument
1511 memcg_page_state(struct mem_cgroup *memcg, int idx) argument
1528 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument
1532 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument
1552 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1558 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1612 struct mem_cgroup *memcg; local
1688 struct mem_cgroup *memcg; local
1735 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument
1754 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument
1759 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument
1825 memcg_kmem_id(struct mem_cgroup *memcg) argument
1836 struct mem_cgroup *memcg; local
1888 memcg_kmem_id(struct mem_cgroup *memcg) argument
1928 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument
[all...]
H A Dvmpressure.h33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
39 extern struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg);
41 extern int vmpressure_register_event(struct mem_cgroup *memcg,
44 extern void vmpressure_unregister_event(struct mem_cgroup *memcg,
47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument
49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, argument
H A Dlist_lru.h33 /* may become negative during memcg reparenting */
70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru,
72 void memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent);
79 * @memcg: the cgroup of the sublist to add the item to.
93 struct mem_cgroup *memcg);
101 * memcg of the sublist is determined by @item list_head. This assumption is
113 * @memcg: the cgroup of the sublist to delete the item from.
122 struct mem_cgroup *memcg);
130 * memcg of the sublist is determined by @item list_head. This assumption is
141 * @memcg
[all...]
H A Dzswap.h35 void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg);
59 static inline void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) {} argument
H A Doom.h37 struct mem_cgroup *memcg; member in struct:oom_control
H A Dshrinker.h13 * Bitmap and deferred work of shrinker::id corresponding to memcg-aware
14 * shrinkers, which have elements charged to the memcg.
54 /* current memcg being shrunk (for memcg aware shrinkers) */
55 struct mem_cgroup *memcg; member in struct:shrink_control
H A Dswap.h407 extern unsigned long try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg,
607 static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) argument
614 if (mem_cgroup_disabled() || mem_cgroup_is_root(memcg))
617 return READ_ONCE(memcg->swappiness);
659 extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg);
677 static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument
H A Dmmzone.h255 * Per-memcg and per-lruvec counters track memory, consumed
451 /* the memcg generation this lru_gen_folio belongs to */
508 * An onlining memcg is added to the tail of a random bin in the old generation.
510 * per-node memcg generation counter, whose reminder (mod MEMCG_NR_GENS) indexes
514 * 1. MEMCG_LRU_HEAD, which moves a memcg to the head of a random bin in its
516 * 2. MEMCG_LRU_TAIL, which moves a memcg to the tail of a random bin in its
518 * 3. MEMCG_LRU_OLD, which moves a memcg to the head of a random bin in the old
520 * 4. MEMCG_LRU_YOUNG, which moves a memcg to the tail of a random bin in the
526 * 2. The first attempt to reclaim a memcg below low, which triggers
528 * 3. The first attempt to reclaim a memcg offline
583 lru_gen_init_memcg(struct mem_cgroup *memcg) argument
587 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument
591 lru_gen_online_memcg(struct mem_cgroup *memcg) argument
595 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument
599 lru_gen_release_memcg(struct mem_cgroup *memcg) argument
603 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument
[all...]
H A Dbacking-dev.h158 void wb_memcg_offline(struct mem_cgroup *memcg);
165 * Cgroup writeback requires support from the filesystem. Also, both memcg and
170 * depending on how memcg and iocg are configured.
186 * Find the wb of @bdi which matches both the memcg and blkcg of %current.
202 * %current's blkcg equals the effective blkcg of its memcg. No
215 * Equivalent to wb_get_create() on %current's memcg. This function is
362 static inline void wb_memcg_offline(struct mem_cgroup *memcg) argument
/linux-master/tools/testing/selftests/cgroup/
H A Dtest_memcontrol.c170 char *memcg; local
172 memcg = cg_name(root, "memcg_test");
173 if (!memcg)
176 if (cg_create(memcg))
179 current = cg_read_long(memcg, "memory.current");
183 if (cg_run(memcg, alloc_anon_50M_check, NULL))
186 if (cg_run(memcg, alloc_pagecache_50M_check, NULL))
192 cg_destroy(memcg);
193 free(memcg);
251 static bool reclaim_until(const char *memcg, lon
503 char *memcg; local
566 char *memcg; local
628 char *memcg; local
686 reclaim_until(const char *memcg, long goal) argument
722 char *memcg; local
825 char *memcg; local
889 char *memcg; local
1050 char *memcg; local
1248 char *memcg; local
[all...]
/linux-master/tools/cgroup/
H A Dmemcg_slabinfo.py42 memcg = container_of(css, 'struct mem_cgroup', 'css')
43 MEMCGS[css.cgroup.kn.id.value_()] = memcg
170 memcg = MEMCGS[cgroup_id]
185 # find memcg pointers belonging to the specified cgroup
186 obj_cgroups.add(memcg.objcg.value_())
188 memcg.objcg_list.address_of_(),
220 memcg.kmem_caches.address_of_(),
/linux-master/include/linux/sched/
H A Dmm.h490 * set_active_memcg - Starts the remote memcg charging scope.
491 * @memcg: memcg to charge.
493 * This function marks the beginning of the remote memcg charging scope. All the
495 * given memcg.
497 * Please, make sure that caller has a reference to the passed memcg structure,
505 set_active_memcg(struct mem_cgroup *memcg) argument
511 this_cpu_write(int_active_memcg, memcg);
514 current->active_memcg = memcg;
521 set_active_memcg(struct mem_cgroup *memcg) argument
[all...]
/linux-master/mm/damon/
H A Dpaddr.c202 struct mem_cgroup *memcg; local
210 memcg = folio_memcg_check(folio);
211 if (!memcg)
214 matched = filter->memcg_id == mem_cgroup_id(memcg);
/linux-master/fs/notify/
H A Dgroup.c27 mem_cgroup_put(group->memcg);
/linux-master/kernel/bpf/
H A Dmemalloc.c208 struct mem_cgroup *memcg = NULL, *old_memcg; local
239 memcg = get_memcg(c);
240 old_memcg = set_active_memcg(memcg);
253 mem_cgroup_put(memcg);
998 struct mem_cgroup *memcg, *old_memcg; local
1000 memcg = get_memcg(c);
1001 old_memcg = set_active_memcg(memcg);
1006 mem_cgroup_put(memcg);

Completed in 689 milliseconds

12