/linux-master/include/linux/ |
H A D | list_lru.h | 27 LRU_STOP, /* stop lru list walking. May drop the lock 47 struct list_lru_one lru; member in struct:list_lru_node 61 void list_lru_destroy(struct list_lru *lru); 62 int __list_lru_init(struct list_lru *lru, bool memcg_aware, 65 #define list_lru_init(lru) \ 66 __list_lru_init((lru), false, NULL, NULL) 67 #define list_lru_init_memcg(lru, shrinker) \ 68 __list_lru_init((lru), true, NULL, shrinker) 70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, 75 * list_lru_add: add an element to the lru lis 153 list_lru_shrink_count(struct list_lru *lru, struct shrink_control *sc) argument 159 list_lru_count(struct list_lru *lru) argument 225 list_lru_shrink_walk(struct list_lru *lru, struct shrink_control *sc, list_lru_walk_cb isolate, void *cb_arg) argument 233 list_lru_shrink_walk_irq(struct list_lru *lru, struct shrink_control *sc, list_lru_walk_cb isolate, void *cb_arg) argument 241 list_lru_walk(struct list_lru *lru, list_lru_walk_cb isolate, void *cb_arg, unsigned long nr_to_walk) argument [all...] |
H A D | mm_inline.h | 39 enum lru_list lru, enum zone_type zid, 47 __mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); 49 NR_ZONE_LRU_BASE + lru, nr_pages); 53 enum lru_list lru, enum zone_type zid, 56 __update_lru_size(lruvec, lru, zid, nr_pages); 58 mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); 63 * __folio_clear_lru_flags - Clear page lru flags before releasing a page. 64 * @folio: The folio that was on lru and now has a zero reference. 89 enum lru_list lru; local 96 lru 38 __update_lru_size(struct lruvec *lruvec, enum lru_list lru, enum zone_type zid, long nr_pages) argument 52 update_lru_size(struct lruvec *lruvec, enum lru_list lru, enum zone_type zid, long nr_pages) argument 181 enum lru_list lru = type * LRU_INACTIVE_FILE; local 321 enum lru_list lru = folio_lru_list(folio); local 335 enum lru_list lru = folio_lru_list(folio); local 349 enum lru_list lru = folio_lru_list(folio); local [all...] |
H A D | balloon_compaction.h | 7 * Balloon page migration makes use of the general non-lru movable page 11 * page->mapping is used in context of non-lru page migration to reference 97 list_add(&page->lru, &balloon->pages); 114 * No touch page.lru field once @page has been isolated 118 list_del(&page->lru); 141 list_add(&page->lru, &balloon->pages); 147 list_del(&page->lru); 166 list_add(&page->lru, pages); 178 struct page *page = list_first_entry_or_null(pages, struct page, lru); 183 list_del(&page->lru); [all...] |
/linux-master/mm/ |
H A D | list_lru.c | 22 static inline bool list_lru_memcg_aware(struct list_lru *lru) argument 24 return lru->memcg_aware; 27 static void list_lru_register(struct list_lru *lru) argument 29 if (!list_lru_memcg_aware(lru)) 33 list_add(&lru->list, &memcg_list_lrus); 37 static void list_lru_unregister(struct list_lru *lru) argument 39 if (!list_lru_memcg_aware(lru)) 43 list_del(&lru->list); 47 static int lru_shrinker_id(struct list_lru *lru) argument 49 return lru 53 list_lru_from_memcg_idx(struct list_lru *lru, int nid, int idx) argument 63 list_lru_register(struct list_lru *lru) argument 67 list_lru_unregister(struct list_lru *lru) argument 71 lru_shrinker_id(struct list_lru *lru) argument 76 list_lru_memcg_aware(struct list_lru *lru) argument 82 list_lru_from_memcg_idx(struct list_lru *lru, int nid, int idx) argument 88 list_lru_add(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument 110 list_lru_add_obj(struct list_lru *lru, struct list_head *item) argument 120 list_lru_del(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument 140 list_lru_del_obj(struct list_lru *lru, struct list_head *item) argument 165 list_lru_count_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg) argument 183 list_lru_count_node(struct list_lru *lru, int nid) argument 193 __list_lru_walk_one(struct list_lru *lru, int nid, int memcg_idx, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 258 list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 274 list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 288 list_lru_walk_node(struct list_lru *lru, int nid, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 343 memcg_list_lru_free(struct list_lru *lru, int src_idx) argument 357 memcg_init_list_lru(struct list_lru *lru, bool memcg_aware) argument 364 memcg_destroy_list_lru(struct list_lru *lru) argument 380 memcg_reparent_list_lru_node(struct list_lru *lru, int nid, int src_idx, struct mem_cgroup *dst_memcg) argument 409 memcg_reparent_list_lru(struct list_lru *lru, int src_idx, struct mem_cgroup *dst_memcg) argument 423 struct list_lru *lru; local 454 memcg_list_lru_allocated(struct mem_cgroup *memcg, struct list_lru *lru) argument 462 memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, gfp_t gfp) argument 540 memcg_init_list_lru(struct list_lru *lru, bool memcg_aware) argument 544 memcg_destroy_list_lru(struct list_lru *lru) argument 549 __list_lru_init(struct list_lru *lru, bool memcg_aware, struct lock_class_key *key, struct shrinker *shrinker) argument 582 list_lru_destroy(struct list_lru *lru) argument [all...] |
H A D | balloon_compaction.c | 48 list_for_each_entry_safe(page, tmp, pages, lru) { 49 list_del(&page->lru); 84 list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) { 104 list_add(&page->lru, pages); 200 return list_first_entry(&pages, struct page, lru); 213 list_del(&page->lru); 226 list_add(&page->lru, &b_dev_info->pages);
|
H A D | mmzone.c | 77 enum lru_list lru; local 83 for_each_lru(lru) 84 INIT_LIST_HEAD(&lruvec->lists[lru]); 88 * (so that their lru fields can be reused to hold mlock_count).
|
H A D | page_reporting.c | 183 list_for_each_entry_safe(page, next, list, lru) { 218 if (!list_is_first(&page->lru, list)) 219 list_rotate_to_front(&page->lru, list); 243 next = list_first_entry(list, struct page, lru); 251 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) 252 list_rotate_to_front(&next->lru, list);
|
H A D | hugetlb_vmemmap.c | 198 list_for_each_entry_safe(page, next, list, lru) 216 list_del(&walk->reuse_page->lru); 227 list_add(&page->lru, walk->vmemmap_pages); 259 page = list_first_entry(walk->vmemmap_pages, struct page, lru); 260 list_del(&page->lru); 340 list_add(&walk.reuse_page->lru, vmemmap_pages); 392 list_add(&page->lru, list); 397 list_for_each_entry_safe(page, next, list, lru) 508 list_for_each_entry_safe(folio, t_folio, folio_list, lru) { 518 list_move(&folio->lru, non_hvo_folio [all...] |
/linux-master/kernel/bpf/ |
H A D | bpf_lru_list.c | 144 static void __bpf_lru_list_rotate_active(struct bpf_lru *lru, argument 158 if (++i == lru->nr_scans || node == first_node) 171 static void __bpf_lru_list_rotate_inactive(struct bpf_lru *lru, argument 187 while (i < lru->nr_scans) { 211 __bpf_lru_list_shrink_inactive(struct bpf_lru *lru, argument 225 } else if (lru->del_from_htab(lru->del_arg, node)) { 232 if (++i == lru->nr_scans) 242 static void __bpf_lru_list_rotate(struct bpf_lru *lru, struct bpf_lru_list *l) argument 245 __bpf_lru_list_rotate_active(lru, 260 __bpf_lru_list_shrink(struct bpf_lru *lru, struct bpf_lru_list *l, unsigned int tgt_nshrink, struct list_head *free_list, enum bpf_lru_list_type tgt_free_type) argument 323 bpf_lru_list_pop_free_to_local(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l) argument 352 __local_list_add_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l, int cpu, struct bpf_lru_node *node, u32 hash) argument 380 __local_list_pop_pending(struct bpf_lru *lru, struct bpf_lru_locallist *loc_l) argument 404 bpf_percpu_lru_pop_free(struct bpf_lru *lru, u32 hash) argument 436 bpf_common_lru_pop_free(struct bpf_lru *lru, u32 hash) argument 499 bpf_lru_pop_free(struct bpf_lru *lru, u32 hash) argument 507 bpf_common_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node) argument 541 bpf_percpu_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node) argument 556 bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node) argument 564 bpf_common_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, u32 elem_size, u32 nr_elems) argument 582 bpf_percpu_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, u32 elem_size, u32 nr_elems) argument 613 bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, u32 elem_size, u32 nr_elems) argument 651 bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, del_from_htab_func del_from_htab, void *del_arg) argument 694 bpf_lru_destroy(struct bpf_lru *lru) argument [all...] |
H A D | bpf_lru_list.h | 71 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, 73 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, 75 void bpf_lru_destroy(struct bpf_lru *lru); 76 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash); 77 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node);
|
/linux-master/net/ceph/ |
H A D | pagelist.c | 32 struct page *page = list_entry(pl->head.prev, struct page, lru); 45 lru); 46 list_del(&page->lru); 61 page = list_first_entry(&pl->free_list, struct page, lru); 62 list_del(&page->lru); 69 list_add_tail(&page->lru, &pl->head); 113 list_add_tail(&page->lru, &pl->free_list); 125 struct page, lru); 126 list_del(&page->lru); 159 page = list_entry(pl->head.prev, struct page, lru); [all...] |
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_gem_shrinker.c | 38 unsigned count = priv->lru.dontneed.count; 41 count += priv->lru.willneed.count; 104 struct drm_gem_lru *lru; member in struct:__anon753 111 { &priv->lru.dontneed, purge, true }, 112 { &priv->lru.willneed, evict, can_swap() }, 113 { &priv->lru.dontneed, active_purge, can_block(sc) }, 114 { &priv->lru.willneed, active_evict, can_swap() && can_block(sc) }, 124 drm_gem_lru_scan(stages[i].lru, nr, 183 &priv->lru.dontneed, 184 &priv->lru [all...] |
/linux-master/samples/bpf/ |
H A D | test_lru_dist.c | 99 static void pfect_lru_init(struct pfect_lru *lru, unsigned int lru_size, argument 102 lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL, 106 assert(lru->map_fd != -1); 108 lru->free_nodes = malloc(lru_size * sizeof(struct pfect_lru_node)); 109 assert(lru->free_nodes); 111 INIT_LIST_HEAD(&lru->list); 112 lru->cur_size = 0; 113 lru->lru_size = lru_size; 114 lru->nr_unique = lru 117 pfect_lru_destroy(struct pfect_lru *lru) argument 123 pfect_lru_lookup_or_insert(struct pfect_lru *lru, unsigned long long key) argument [all...] |
/linux-master/fs/bcachefs/ |
H A D | lru.c | 9 #include "lru.h" 21 "lru entry at time=0"); 29 const struct bch_lru *lru = bkey_s_c_to_lru(k).v; local 31 prt_printf(out, "idx %llu", le64_to_cpu(lru->idx)); 34 void bch2_lru_pos_to_text(struct printbuf *out, struct bpos lru) argument 37 lru_pos_id(lru), 38 lru_pos_time(lru), 39 u64_to_bucket(lru.offset).inode, 40 u64_to_bucket(lru.offset).offset); 99 "lru ke [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 38 struct list_head lru; member in struct:nv50_instmem 53 struct list_head lru; member in struct:nv50_instobj 143 eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru); 149 list_del_init(&eobj->lru); 205 if (likely(iobj->lru.next) && iobj->map) { 206 BUG_ON(!list_empty(&iobj->lru)); 207 list_add_tail(&iobj->lru, &imem->lru); 249 if (likely(iobj->lru [all...] |
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_resource.c | 73 list_bulk_move_tail(&man->lru[j], &pos->first->lru, 74 &pos->last->lru); 93 pos->first = list_next_entry(res, lru); 94 list_move(&res->lru, &pos->last->lru); 125 pos->first = list_next_entry(res, lru); 127 pos->last = list_prev_entry(res, lru); 129 list_move(&res->lru, &pos->last->lru); [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_gem.c | 502 * Move folios to appropriate lru and release the folios, decrementing the 1343 * @lru: The LRU to initialize 1347 drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock) argument 1349 lru->lock = lock; 1350 lru->count = 0; 1351 INIT_LIST_HEAD(&lru->list); 1358 obj->lru->count -= obj->size >> PAGE_SHIFT; 1359 WARN_ON(obj->lru->count < 0); 1361 obj->lru = NULL; 1374 struct drm_gem_lru *lru local 1394 drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj) argument 1418 drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj) argument 1442 drm_gem_lru_scan(struct drm_gem_lru *lru, unsigned int nr_to_scan, unsigned long *remaining, bool (*shrink)(struct drm_gem_object *obj)) argument [all...] |
/linux-master/include/trace/events/ |
H A D | pagemap.h | 37 __field(enum lru_list, lru ) 44 __entry->lru = folio_lru_list(folio); 49 TP_printk("folio=%p pfn=0x%lx lru=%d flags=%s%s%s%s%s%s", 52 __entry->lru,
|
/linux-master/fs/nfs/ |
H A D | nfs42xattr.c | 71 struct list_head lru; member in struct:nfs4_xattr_cache 82 struct list_head lru; member in struct:nfs4_xattr_entry 130 struct list_lru *lru; local 132 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? 135 return list_lru_add_obj(lru, &entry->lru); 141 struct list_lru *lru; local 143 lru = (entry->flags & NFS4_XATTR_ENTRY_EXTVAL) ? 146 return list_lru_del_obj(lru, &entry->lru); 804 cache_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument 869 entry_lru_isolate(struct list_head *item, struct list_lru_one *lru, spinlock_t *lru_lock, void *arg) argument 925 struct list_lru *lru; local 953 struct list_lru *lru; local 980 nfs4_xattr_shrinker_init(struct shrinker **shrinker, struct list_lru *lru, const char *name, count_objects_cb count, scan_objects_cb scan, long batch, int seeks) argument 1007 nfs4_xattr_shrinker_destroy(struct shrinker *shrinker, struct list_lru *lru) argument [all...] |
/linux-master/drivers/md/ |
H A D | dm-bufio.c | 86 struct lru *lru; member in struct:lru_iter 92 struct lru { struct 101 static void lru_init(struct lru *lru) argument 103 lru->cursor = NULL; 104 lru->count = 0; 105 INIT_LIST_HEAD(&lru->iterators); 108 static void lru_destroy(struct lru *lru) argument 117 lru_insert(struct lru *lru, struct lru_entry *le) argument 147 lru_iter_begin(struct lru *lru, struct lru_iter *it) argument 196 lru_iter_invalidate(struct lru *lru, struct lru_entry *e) argument 222 lru_remove(struct lru *lru, struct lru_entry *le) argument 257 lru_evict(struct lru *lru, le_predicate pred, void *context, bool no_sleep) argument 347 struct lru_entry lru; member in struct:dm_buffer 394 struct lru lru[LIST_SIZE]; member in struct:dm_buffer_cache 783 struct lru *lru = &bc->lru[list_mode]; local [all...] |
/linux-master/arch/mips/mm/ |
H A D | cerr-sb1.c | 326 uint8_t lru; local 348 lru = (taghi >> 14) & 0xff; 352 (lru & 0x3), 353 ((lru >> 2) & 0x3), 354 ((lru >> 4) & 0x3), 355 ((lru >> 6) & 0x3)); 481 uint8_t ecc, lru; local 503 lru = (taghi >> 14) & 0xff; 507 (lru & 0x3), 508 ((lru >> [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | wq.c | 50 } lru SEC(".maps"); 65 if (map == &lru && 179 return test_elem_callback(&lru, &key, wq_callback);
|
H A D | wq_failures.c | 28 } lru SEC(".maps"); 88 if (bpf_wq_init(wq, &lru, 0) != 0)
|
/linux-master/drivers/iommu/ |
H A D | iommu-pages.h | 168 * @page: the head of the lru list to be freed. 178 struct page *p = list_entry(page->prev, struct page, lru); 180 list_del(&p->lru);
|
/linux-master/fs/proc/ |
H A D | meminfo.c | 42 int lru; local 53 for (lru = LRU_BASE; lru < NR_LRU_LISTS; lru++) 54 pages[lru] = global_node_page_state(NR_LRU_BASE + lru);
|