Loading...
#ifndef IOU_ALLOC_CACHE_H #define IOU_ALLOC_CACHE_H /* * Don't allow the cache to grow beyond this size. */ #define IO_ALLOC_CACHE_MAX 512 struct io_cache_entry { struct io_wq_work_node node; }; static inline bool io_alloc_cache_put(struct io_alloc_cache *cache, struct io_cache_entry *entry) { if (cache->nr_cached < cache->max_cached) { cache->nr_cached++; wq_stack_add_head(&entry->node, &cache->list); /* KASAN poisons object */ kasan_slab_free_mempool(entry); return true; } return false; } static inline bool io_alloc_cache_empty(struct io_alloc_cache *cache) { return !cache->list.next; } static inline struct io_cache_entry *io_alloc_cache_get(struct io_alloc_cache *cache) { if (cache->list.next) { struct io_cache_entry *entry; entry = container_of(cache->list.next, struct io_cache_entry, node); kasan_unpoison_range(entry, cache->elem_size); cache->list.next = cache->list.next->next; cache->nr_cached--; return entry; } return NULL; } static inline void io_alloc_cache_init(struct io_alloc_cache *cache, unsigned max_nr, size_t size) { cache->list.next = NULL; cache->nr_cached = 0; cache->max_cached = max_nr; cache->elem_size = size; } static inline void io_alloc_cache_free(struct io_alloc_cache *cache, void (*free)(struct io_cache_entry *)) { while (1) { struct io_cache_entry *entry = io_alloc_cache_get(cache); if (!entry) break; free(entry); } cache->nr_cached = 0; } #endif |