xref: /openbsd-src/sys/dev/pci/drm/i915/gt/intel_gt_buffer_pool.c (revision f005ef32267c16bdb134f0e9fa4477dbe07c263a)
1ad8b1aafSjsg // SPDX-License-Identifier: MIT
2ad8b1aafSjsg /*
3ad8b1aafSjsg  * Copyright © 2014-2018 Intel Corporation
4ad8b1aafSjsg  */
5ad8b1aafSjsg 
61bb76ff1Sjsg #include "gem/i915_gem_internal.h"
7ad8b1aafSjsg #include "gem/i915_gem_object.h"
8ad8b1aafSjsg 
9ad8b1aafSjsg #include "i915_drv.h"
10ad8b1aafSjsg #include "intel_engine_pm.h"
11ad8b1aafSjsg #include "intel_gt_buffer_pool.h"
12ad8b1aafSjsg 
13ad8b1aafSjsg static struct list_head *
bucket_for_size(struct intel_gt_buffer_pool * pool,size_t sz)14ad8b1aafSjsg bucket_for_size(struct intel_gt_buffer_pool *pool, size_t sz)
15ad8b1aafSjsg {
16ad8b1aafSjsg 	int n;
17ad8b1aafSjsg 
18ad8b1aafSjsg 	/*
19ad8b1aafSjsg 	 * Compute a power-of-two bucket, but throw everything greater than
20ad8b1aafSjsg 	 * 16KiB into the same bucket: i.e. the buckets hold objects of
21ad8b1aafSjsg 	 * (1 page, 2 pages, 4 pages, 8+ pages).
22ad8b1aafSjsg 	 */
23ad8b1aafSjsg 	n = fls(sz >> PAGE_SHIFT) - 1;
24ad8b1aafSjsg 	if (n >= ARRAY_SIZE(pool->cache_list))
25ad8b1aafSjsg 		n = ARRAY_SIZE(pool->cache_list) - 1;
26ad8b1aafSjsg 
27ad8b1aafSjsg 	return &pool->cache_list[n];
28ad8b1aafSjsg }
29ad8b1aafSjsg 
node_free(struct intel_gt_buffer_pool_node * node)30ad8b1aafSjsg static void node_free(struct intel_gt_buffer_pool_node *node)
31ad8b1aafSjsg {
32ad8b1aafSjsg 	i915_gem_object_put(node->obj);
33ad8b1aafSjsg 	i915_active_fini(&node->active);
34ad8b1aafSjsg 	kfree_rcu(node, rcu);
35ad8b1aafSjsg }
36ad8b1aafSjsg 
pool_free_older_than(struct intel_gt_buffer_pool * pool,long keep)37ad8b1aafSjsg static bool pool_free_older_than(struct intel_gt_buffer_pool *pool, long keep)
38ad8b1aafSjsg {
39ad8b1aafSjsg 	struct intel_gt_buffer_pool_node *node, *stale = NULL;
40ad8b1aafSjsg 	bool active = false;
41ad8b1aafSjsg 	int n;
42ad8b1aafSjsg 
43ad8b1aafSjsg 	/* Free buffers that have not been used in the past second */
44ad8b1aafSjsg 	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) {
45ad8b1aafSjsg 		struct list_head *list = &pool->cache_list[n];
46ad8b1aafSjsg 
47ad8b1aafSjsg 		if (list_empty(list))
48ad8b1aafSjsg 			continue;
49ad8b1aafSjsg 
50ad8b1aafSjsg 		if (spin_trylock_irq(&pool->lock)) {
51ad8b1aafSjsg 			struct list_head *pos;
52ad8b1aafSjsg 
53ad8b1aafSjsg 			/* Most recent at head; oldest at tail */
54ad8b1aafSjsg 			list_for_each_prev(pos, list) {
55ad8b1aafSjsg 				unsigned long age;
56ad8b1aafSjsg 
57ad8b1aafSjsg 				node = list_entry(pos, typeof(*node), link);
58ad8b1aafSjsg 
59ad8b1aafSjsg 				age = READ_ONCE(node->age);
60ad8b1aafSjsg 				if (!age || jiffies - age < keep)
61ad8b1aafSjsg 					break;
62ad8b1aafSjsg 
63ad8b1aafSjsg 				/* Check we are the first to claim this node */
64ad8b1aafSjsg 				if (!xchg(&node->age, 0))
65ad8b1aafSjsg 					break;
66ad8b1aafSjsg 
67ad8b1aafSjsg 				node->free = stale;
68ad8b1aafSjsg 				stale = node;
69ad8b1aafSjsg 			}
70ad8b1aafSjsg 			if (!list_is_last(pos, list))
71ad8b1aafSjsg 				__list_del_many(pos, list);
72ad8b1aafSjsg 
73ad8b1aafSjsg 			spin_unlock_irq(&pool->lock);
74ad8b1aafSjsg 		}
75ad8b1aafSjsg 
76ad8b1aafSjsg 		active |= !list_empty(list);
77ad8b1aafSjsg 	}
78ad8b1aafSjsg 
79ad8b1aafSjsg 	while ((node = stale)) {
80ad8b1aafSjsg 		stale = stale->free;
81ad8b1aafSjsg 		node_free(node);
82ad8b1aafSjsg 	}
83ad8b1aafSjsg 
84ad8b1aafSjsg 	return active;
85ad8b1aafSjsg }
86ad8b1aafSjsg 
pool_free_work(struct work_struct * wrk)87ad8b1aafSjsg static void pool_free_work(struct work_struct *wrk)
88ad8b1aafSjsg {
89ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool =
90ad8b1aafSjsg 		container_of(wrk, typeof(*pool), work.work);
91*f005ef32Sjsg 	struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
92ad8b1aafSjsg 
93ad8b1aafSjsg 	if (pool_free_older_than(pool, HZ))
94*f005ef32Sjsg 		queue_delayed_work(gt->i915->unordered_wq, &pool->work,
95ad8b1aafSjsg 				   round_jiffies_up_relative(HZ));
96ad8b1aafSjsg }
97ad8b1aafSjsg 
pool_retire(struct i915_active * ref)98ad8b1aafSjsg static void pool_retire(struct i915_active *ref)
99ad8b1aafSjsg {
100ad8b1aafSjsg 	struct intel_gt_buffer_pool_node *node =
101ad8b1aafSjsg 		container_of(ref, typeof(*node), active);
102ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool = node->pool;
103*f005ef32Sjsg 	struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
104ad8b1aafSjsg 	struct list_head *list = bucket_for_size(pool, node->obj->base.size);
105ad8b1aafSjsg 	unsigned long flags;
106ad8b1aafSjsg 
1075ca02815Sjsg 	if (node->pinned) {
108ad8b1aafSjsg 		i915_gem_object_unpin_pages(node->obj);
109ad8b1aafSjsg 
110ad8b1aafSjsg 		/* Return this object to the shrinker pool */
111ad8b1aafSjsg 		i915_gem_object_make_purgeable(node->obj);
1125ca02815Sjsg 		node->pinned = false;
1135ca02815Sjsg 	}
114ad8b1aafSjsg 
115ad8b1aafSjsg 	GEM_BUG_ON(node->age);
116ad8b1aafSjsg 	spin_lock_irqsave(&pool->lock, flags);
117ad8b1aafSjsg 	list_add_rcu(&node->link, list);
118ad8b1aafSjsg 	WRITE_ONCE(node->age, jiffies ?: 1); /* 0 reserved for active nodes */
119ad8b1aafSjsg 	spin_unlock_irqrestore(&pool->lock, flags);
120ad8b1aafSjsg 
121*f005ef32Sjsg 	queue_delayed_work(gt->i915->unordered_wq, &pool->work,
122ad8b1aafSjsg 			   round_jiffies_up_relative(HZ));
123ad8b1aafSjsg }
124ad8b1aafSjsg 
intel_gt_buffer_pool_mark_used(struct intel_gt_buffer_pool_node * node)1255ca02815Sjsg void intel_gt_buffer_pool_mark_used(struct intel_gt_buffer_pool_node *node)
1265ca02815Sjsg {
1275ca02815Sjsg 	assert_object_held(node->obj);
1285ca02815Sjsg 
1295ca02815Sjsg 	if (node->pinned)
1305ca02815Sjsg 		return;
1315ca02815Sjsg 
1325ca02815Sjsg 	__i915_gem_object_pin_pages(node->obj);
1335ca02815Sjsg 	/* Hide this pinned object from the shrinker until retired */
1345ca02815Sjsg 	i915_gem_object_make_unshrinkable(node->obj);
1355ca02815Sjsg 	node->pinned = true;
1365ca02815Sjsg }
1375ca02815Sjsg 
138ad8b1aafSjsg static struct intel_gt_buffer_pool_node *
node_create(struct intel_gt_buffer_pool * pool,size_t sz,enum i915_map_type type)1395ca02815Sjsg node_create(struct intel_gt_buffer_pool *pool, size_t sz,
1405ca02815Sjsg 	    enum i915_map_type type)
141ad8b1aafSjsg {
1421bb76ff1Sjsg 	struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
143ad8b1aafSjsg 	struct intel_gt_buffer_pool_node *node;
144ad8b1aafSjsg 	struct drm_i915_gem_object *obj;
145ad8b1aafSjsg 
146ad8b1aafSjsg 	node = kmalloc(sizeof(*node),
147ad8b1aafSjsg 		       GFP_KERNEL | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
148ad8b1aafSjsg 	if (!node)
149ad8b1aafSjsg 		return ERR_PTR(-ENOMEM);
150ad8b1aafSjsg 
151ad8b1aafSjsg 	node->age = 0;
152ad8b1aafSjsg 	node->pool = pool;
1535ca02815Sjsg 	node->pinned = false;
1545ca02815Sjsg 	i915_active_init(&node->active, NULL, pool_retire, 0);
155ad8b1aafSjsg 
156ad8b1aafSjsg 	obj = i915_gem_object_create_internal(gt->i915, sz);
157ad8b1aafSjsg 	if (IS_ERR(obj)) {
158ad8b1aafSjsg 		i915_active_fini(&node->active);
159ad8b1aafSjsg 		kfree(node);
160ad8b1aafSjsg 		return ERR_CAST(obj);
161ad8b1aafSjsg 	}
162ad8b1aafSjsg 
163ad8b1aafSjsg 	i915_gem_object_set_readonly(obj);
164ad8b1aafSjsg 
1655ca02815Sjsg 	node->type = type;
166ad8b1aafSjsg 	node->obj = obj;
167ad8b1aafSjsg 	return node;
168ad8b1aafSjsg }
169ad8b1aafSjsg 
170ad8b1aafSjsg struct intel_gt_buffer_pool_node *
intel_gt_get_buffer_pool(struct intel_gt * gt,size_t size,enum i915_map_type type)1715ca02815Sjsg intel_gt_get_buffer_pool(struct intel_gt *gt, size_t size,
1725ca02815Sjsg 			 enum i915_map_type type)
173ad8b1aafSjsg {
174ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
175ad8b1aafSjsg 	struct intel_gt_buffer_pool_node *node;
176ad8b1aafSjsg 	struct list_head *list;
177ad8b1aafSjsg 	int ret;
178ad8b1aafSjsg 
179ad8b1aafSjsg 	size = PAGE_ALIGN(size);
180ad8b1aafSjsg 	list = bucket_for_size(pool, size);
181ad8b1aafSjsg 
182ad8b1aafSjsg 	rcu_read_lock();
183ad8b1aafSjsg 	list_for_each_entry_rcu(node, list, link) {
184ad8b1aafSjsg 		unsigned long age;
185ad8b1aafSjsg 
186ad8b1aafSjsg 		if (node->obj->base.size < size)
187ad8b1aafSjsg 			continue;
188ad8b1aafSjsg 
1895ca02815Sjsg 		if (node->type != type)
1905ca02815Sjsg 			continue;
1915ca02815Sjsg 
192ad8b1aafSjsg 		age = READ_ONCE(node->age);
193ad8b1aafSjsg 		if (!age)
194ad8b1aafSjsg 			continue;
195ad8b1aafSjsg 
196ad8b1aafSjsg 		if (cmpxchg(&node->age, age, 0) == age) {
197ad8b1aafSjsg 			spin_lock_irq(&pool->lock);
198ad8b1aafSjsg 			list_del_rcu(&node->link);
199ad8b1aafSjsg 			spin_unlock_irq(&pool->lock);
200ad8b1aafSjsg 			break;
201ad8b1aafSjsg 		}
202ad8b1aafSjsg 	}
203ad8b1aafSjsg 	rcu_read_unlock();
204ad8b1aafSjsg 
205ad8b1aafSjsg 	if (&node->link == list) {
2065ca02815Sjsg 		node = node_create(pool, size, type);
207ad8b1aafSjsg 		if (IS_ERR(node))
208ad8b1aafSjsg 			return node;
209ad8b1aafSjsg 	}
210ad8b1aafSjsg 
211ad8b1aafSjsg 	ret = i915_active_acquire(&node->active);
212ad8b1aafSjsg 	if (ret) {
213ad8b1aafSjsg 		node_free(node);
214ad8b1aafSjsg 		return ERR_PTR(ret);
215ad8b1aafSjsg 	}
216ad8b1aafSjsg 
217ad8b1aafSjsg 	return node;
218ad8b1aafSjsg }
219ad8b1aafSjsg 
intel_gt_init_buffer_pool(struct intel_gt * gt)220ad8b1aafSjsg void intel_gt_init_buffer_pool(struct intel_gt *gt)
221ad8b1aafSjsg {
222ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
223ad8b1aafSjsg 	int n;
224ad8b1aafSjsg 
225ad8b1aafSjsg 	mtx_init(&pool->lock, IPL_TTY);
226ad8b1aafSjsg 	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
227ad8b1aafSjsg 		INIT_LIST_HEAD(&pool->cache_list[n]);
228ad8b1aafSjsg 	INIT_DELAYED_WORK(&pool->work, pool_free_work);
229ad8b1aafSjsg }
230ad8b1aafSjsg 
intel_gt_flush_buffer_pool(struct intel_gt * gt)231ad8b1aafSjsg void intel_gt_flush_buffer_pool(struct intel_gt *gt)
232ad8b1aafSjsg {
233ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
234ad8b1aafSjsg 
235ad8b1aafSjsg 	do {
236ad8b1aafSjsg 		while (pool_free_older_than(pool, 0))
237ad8b1aafSjsg 			;
238ad8b1aafSjsg 	} while (cancel_delayed_work_sync(&pool->work));
239ad8b1aafSjsg }
240ad8b1aafSjsg 
intel_gt_fini_buffer_pool(struct intel_gt * gt)241ad8b1aafSjsg void intel_gt_fini_buffer_pool(struct intel_gt *gt)
242ad8b1aafSjsg {
243ad8b1aafSjsg 	struct intel_gt_buffer_pool *pool = &gt->buffer_pool;
244ad8b1aafSjsg 	int n;
245ad8b1aafSjsg 
246ad8b1aafSjsg 	for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
247ad8b1aafSjsg 		GEM_BUG_ON(!list_empty(&pool->cache_list[n]));
248ad8b1aafSjsg }
249