848a4e5c09
In order to avoid flush_scheduled_work() usage, add a dedicated workqueue in the drm_i915_private structure. In this way, we don't need to use the system queue anymore. This change is mostly mechanical and based on Tetsuo's original patch[1]. v6 by Jani: - Also create unordered_wq for mock device Link: https://patchwork.freedesktop.org/series/114608/ [1] Cc: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp> Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Cc: Jani Nikula <jani.nikula@intel.com> Cc: Ville Syrjälä <ville.syrjala@linux.intel.com> Reviewed-by: Jani Nikula <jani.nikula@intel.com> Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com> Signed-off-by: Luca Coelho <luciano.coelho@intel.com> Signed-off-by: Jani Nikula <jani.nikula@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/c816ebe17ef08d363981942a096a586a7658a65e.1686231190.git.jani.nikula@intel.com
249 lines
5.7 KiB
C
249 lines
5.7 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2014-2018 Intel Corporation
|
|
*/
|
|
|
|
#include "gem/i915_gem_internal.h"
|
|
#include "gem/i915_gem_object.h"
|
|
|
|
#include "i915_drv.h"
|
|
#include "intel_engine_pm.h"
|
|
#include "intel_gt_buffer_pool.h"
|
|
|
|
static struct list_head *
|
|
bucket_for_size(struct intel_gt_buffer_pool *pool, size_t sz)
|
|
{
|
|
int n;
|
|
|
|
/*
|
|
* Compute a power-of-two bucket, but throw everything greater than
|
|
* 16KiB into the same bucket: i.e. the buckets hold objects of
|
|
* (1 page, 2 pages, 4 pages, 8+ pages).
|
|
*/
|
|
n = fls(sz >> PAGE_SHIFT) - 1;
|
|
if (n >= ARRAY_SIZE(pool->cache_list))
|
|
n = ARRAY_SIZE(pool->cache_list) - 1;
|
|
|
|
return &pool->cache_list[n];
|
|
}
|
|
|
|
static void node_free(struct intel_gt_buffer_pool_node *node)
|
|
{
|
|
i915_gem_object_put(node->obj);
|
|
i915_active_fini(&node->active);
|
|
kfree_rcu(node, rcu);
|
|
}
|
|
|
|
static bool pool_free_older_than(struct intel_gt_buffer_pool *pool, long keep)
|
|
{
|
|
struct intel_gt_buffer_pool_node *node, *stale = NULL;
|
|
bool active = false;
|
|
int n;
|
|
|
|
/* Free buffers that have not been used in the past second */
|
|
for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++) {
|
|
struct list_head *list = &pool->cache_list[n];
|
|
|
|
if (list_empty(list))
|
|
continue;
|
|
|
|
if (spin_trylock_irq(&pool->lock)) {
|
|
struct list_head *pos;
|
|
|
|
/* Most recent at head; oldest at tail */
|
|
list_for_each_prev(pos, list) {
|
|
unsigned long age;
|
|
|
|
node = list_entry(pos, typeof(*node), link);
|
|
|
|
age = READ_ONCE(node->age);
|
|
if (!age || jiffies - age < keep)
|
|
break;
|
|
|
|
/* Check we are the first to claim this node */
|
|
if (!xchg(&node->age, 0))
|
|
break;
|
|
|
|
node->free = stale;
|
|
stale = node;
|
|
}
|
|
if (!list_is_last(pos, list))
|
|
__list_del_many(pos, list);
|
|
|
|
spin_unlock_irq(&pool->lock);
|
|
}
|
|
|
|
active |= !list_empty(list);
|
|
}
|
|
|
|
while ((node = stale)) {
|
|
stale = stale->free;
|
|
node_free(node);
|
|
}
|
|
|
|
return active;
|
|
}
|
|
|
|
static void pool_free_work(struct work_struct *wrk)
|
|
{
|
|
struct intel_gt_buffer_pool *pool =
|
|
container_of(wrk, typeof(*pool), work.work);
|
|
struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
|
|
|
|
if (pool_free_older_than(pool, HZ))
|
|
queue_delayed_work(gt->i915->unordered_wq, &pool->work,
|
|
round_jiffies_up_relative(HZ));
|
|
}
|
|
|
|
static void pool_retire(struct i915_active *ref)
|
|
{
|
|
struct intel_gt_buffer_pool_node *node =
|
|
container_of(ref, typeof(*node), active);
|
|
struct intel_gt_buffer_pool *pool = node->pool;
|
|
struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
|
|
struct list_head *list = bucket_for_size(pool, node->obj->base.size);
|
|
unsigned long flags;
|
|
|
|
if (node->pinned) {
|
|
i915_gem_object_unpin_pages(node->obj);
|
|
|
|
/* Return this object to the shrinker pool */
|
|
i915_gem_object_make_purgeable(node->obj);
|
|
node->pinned = false;
|
|
}
|
|
|
|
GEM_BUG_ON(node->age);
|
|
spin_lock_irqsave(&pool->lock, flags);
|
|
list_add_rcu(&node->link, list);
|
|
WRITE_ONCE(node->age, jiffies ?: 1); /* 0 reserved for active nodes */
|
|
spin_unlock_irqrestore(&pool->lock, flags);
|
|
|
|
queue_delayed_work(gt->i915->unordered_wq, &pool->work,
|
|
round_jiffies_up_relative(HZ));
|
|
}
|
|
|
|
void intel_gt_buffer_pool_mark_used(struct intel_gt_buffer_pool_node *node)
|
|
{
|
|
assert_object_held(node->obj);
|
|
|
|
if (node->pinned)
|
|
return;
|
|
|
|
__i915_gem_object_pin_pages(node->obj);
|
|
/* Hide this pinned object from the shrinker until retired */
|
|
i915_gem_object_make_unshrinkable(node->obj);
|
|
node->pinned = true;
|
|
}
|
|
|
|
static struct intel_gt_buffer_pool_node *
|
|
node_create(struct intel_gt_buffer_pool *pool, size_t sz,
|
|
enum i915_map_type type)
|
|
{
|
|
struct intel_gt *gt = container_of(pool, struct intel_gt, buffer_pool);
|
|
struct intel_gt_buffer_pool_node *node;
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
node = kmalloc(sizeof(*node),
|
|
GFP_KERNEL | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
|
|
if (!node)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
node->age = 0;
|
|
node->pool = pool;
|
|
node->pinned = false;
|
|
i915_active_init(&node->active, NULL, pool_retire, 0);
|
|
|
|
obj = i915_gem_object_create_internal(gt->i915, sz);
|
|
if (IS_ERR(obj)) {
|
|
i915_active_fini(&node->active);
|
|
kfree(node);
|
|
return ERR_CAST(obj);
|
|
}
|
|
|
|
i915_gem_object_set_readonly(obj);
|
|
|
|
node->type = type;
|
|
node->obj = obj;
|
|
return node;
|
|
}
|
|
|
|
struct intel_gt_buffer_pool_node *
|
|
intel_gt_get_buffer_pool(struct intel_gt *gt, size_t size,
|
|
enum i915_map_type type)
|
|
{
|
|
struct intel_gt_buffer_pool *pool = >->buffer_pool;
|
|
struct intel_gt_buffer_pool_node *node;
|
|
struct list_head *list;
|
|
int ret;
|
|
|
|
size = PAGE_ALIGN(size);
|
|
list = bucket_for_size(pool, size);
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(node, list, link) {
|
|
unsigned long age;
|
|
|
|
if (node->obj->base.size < size)
|
|
continue;
|
|
|
|
if (node->type != type)
|
|
continue;
|
|
|
|
age = READ_ONCE(node->age);
|
|
if (!age)
|
|
continue;
|
|
|
|
if (cmpxchg(&node->age, age, 0) == age) {
|
|
spin_lock_irq(&pool->lock);
|
|
list_del_rcu(&node->link);
|
|
spin_unlock_irq(&pool->lock);
|
|
break;
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
if (&node->link == list) {
|
|
node = node_create(pool, size, type);
|
|
if (IS_ERR(node))
|
|
return node;
|
|
}
|
|
|
|
ret = i915_active_acquire(&node->active);
|
|
if (ret) {
|
|
node_free(node);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
return node;
|
|
}
|
|
|
|
void intel_gt_init_buffer_pool(struct intel_gt *gt)
|
|
{
|
|
struct intel_gt_buffer_pool *pool = >->buffer_pool;
|
|
int n;
|
|
|
|
spin_lock_init(&pool->lock);
|
|
for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
|
|
INIT_LIST_HEAD(&pool->cache_list[n]);
|
|
INIT_DELAYED_WORK(&pool->work, pool_free_work);
|
|
}
|
|
|
|
void intel_gt_flush_buffer_pool(struct intel_gt *gt)
|
|
{
|
|
struct intel_gt_buffer_pool *pool = >->buffer_pool;
|
|
|
|
do {
|
|
while (pool_free_older_than(pool, 0))
|
|
;
|
|
} while (cancel_delayed_work_sync(&pool->work));
|
|
}
|
|
|
|
void intel_gt_fini_buffer_pool(struct intel_gt *gt)
|
|
{
|
|
struct intel_gt_buffer_pool *pool = >->buffer_pool;
|
|
int n;
|
|
|
|
for (n = 0; n < ARRAY_SIZE(pool->cache_list); n++)
|
|
GEM_BUG_ON(!list_empty(&pool->cache_list[n]));
|
|
}
|