drm/i915: move request slabs to direct module init/exit

With the global kmem_cache shrink infrastructure gone there's nothing
special and we can convert them over.

I'm doing this split up into each patch because there's quite a bit of
noise with removing the static global.slab_requests|execute_cbs to just a
slab_requests|execute_cbs.

v2: Make slab static (Jason, 0day)

Reviewed-by: Jason Ekstrand <jason@jlekstrand.net>
Cc: Jason Ekstrand <jason@jlekstrand.net>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20210727121037.2041102-7-daniel.vetter@ffwll.ch
This commit is contained in:
Daniel Vetter
2021-07-27 14:10:33 +02:00
parent c8ad09affd
commit 47514ac752
4 changed files with 24 additions and 30 deletions

View File

@@ -42,7 +42,6 @@
#include "i915_active.h"
#include "i915_drv.h"
#include "i915_globals.h"
#include "i915_trace.h"
#include "intel_pm.h"
@@ -52,11 +51,8 @@ struct execute_cb {
struct i915_request *signal;
};
static struct i915_global_request {
struct i915_global base;
struct kmem_cache *slab_requests;
struct kmem_cache *slab_execute_cbs;
} global;
static struct kmem_cache *slab_requests;
static struct kmem_cache *slab_execute_cbs;
static const char *i915_fence_get_driver_name(struct dma_fence *fence)
{
@@ -107,7 +103,7 @@ static signed long i915_fence_wait(struct dma_fence *fence,
struct kmem_cache *i915_request_slab_cache(void)
{
return global.slab_requests;
return slab_requests;
}
static void i915_fence_release(struct dma_fence *fence)
@@ -140,7 +136,7 @@ static void i915_fence_release(struct dma_fence *fence)
intel_context_put(rq->context);
kmem_cache_free(global.slab_requests, rq);
kmem_cache_free(slab_requests, rq);
}
const struct dma_fence_ops i915_fence_ops = {
@@ -157,7 +153,7 @@ static void irq_execute_cb(struct irq_work *wrk)
struct execute_cb *cb = container_of(wrk, typeof(*cb), work);
i915_sw_fence_complete(cb->fence);
kmem_cache_free(global.slab_execute_cbs, cb);
kmem_cache_free(slab_execute_cbs, cb);
}
static __always_inline void
@@ -462,7 +458,7 @@ __await_execution(struct i915_request *rq,
if (i915_request_is_active(signal))
return 0;
cb = kmem_cache_alloc(global.slab_execute_cbs, gfp);
cb = kmem_cache_alloc(slab_execute_cbs, gfp);
if (!cb)
return -ENOMEM;
@@ -808,7 +804,7 @@ request_alloc_slow(struct intel_timeline *tl,
rq = list_first_entry(&tl->requests, typeof(*rq), link);
i915_request_retire(rq);
rq = kmem_cache_alloc(global.slab_requests,
rq = kmem_cache_alloc(slab_requests,
gfp | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
if (rq)
return rq;
@@ -821,7 +817,7 @@ request_alloc_slow(struct intel_timeline *tl,
retire_requests(tl);
out:
return kmem_cache_alloc(global.slab_requests, gfp);
return kmem_cache_alloc(slab_requests, gfp);
}
static void __i915_request_ctor(void *arg)
@@ -882,7 +878,7 @@ __i915_request_create(struct intel_context *ce, gfp_t gfp)
*
* Do not use kmem_cache_zalloc() here!
*/
rq = kmem_cache_alloc(global.slab_requests,
rq = kmem_cache_alloc(slab_requests,
gfp | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
if (unlikely(!rq)) {
rq = request_alloc_slow(tl, &ce->engine->request_pool, gfp);
@@ -984,7 +980,7 @@ err_unwind:
err_free:
intel_context_put(ce);
kmem_cache_free(global.slab_requests, rq);
kmem_cache_free(slab_requests, rq);
err_unreserve:
intel_context_unpin(ce);
return ERR_PTR(ret);
@@ -2080,19 +2076,15 @@ enum i915_request_state i915_test_request_state(struct i915_request *rq)
#include "selftests/i915_request.c"
#endif
static void i915_global_request_exit(void)
void i915_request_module_exit(void)
{
kmem_cache_destroy(global.slab_execute_cbs);
kmem_cache_destroy(global.slab_requests);
kmem_cache_destroy(slab_execute_cbs);
kmem_cache_destroy(slab_requests);
}
static struct i915_global_request global = { {
.exit = i915_global_request_exit,
} };
int __init i915_global_request_init(void)
int __init i915_request_module_init(void)
{
global.slab_requests =
slab_requests =
kmem_cache_create("i915_request",
sizeof(struct i915_request),
__alignof__(struct i915_request),
@@ -2100,20 +2092,19 @@ int __init i915_global_request_init(void)
SLAB_RECLAIM_ACCOUNT |
SLAB_TYPESAFE_BY_RCU,
__i915_request_ctor);
if (!global.slab_requests)
if (!slab_requests)
return -ENOMEM;
global.slab_execute_cbs = KMEM_CACHE(execute_cb,
slab_execute_cbs = KMEM_CACHE(execute_cb,
SLAB_HWCACHE_ALIGN |
SLAB_RECLAIM_ACCOUNT |
SLAB_TYPESAFE_BY_RCU);
if (!global.slab_execute_cbs)
if (!slab_execute_cbs)
goto err_requests;
i915_global_register(&global.base);
return 0;
err_requests:
kmem_cache_destroy(global.slab_requests);
kmem_cache_destroy(slab_requests);
return -ENOMEM;
}