mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

In situations where the GPU is mostly idle, all or nearly all buffer objects will be in the inactive list. But if the system is under memory pressure (from something other than GPU), we could still get a lot of shrinker calls. Which results in traversing a list of thousands of objs and in the end finding nothing to shrink. Which isn't so efficient. Instead split the inactive_list into two lists, one inactive objs which are shrinkable, and a second one for those that are not. This way we can avoid traversing objs which we know are not shrinker candidates. v2: Fix inverted logic think-o Signed-off-by: Rob Clark <robdclark@chromium.org>
157 lines
3.6 KiB
C
157 lines
3.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2016 Red Hat
|
|
* Author: Rob Clark <robdclark@gmail.com>
|
|
*/
|
|
|
|
#include "msm_drv.h"
|
|
#include "msm_gem.h"
|
|
#include "msm_gpu.h"
|
|
#include "msm_gpu_trace.h"
|
|
|
|
static unsigned long
|
|
msm_gem_shrinker_count(struct shrinker *shrinker, struct shrink_control *sc)
|
|
{
|
|
struct msm_drm_private *priv =
|
|
container_of(shrinker, struct msm_drm_private, shrinker);
|
|
struct msm_gem_object *msm_obj;
|
|
unsigned long count = 0;
|
|
|
|
mutex_lock(&priv->mm_lock);
|
|
|
|
list_for_each_entry(msm_obj, &priv->inactive_dontneed, mm_list) {
|
|
if (!msm_gem_trylock(&msm_obj->base))
|
|
continue;
|
|
if (is_purgeable(msm_obj))
|
|
count += msm_obj->base.size >> PAGE_SHIFT;
|
|
msm_gem_unlock(&msm_obj->base);
|
|
}
|
|
|
|
mutex_unlock(&priv->mm_lock);
|
|
|
|
return count;
|
|
}
|
|
|
|
static unsigned long
|
|
msm_gem_shrinker_scan(struct shrinker *shrinker, struct shrink_control *sc)
|
|
{
|
|
struct msm_drm_private *priv =
|
|
container_of(shrinker, struct msm_drm_private, shrinker);
|
|
struct msm_gem_object *msm_obj;
|
|
unsigned long freed = 0;
|
|
|
|
mutex_lock(&priv->mm_lock);
|
|
|
|
list_for_each_entry(msm_obj, &priv->inactive_dontneed, mm_list) {
|
|
if (freed >= sc->nr_to_scan)
|
|
break;
|
|
if (!msm_gem_trylock(&msm_obj->base))
|
|
continue;
|
|
if (is_purgeable(msm_obj)) {
|
|
msm_gem_purge(&msm_obj->base);
|
|
freed += msm_obj->base.size >> PAGE_SHIFT;
|
|
}
|
|
msm_gem_unlock(&msm_obj->base);
|
|
}
|
|
|
|
mutex_unlock(&priv->mm_lock);
|
|
|
|
if (freed > 0)
|
|
trace_msm_gem_purge(freed << PAGE_SHIFT);
|
|
|
|
return freed;
|
|
}
|
|
|
|
/* since we don't know any better, lets bail after a few
|
|
* and if necessary the shrinker will be invoked again.
|
|
* Seems better than unmapping *everything*
|
|
*/
|
|
static const int vmap_shrink_limit = 15;
|
|
|
|
static unsigned
|
|
vmap_shrink(struct list_head *mm_list)
|
|
{
|
|
struct msm_gem_object *msm_obj;
|
|
unsigned unmapped = 0;
|
|
|
|
list_for_each_entry(msm_obj, mm_list, mm_list) {
|
|
if (!msm_gem_trylock(&msm_obj->base))
|
|
continue;
|
|
if (is_vunmapable(msm_obj)) {
|
|
msm_gem_vunmap(&msm_obj->base);
|
|
unmapped++;
|
|
}
|
|
msm_gem_unlock(&msm_obj->base);
|
|
|
|
if (++unmapped >= vmap_shrink_limit)
|
|
break;
|
|
}
|
|
|
|
return unmapped;
|
|
}
|
|
|
|
static int
|
|
msm_gem_shrinker_vmap(struct notifier_block *nb, unsigned long event, void *ptr)
|
|
{
|
|
struct msm_drm_private *priv =
|
|
container_of(nb, struct msm_drm_private, vmap_notifier);
|
|
struct list_head *mm_lists[] = {
|
|
&priv->inactive_dontneed,
|
|
&priv->inactive_willneed,
|
|
priv->gpu ? &priv->gpu->active_list : NULL,
|
|
NULL,
|
|
};
|
|
unsigned idx, unmapped = 0;
|
|
|
|
mutex_lock(&priv->mm_lock);
|
|
|
|
for (idx = 0; mm_lists[idx]; idx++) {
|
|
unmapped += vmap_shrink(mm_lists[idx]);
|
|
|
|
if (unmapped >= vmap_shrink_limit)
|
|
break;
|
|
}
|
|
|
|
mutex_unlock(&priv->mm_lock);
|
|
|
|
*(unsigned long *)ptr += unmapped;
|
|
|
|
if (unmapped > 0)
|
|
trace_msm_gem_purge_vmaps(unmapped);
|
|
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
/**
|
|
* msm_gem_shrinker_init - Initialize msm shrinker
|
|
* @dev_priv: msm device
|
|
*
|
|
* This function registers and sets up the msm shrinker.
|
|
*/
|
|
void msm_gem_shrinker_init(struct drm_device *dev)
|
|
{
|
|
struct msm_drm_private *priv = dev->dev_private;
|
|
priv->shrinker.count_objects = msm_gem_shrinker_count;
|
|
priv->shrinker.scan_objects = msm_gem_shrinker_scan;
|
|
priv->shrinker.seeks = DEFAULT_SEEKS;
|
|
WARN_ON(register_shrinker(&priv->shrinker));
|
|
|
|
priv->vmap_notifier.notifier_call = msm_gem_shrinker_vmap;
|
|
WARN_ON(register_vmap_purge_notifier(&priv->vmap_notifier));
|
|
}
|
|
|
|
/**
|
|
* msm_gem_shrinker_cleanup - Clean up msm shrinker
|
|
* @dev_priv: msm device
|
|
*
|
|
* This function unregisters the msm shrinker.
|
|
*/
|
|
void msm_gem_shrinker_cleanup(struct drm_device *dev)
|
|
{
|
|
struct msm_drm_private *priv = dev->dev_private;
|
|
|
|
if (priv->shrinker.nr_deferred) {
|
|
WARN_ON(unregister_vmap_purge_notifier(&priv->vmap_notifier));
|
|
unregister_shrinker(&priv->shrinker);
|
|
}
|
|
}
|