[Intel-gfx] [PATCH 1/3] drm/i915/gem: Avoid kmalloc under i915->mm_lock
Chris Wilson
chris at chris-wilson.co.uk
Fri Jun 19 19:36:05 UTC 2020
Quoting Matthew Auld (2020-06-19 20:29:15)
> On Fri, 19 Jun 2020 at 15:31, Chris Wilson <chris at chris-wilson.co.uk> wrote:
> >
> > Rearrange the allocation of the mm_struct registration to avoid
> > allocating underneath the i915->mm_lock, so that we avoid tainting the
> > lock (and in turn many other locks that may be held as i915->mm_lock is
> > taken, and those locks we may want on the free [shrinker] paths). In
> > doing so, we convert the lookup to be RCU protected by courtesy of
> > converting the free-worker to be an rcu_work.
> >
> > Signed-off-by: Chris Wilson <chris at chris-wilson.co.uk>
> > ---
> > drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 123 +++++++++-----------
> > drivers/gpu/drm/i915/i915_drv.h | 2 +-
> > 2 files changed, 59 insertions(+), 66 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > index 9c53eb883400..84766414a1f0 100644
> > --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c
> > @@ -21,7 +21,7 @@ struct i915_mm_struct {
> > struct i915_mmu_notifier *mn;
> > struct hlist_node node;
> > struct kref kref;
> > - struct work_struct work;
> > + struct rcu_work work;
> > };
> >
> > #if defined(CONFIG_MMU_NOTIFIER)
> > @@ -189,40 +189,31 @@ i915_gem_userptr_release__mmu_notifier(struct drm_i915_gem_object *obj)
> > static struct i915_mmu_notifier *
> > i915_mmu_notifier_find(struct i915_mm_struct *mm)
> > {
> > - struct i915_mmu_notifier *mn;
> > - int err = 0;
> > + struct i915_mmu_notifier *mn, *old;
> > + int err;
> >
> > - mn = mm->mn;
> > - if (mn)
> > + mn = READ_ONCE(mm->mn);
> > + if (likely(mn))
> > return mn;
> >
> > mn = i915_mmu_notifier_create(mm);
> > if (IS_ERR(mn))
> > - err = PTR_ERR(mn);
> > -
> > - mmap_write_lock(mm->mm);
> > - mutex_lock(&mm->i915->mm_lock);
> > - if (mm->mn == NULL && !err) {
> > - /* Protected by mmap_lock (write-lock) */
> > - err = __mmu_notifier_register(&mn->mn, mm->mm);
> > - if (!err) {
> > - /* Protected by mm_lock */
> > - mm->mn = fetch_and_zero(&mn);
> > - }
> > - } else if (mm->mn) {
> > - /*
> > - * Someone else raced and successfully installed the mmu
> > - * notifier, we can cancel our own errors.
> > - */
> > - err = 0;
> > + return mn;
> > +
> > + err = mmu_notifier_register(&mn->mn, mm->mm);
> > + if (err) {
> > + kfree(mn);
> > + return ERR_PTR(err);
> > }
> > - mutex_unlock(&mm->i915->mm_lock);
> > - mmap_write_unlock(mm->mm);
> >
> > - if (mn && !IS_ERR(mn))
> > + old = cmpxchg(&mm->mn, NULL, mn);
> > + if (old) {
> > + mmu_notifier_unregister(&mn->mn, mm->mm);
> > kfree(mn);
> > + mn = old;
> > + }
> >
> > - return err ? ERR_PTR(err) : mm->mn;
> > + return mn;
> > }
> >
> > static int
> > @@ -301,23 +292,26 @@ i915_mmu_notifier_free(struct i915_mmu_notifier *mn,
> > #endif
> >
> > static struct i915_mm_struct *
> > -__i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real)
> > +__i915_mm_struct_find(struct drm_i915_private *i915, struct mm_struct *real)
> > {
> > - struct i915_mm_struct *mm;
> > + struct i915_mm_struct *it, *mm = NULL;
> >
> > - /* Protected by dev_priv->mm_lock */
> > - hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real)
> > - if (mm->mm == real)
> > - return mm;
> > + rcu_read_lock();
> > + hash_for_each_possible(i915->mm_structs, it, node, (unsigned long)real)
> > + if (it->mm == real && kref_get_unless_zero(&it->kref)) {
> > + mm = it;
> > + break;
> > + }
> > + rcu_read_unlock();
> >
> > - return NULL;
> > + return mm;
> > }
> >
> > static int
> > i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> > {
> > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev);
> > - struct i915_mm_struct *mm;
> > + struct drm_i915_private *i915 = to_i915(obj->base.dev);
> > + struct i915_mm_struct *mm, *new;
> > int ret = 0;
> >
> > /* During release of the GEM object we hold the struct_mutex. This
> > @@ -330,39 +324,40 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj)
> > * struct_mutex, i.e. we need to schedule a worker to do the clean
> > * up.
> > */
> > - mutex_lock(&dev_priv->mm_lock);
> > - mm = __i915_mm_struct_find(dev_priv, current->mm);
> > - if (mm == NULL) {
> > - mm = kmalloc(sizeof(*mm), GFP_KERNEL);
> > - if (mm == NULL) {
> > - ret = -ENOMEM;
> > - goto out;
> > - }
> > + mm = __i915_mm_struct_find(i915, current->mm);
>
> Is this really safe without the mm_lock, assuming concurrent
> hash_add/has_del with hash_for_each?
Hmm, not quite. There's hash_add_rcu, hash_del_rcu,
hash_for_each_possible_rcu, to ensure the write ordering is safe for the
concurrent readers. Good catch.
-Chris
More information about the Intel-gfx
mailing list