[RFC PATCH 3/5] mm/vma: add support for peer to peer to device vma
Jason Gunthorpe
jgg at mellanox.com
Wed Jan 30 22:51:55 UTC 2019
On Wed, Jan 30, 2019 at 05:47:05PM -0500, Jerome Glisse wrote:
> On Wed, Jan 30, 2019 at 10:33:04PM +0000, Jason Gunthorpe wrote:
> > On Wed, Jan 30, 2019 at 05:30:27PM -0500, Jerome Glisse wrote:
> >
> > > > What is the problem in the HMM mirror that it needs this restriction?
> > >
> > > No restriction at all here. I think i just wasn't understood.
> >
> > Are you are talking about from the exporting side - where the thing
> > creating the VMA can really only put one distinct object into it?
>
> The message i was trying to get accross is that HMM mirror will
> always succeed for everything* except for special vma ie mmap of
> device file. For those it can only succeed if a p2p_map() call
> succeed.
>
> So any user of HMM mirror might to know why the mirroring fail ie
> was it because something exceptional is happening ? Or is it because
> i was trying to map a special vma which can be forbiden.
>
> Hence why i assume that you might want to know about such p2p_map
> failure at the time you create the umem odp object as it might be
> some failure you might want to report differently and handle
> differently. If you do not care about differentiating OOM or
> exceptional failure from p2p_map failure than you have nothing to
> worry about you will get the same error from HMM for both.
I think my hope here was that we could have some kind of 'trial'
interface where very eary users can call
'hmm_mirror_is_maybe_supported(dev, user_ptr, len)' and get a failure
indication.
We probably wouldn't call this on the full address space though
Beyond that it is just inevitable there can be problems faulting if
the memory map is messed with after MR is created.
And here again, I don't want to worry about any particular VMA
boundaries..
Jason
More information about the dri-devel
mailing list