Implement svm without BO concept in xe driver

Felix Kuehling felix.kuehling at amd.com
Wed Aug 16 21:55:23 UTC 2023


On 2023-08-16 13:30, Zeng, Oak wrote:
> I spoke with Thomas. We discussed two approaches:
>
> 1) make ttm_resource a central place for vram management functions such as eviction, cgroup memory accounting. Both the BO-based driver and BO-less SVM codes call into ttm_resource_alloc/free functions for vram allocation/free.
>      *This way BO driver and SVM driver shares the eviction/cgroup logic, no need to reimplment LRU eviction list in SVM driver. Cgroup logic should be in ttm_resource layer. +Maarten.
>      *ttm_resource is not a perfect match for SVM to allocate vram. It is still a big overhead. The *bo* member of ttm_resource is not needed for SVM - this might end up with invasive changes to ttm...need to look into more details

Overhead is a problem. We'd want to be able to allocate, free and evict 
memory at a similar granularity as our preferred migration and page 
fault granularity, which defaults to 2MB in our SVM implementation.


> 	
> 2) svm code allocate memory directly from drm-buddy allocator, and expose memory eviction functions from both ttm and svm so they can evict memory from each other. For example, expose the ttm_mem_evict_first function from ttm side so hmm/svm code can call it; expose a similar function from svm side so ttm can evict hmm memory.

I like this option. One thing that needs some thought with this is how 
to get some semblance of fairness between the two types of clients. 
Basically how to choose what to evict. And what share of the available 
memory does each side get to use on average. E.g. an idle client may get 
all its memory evicted while a busy client may get a bigger share of the 
available memory.

Regards,
   Felix


>
>
> Today we don't know which approach is better. I will work on some prove of concept codes, starting with #1 approach firstly.
>
> Btw, I talked with application engineers and they said most applications actually use a mixture of gem_bo create and malloc, so we definitely need to solve this problem.
>
> Cheers,
> Oak
>
>> -----Original Message-----
>> From: Christian König <christian.koenig at amd.com>
>> Sent: August 16, 2023 2:06 AM
>> To: Zeng, Oak <oak.zeng at intel.com>; Felix Kuehling <felix.kuehling at amd.com>;
>> Thomas Hellström <thomas.hellstrom at linux.intel.com>; Brost, Matthew
>> <matthew.brost at intel.com>; Vishwanathapura, Niranjana
>> <niranjana.vishwanathapura at intel.com>; Welty, Brian <brian.welty at intel.com>;
>> Philip Yang <Philip.Yang at amd.com>; intel-xe at lists.freedesktop.org; dri-
>> devel at lists.freedesktop.org
>> Subject: Re: Implement svm without BO concept in xe driver
>>
>> Hi Oak,
>>
>> yeah, I completely agree with you and Felix. The main problem here is
>> getting the memory pressure visible on both sides.
>>
>> At the moment I have absolutely no idea how to handle that, maybe
>> something like the ttm_resource object shared between TTM and HMM?
>>
>> Regards,
>> Christian.
>>
>> Am 16.08.23 um 05:47 schrieb Zeng, Oak:
>>> Hi Felix,
>>>
>>> It is great to hear from you!
>>>
>>> When I implement the HMM-based SVM for intel devices, I found this
>> interesting problem: HMM uses struct page based memory management scheme
>> which is completely different against the BO/TTM style memory management
>> philosophy. Writing SVM code upon the BO/TTM concept seems overkill and
>> awkward. So I thought we better make the SVM code BO-less and TTM-less. But
>> on the other hand, currently vram eviction and cgroup memory accounting are all
>> hooked to the TTM layer, which means a TTM-less SVM driver won't be able to
>> evict vram allocated through TTM/gpu_vram_mgr.
>>> Ideally HMM migration should use drm-buddy for vram allocation, but we need
>> to solve this TTM/HMM mutual eviction problem as you pointed out (I am
>> working with application engineers to figure out whether mutual eviction can
>> truly benefit applications). Maybe we can implement a TTM-less vram
>> management block which can be shared b/t the HMM-based driver and the BO-
>> based driver:
>>>      * allocate/free memory from drm-buddy, buddy-block based
>>>      * memory eviction logics, allow driver to specify which allocation is evictable
>>>      * memory accounting, cgroup logic
>>>
>>> Maybe such a block can be placed at drm layer (say, call it drm_vram_mgr for
>> now), so it can be shared b/t amd and intel. So I involved amd folks. Today both
>> amd and intel-xe driver implemented a TTM-based vram manager which doesn't
>> serve above design goal. Once the drm_vram_mgr is implemented, both amd
>> and intel's BO-based/TTM-based vram manager, and the HMM-based vram
>> manager can call into this drm-vram-mgr.
>>> Thanks again,
>>> Oak
>>>
>>>> -----Original Message-----
>>>> From: Felix Kuehling <felix.kuehling at amd.com>
>>>> Sent: August 15, 2023 6:17 PM
>>>> To: Zeng, Oak <oak.zeng at intel.com>; Thomas Hellström
>>>> <thomas.hellstrom at linux.intel.com>; Brost, Matthew
>>>> <matthew.brost at intel.com>; Vishwanathapura, Niranjana
>>>> <niranjana.vishwanathapura at intel.com>; Welty, Brian
>> <brian.welty at intel.com>;
>>>> Christian König <christian.koenig at amd.com>; Philip Yang
>>>> <Philip.Yang at amd.com>; intel-xe at lists.freedesktop.org; dri-
>>>> devel at lists.freedesktop.org
>>>> Subject: Re: Implement svm without BO concept in xe driver
>>>>
>>>> Hi Oak,
>>>>
>>>> I'm not sure what you're looking for from AMD? Are we just CC'ed FYI? Or
>>>> are you looking for comments about
>>>>
>>>>     * Our plans for VRAM management with HMM
>>>>     * Our experience with BO-based VRAM management
>>>>     * Something else?
>>>>
>>>> IMO, having separate memory pools for HMM and TTM is a non-starter for
>>>> AMD. We need access to the full VRAM in either of the APIs for it to be
>>>> useful. That also means, we need to handle memory pressure in both
>>>> directions. That's one of the main reasons we went with the BO-based
>>>> approach initially. I think in the long run, using the buddy allocator,
>>>> or the amdgpu_vram_mgr directly for HMM migrations would be better,
>>>> assuming we can handle memory pressure in both directions between HMM
>>>> and TTM sharing the same pool of physical memory.
>>>>
>>>> Regards,
>>>>      Felix
>>>>
>>>>
>>>> On 2023-08-15 16:34, Zeng, Oak wrote:
>>>>> Also + Christian
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Oak
>>>>>
>>>>> *From:*Intel-xe <intel-xe-bounces at lists.freedesktop.org> *On Behalf Of
>>>>> *Zeng, Oak
>>>>> *Sent:* August 14, 2023 11:38 PM
>>>>> *To:* Thomas Hellström <thomas.hellstrom at linux.intel.com>; Brost,
>>>>> Matthew <matthew.brost at intel.com>; Vishwanathapura, Niranjana
>>>>> <niranjana.vishwanathapura at intel.com>; Welty, Brian
>>>>> <brian.welty at intel.com>; Felix Kuehling <felix.kuehling at amd.com>;
>>>>> Philip Yang <Philip.Yang at amd.com>; intel-xe at lists.freedesktop.org;
>>>>> dri-devel at lists.freedesktop.org
>>>>> *Subject:* [Intel-xe] Implement svm without BO concept in xe driver
>>>>>
>>>>> Hi Thomas, Matt and all,
>>>>>
>>>>> This came up when I port i915 svm codes to xe driver. In i915
>>>>> implementation, we have i915_buddy manage gpu vram and svm codes
>>>>> directly call i915_buddy layer to allocate/free vram. There is no
>>>>> gem_bo/ttm bo concept involved in the svm implementation.
>>>>>
>>>>> In xe driver,  we have drm_buddy, xe_ttm_vram_mgr and ttm layer to
>>>>> manage vram. Drm_buddy is initialized during xe_ttm_vram_mgr
>>>>> initialization. Vram allocation/free is done through xe_ttm_vram_mgr
>>>>> functions which call into drm_buddy layer to allocate vram blocks.
>>>>>
>>>>> I plan to implement xe svm driver the same way as we did in i915,
>>>>> which means there will not be bo concept in the svm implementation.
>>>>> Drm_buddy will be passed to svm layer during vram initialization and
>>>>> svm will allocate/free memory directly from drm_buddy, bypassing
>>>>> ttm/xee vram manager. Here are a few considerations/things we are
>>>>> aware of:
>>>>>
>>>>>    1. This approach seems match hmm design better than bo concept. Our
>>>>>       svm implementation will be based on hmm. In hmm design, each vram
>>>>>       page is backed by a struct page. It is very easy to perform page
>>>>>       granularity migrations (b/t vram and system memory). If BO concept
>>>>>       is involved, we will have to split/remerge BOs during page
>>>>>       granularity migrations.
>>>>>
>>>>>    2. We have a prove of concept of this approach in i915, originally
>>>>>       implemented by Niranjana. It seems work but it only has basic
>>>>>       functionalities for now. We don’t have advanced features such as
>>>>>       memory eviction etc.
>>>>>
>>>>>    3. With this approach, vram will divided into two separate pools: one
>>>>>       for xe_gem_created BOs and one for vram used by svm. Those two
>>>>>       pools are not connected: memory pressure from one pool won’t be
>>>>>       able to evict vram from another pool. At this point, we don’t
>>>>>       whether this aspect is good or not.
>>>>>
>>>>>    4. Amdkfd svm went different approach which is BO based. The benefit
>>>>>       of this approach is a lot of existing driver facilities (such as
>>>>>       memory eviction/cgroup/accounting) can be reused
>>>>>
>>>>> Do you have any comment to this approach? Should I come back with a
>>>>> RFC of some POC codes?
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Oak
>>>>>


More information about the dri-devel mailing list