Implement svm without BO concept in xe driver
Christian König
christian.koenig at amd.com
Wed Aug 16 06:05:44 UTC 2023
Hi Oak,
yeah, I completely agree with you and Felix. The main problem here is
getting the memory pressure visible on both sides.
At the moment I have absolutely no idea how to handle that, maybe
something like the ttm_resource object shared between TTM and HMM?
Regards,
Christian.
Am 16.08.23 um 05:47 schrieb Zeng, Oak:
> Hi Felix,
>
> It is great to hear from you!
>
> When I implement the HMM-based SVM for intel devices, I found this interesting problem: HMM uses struct page based memory management scheme which is completely different against the BO/TTM style memory management philosophy. Writing SVM code upon the BO/TTM concept seems overkill and awkward. So I thought we better make the SVM code BO-less and TTM-less. But on the other hand, currently vram eviction and cgroup memory accounting are all hooked to the TTM layer, which means a TTM-less SVM driver won't be able to evict vram allocated through TTM/gpu_vram_mgr.
>
> Ideally HMM migration should use drm-buddy for vram allocation, but we need to solve this TTM/HMM mutual eviction problem as you pointed out (I am working with application engineers to figure out whether mutual eviction can truly benefit applications). Maybe we can implement a TTM-less vram management block which can be shared b/t the HMM-based driver and the BO-based driver:
> * allocate/free memory from drm-buddy, buddy-block based
> * memory eviction logics, allow driver to specify which allocation is evictable
> * memory accounting, cgroup logic
>
> Maybe such a block can be placed at drm layer (say, call it drm_vram_mgr for now), so it can be shared b/t amd and intel. So I involved amd folks. Today both amd and intel-xe driver implemented a TTM-based vram manager which doesn't serve above design goal. Once the drm_vram_mgr is implemented, both amd and intel's BO-based/TTM-based vram manager, and the HMM-based vram manager can call into this drm-vram-mgr.
>
> Thanks again,
> Oak
>
>> -----Original Message-----
>> From: Felix Kuehling <felix.kuehling at amd.com>
>> Sent: August 15, 2023 6:17 PM
>> To: Zeng, Oak <oak.zeng at intel.com>; Thomas Hellström
>> <thomas.hellstrom at linux.intel.com>; Brost, Matthew
>> <matthew.brost at intel.com>; Vishwanathapura, Niranjana
>> <niranjana.vishwanathapura at intel.com>; Welty, Brian <brian.welty at intel.com>;
>> Christian König <christian.koenig at amd.com>; Philip Yang
>> <Philip.Yang at amd.com>; intel-xe at lists.freedesktop.org; dri-
>> devel at lists.freedesktop.org
>> Subject: Re: Implement svm without BO concept in xe driver
>>
>> Hi Oak,
>>
>> I'm not sure what you're looking for from AMD? Are we just CC'ed FYI? Or
>> are you looking for comments about
>>
>> * Our plans for VRAM management with HMM
>> * Our experience with BO-based VRAM management
>> * Something else?
>>
>> IMO, having separate memory pools for HMM and TTM is a non-starter for
>> AMD. We need access to the full VRAM in either of the APIs for it to be
>> useful. That also means, we need to handle memory pressure in both
>> directions. That's one of the main reasons we went with the BO-based
>> approach initially. I think in the long run, using the buddy allocator,
>> or the amdgpu_vram_mgr directly for HMM migrations would be better,
>> assuming we can handle memory pressure in both directions between HMM
>> and TTM sharing the same pool of physical memory.
>>
>> Regards,
>> Felix
>>
>>
>> On 2023-08-15 16:34, Zeng, Oak wrote:
>>> Also + Christian
>>>
>>> Thanks,
>>>
>>> Oak
>>>
>>> *From:*Intel-xe <intel-xe-bounces at lists.freedesktop.org> *On Behalf Of
>>> *Zeng, Oak
>>> *Sent:* August 14, 2023 11:38 PM
>>> *To:* Thomas Hellström <thomas.hellstrom at linux.intel.com>; Brost,
>>> Matthew <matthew.brost at intel.com>; Vishwanathapura, Niranjana
>>> <niranjana.vishwanathapura at intel.com>; Welty, Brian
>>> <brian.welty at intel.com>; Felix Kuehling <felix.kuehling at amd.com>;
>>> Philip Yang <Philip.Yang at amd.com>; intel-xe at lists.freedesktop.org;
>>> dri-devel at lists.freedesktop.org
>>> *Subject:* [Intel-xe] Implement svm without BO concept in xe driver
>>>
>>> Hi Thomas, Matt and all,
>>>
>>> This came up when I port i915 svm codes to xe driver. In i915
>>> implementation, we have i915_buddy manage gpu vram and svm codes
>>> directly call i915_buddy layer to allocate/free vram. There is no
>>> gem_bo/ttm bo concept involved in the svm implementation.
>>>
>>> In xe driver, we have drm_buddy, xe_ttm_vram_mgr and ttm layer to
>>> manage vram. Drm_buddy is initialized during xe_ttm_vram_mgr
>>> initialization. Vram allocation/free is done through xe_ttm_vram_mgr
>>> functions which call into drm_buddy layer to allocate vram blocks.
>>>
>>> I plan to implement xe svm driver the same way as we did in i915,
>>> which means there will not be bo concept in the svm implementation.
>>> Drm_buddy will be passed to svm layer during vram initialization and
>>> svm will allocate/free memory directly from drm_buddy, bypassing
>>> ttm/xee vram manager. Here are a few considerations/things we are
>>> aware of:
>>>
>>> 1. This approach seems match hmm design better than bo concept. Our
>>> svm implementation will be based on hmm. In hmm design, each vram
>>> page is backed by a struct page. It is very easy to perform page
>>> granularity migrations (b/t vram and system memory). If BO concept
>>> is involved, we will have to split/remerge BOs during page
>>> granularity migrations.
>>>
>>> 2. We have a prove of concept of this approach in i915, originally
>>> implemented by Niranjana. It seems work but it only has basic
>>> functionalities for now. We don’t have advanced features such as
>>> memory eviction etc.
>>>
>>> 3. With this approach, vram will divided into two separate pools: one
>>> for xe_gem_created BOs and one for vram used by svm. Those two
>>> pools are not connected: memory pressure from one pool won’t be
>>> able to evict vram from another pool. At this point, we don’t
>>> whether this aspect is good or not.
>>>
>>> 4. Amdkfd svm went different approach which is BO based. The benefit
>>> of this approach is a lot of existing driver facilities (such as
>>> memory eviction/cgroup/accounting) can be reused
>>>
>>> Do you have any comment to this approach? Should I come back with a
>>> RFC of some POC codes?
>>>
>>> Thanks,
>>>
>>> Oak
>>>
More information about the dri-devel
mailing list