[PATCH v2 0/8] Engine Busyness

Tvrtko Ursulin tvrtko.ursulin at linux.intel.com
Thu Dec 21 09:36:41 UTC 2023


On 20/12/2023 23:58, Umesh Nerlige Ramappa wrote:
> On Wed, Dec 20, 2023 at 09:00:34AM +0000, Tvrtko Ursulin wrote:
>>
>> On 20/12/2023 05:36, Umesh Nerlige Ramappa wrote:
>>> On Thu, Dec 14, 2023 at 08:06:46AM +0000, Tvrtko Ursulin wrote:
>>>>
>>>> On 14/12/2023 01:56, Umesh Nerlige Ramappa wrote:
>>>>> On Thu, Dec 07, 2023 at 02:45:47PM +0000, Tvrtko Ursulin wrote:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> On 07/12/2023 12:57, Riana Tauro wrote:
>>>>>>> GuC provides engine busyness ticks as a 64 bit counter which count
>>>>>>> as clock ticks. These counters are maintained in a
>>>>>>> shared memory buffer and internally updated on a continuous basis.
>>>>>>>
>>>>>>> GuC also provides a periodically total active ticks that GT has been
>>>>>>> active for. This counter is exposed to the user such that 
>>>>>>> busyness can
>>>>>>> be calculated as a percentage using
>>>>>>>
>>>>>>> busyness % = (engine active ticks/total active ticks) * 100.
>>>>>>
>>>>>> I think I've asked this before but don't remember it was clarified 
>>>>>> - what are the semantics of "active" with total active ticks? In 
>>>>>> other words considering activity timelines like:
>>>>>>
>>>>>> 1)
>>>>>>     0          1s
>>>>>> rcs0 |xxxxx-----|
>>>>>> bcs0 |-----xxxxx|
>>>>>>
>>>>>> 2)
>>>>>>     0          1s
>>>>>> rcs0 |xxxxx-----|
>>>>>> bcs0 |xxxxx-----|
>>>>>>
>>>>>> Assuming 1s sampling, would the above formula correctly say 50% 
>>>>>> for both engines in both cases?
>>>>>
>>>>> Yes. What is the significance of case 2? Are you saying rcs and bcs 
>>>>> are executing in parallel?
>>>>
>>>> In parallel yes. Complete overlap, no overlap, or any overlap of 
>>>> activity in between the two.
>>>
>>> GuC accumulates this on context switches, so the overlap does not 
>>> matter.
>>>
>>>>
>>>>> Either ways, when total active ticks is queried it would provide 
>>>>> the latest value of the active time (does not depend on gt 
>>>>> park/unpark since the value is either obtained on demand from GuC 
>>>>> or is a value that is frequently updated by GuC.
>>>>>
>>>>> The duration of context (in to out) is accumulated for the each 
>>>>> engine.
>>>>
>>>> But why is the total *active* tick moving during the 0.5s - 1s time 
>>>> of the 2nd diagram though? What does it mean by "active" if nothing 
>>>> was active during that period?
>>>
>>> VF was still using it's allotted time and hence was active.
>>
>> And if we leave SR-IOV out for a moment?
> 
> Then it is just a periodically sampled (by GuC) value of GT ticks. The 
> period being 100ms.
> 
>>
>> "GuC also provides a periodically total active ticks that GT has been 
>> active for."
>>
>> How many time worth of total GT active ticks does GuC report in 
>> diagram 2 above?
> 
> Every 100ms we would see an updated value. For the duration of 0.5s, it 
> would be 500ms. Sampled at 1s, it will be 1000ms. Until 0.5s it should 
> be 100% busyness but there is an error margin of 100ms. From then on, 
> the busyness % will decrease as time progresses. The error margin is 
> more pronounced for very short workloads, so IGTs were changed to use 2s 
> batch durations rather than 500ms. Haven't checked if IGTs have been 
> posted yet though.

Sorry somehow it is still not clear to me. :)

GuC updates the GT total active ticks _constantly_? With a 100ms 
sampling so like:

a)

  while (true)
    if (gt_active)
       gt_total_active += 100ms
    sleep(100ms)

Or b):

  while (true)
    gt_total_active += 100ms
    sleep(100ms)


?

If a) then diagram 2) above would show 50% rcs0, no? (When sampled at 
T=0 and T=1s and deltas calculated.)

If b) then "...total active ticks that GT has been active for." uses a 
different definition of "GT active" than I am assuming? Like no relation 
to whether any of the engines is used, just the fact GuC is loaded and 
running?

Regards,

Tvrtko

> 
> Regards,
> Umesh
> 
>>
>> Regards,
>>
>> Tvrtko
>>
>>>
>>> Regards,
>>> Umesh
>>>
>>>>
>>>>>> I am also curious if there are plans to add support to 
>>>>>> intel_gpu_top in which case please copy me on the required 
>>>>>> refactorings.
>>>>>>
>>>>>
>>>>> Certainly. It's in the works.
>>>>
>>>> Cool.
>>>>
>>>> Regards,
>>>>
>>>> Tvrtko


More information about the Intel-xe mailing list