<html>
<head>
<base href="https://bugs.freedesktop.org/" />
</head>
<body>
<p>
<div>
<b><a class="bz_bug_link
bz_status_RESOLVED bz_closed"
title="RESOLVED NOTABUG - Performance improvement : Please consider hardware ɢᴘᴜ rendering in llvmpipe"
href="https://bugs.freedesktop.org/show_bug.cgi?id=93686#c23">Comment # 23</a>
on <a class="bz_bug_link
bz_status_RESOLVED bz_closed"
title="RESOLVED NOTABUG - Performance improvement : Please consider hardware ɢᴘᴜ rendering in llvmpipe"
href="https://bugs.freedesktop.org/show_bug.cgi?id=93686">bug 93686</a>
from <span class="vcard"><a class="email" href="mailto:jan.vesely@rutgers.edu" title="Jan Vesely <jan.vesely@rutgers.edu>"> <span class="fn">Jan Vesely</span></a>
</span></b>
<pre>(In reply to ytrezq from <a href="show_bug.cgi?id=93686#c16">comment #16</a>)
<span class="quote">> (In reply to Alex Deucher from <a href="show_bug.cgi?id=93686#c13">comment #13</a>)
> > I can't speak for intel, but on AMD APUs, while the GPU appears as a device
> > on the PCIE bus, it actually has a much faster internal connection to the
> > memory controller.
>
> For simplifying let’s pick up the comparison of 2 virtual machines :
>
> each one is well parametrized so they none of them can fill the ram, even if
> they are alone (ex :max use 4Gb of ʀᴀᴍ and 64Gb available on the host).
>
> Does the first virtual machine can read the ʀᴀᴍ of the second ? No !
> Does the second virtual machine can read the ʀᴀᴍ of the first ? Neither !
> In both case they share the same memory hardware.
>
> The same apply to ɢᴘᴜs with ᴄᴘᴜs on ᴀᴘᴜs but with the hypervisor ᴏꜱ being
> replaced by a pure hardware one, so there’s no software parts (with the
> exception of the memory allocated memory ratio being controllable in the
> ʙɪᴏꜱ or the ᴜᴇꜰɪ firmware)</span >
the comparison with VMs is wrong, and the information about APUs is also wrong.
AMD APUs have complete access to the entire physical memory of the system. they
use both coherent and non-coherent links which are faster (higher bw + lower
latency) than PCIe lanes. see [0] if you want to learn more about APU memory
(it's a bit dated so the numbers are different for latest products).
coherent memory access allows you to avoid synchronization overhead (or pay it
on every access). the reason to mention APUs is that the difference between
"GPU memory" and coherent system ram is much smaller than dGPUs. the reason to
mention HSA is because they implement this approach wrt. compute; agents with
different capabilities sharing coherent view of the memory.
[0] <a href="http://developer.amd.com/wordpress/media/2013/06/1004_final.pdf">http://developer.amd.com/wordpress/media/2013/06/1004_final.pdf</a></pre>
</div>
</p>
<hr>
<span>You are receiving this mail because:</span>
<ul>
<li>You are the QA Contact for the bug.</li>
<li>You are the assignee for the bug.</li>
</ul>
</body>
</html>