[Intel-gfx] [PATCH] drm/i915: Use SSE4.1 movntdqa to accelerate reads from WC memory
Tvrtko Ursulin
tvrtko.ursulin at linux.intel.com
Mon Jul 18 12:56:52 UTC 2016
On 18/07/16 12:57, Dave Gordon wrote:
> On 18/07/16 12:35, Chris Wilson wrote:
>> On Mon, Jul 18, 2016 at 12:15:32PM +0100, Tvrtko Ursulin wrote:
>>> I am not sure about this, but looking at the raid6 for example, it
>>> has a lot more annotations in cases like this.
>>>
>>> It seems to be telling the compiler which memory ranges does each
>>> instruction access, and also uses "asm volatile" - whether or not
>>> that is really needed I don't know.
>>>
>>> For example:
>>> asm volatile("movdqa %0,%%xmm4" :: "m" (dptr[z0][d]));
>>>
>>> And:
>>> asm volatile("movdqa %%xmm4,%0" : "=m" (q[d]));
>>>
>>> Each one is telling the compiler the instruction is either reading
>>> or writing respectively from a certain memory address.
>>>
>>> You don't have any of that, and don't even specify nothing as an
>>> output parameter so I am not sure if your code is safe.
>>
>> The asm is correct. We do not modify either of the two pointers which we
>> pass in via register inputs, but the memory behind them - hence the
>> memory
>> clobber.
>
> This is a choice of how much we let the compiler decide about
> addressing, and how much we tell it about what the asm code really does.
> The examples above get the compiler to generate *any* suitable
> addressing mode for each specific location involved in the transfers, so
> the compiler knows a lot about what's happening and can track where each
> datum comes from and goes to.
>
> OTOH Chris' code
>
> + asm("movntdqa (%0), %%xmm0\n"
> + "movntdqa 16(%0), %%xmm1\n"
> + "movntdqa 32(%0), %%xmm2\n"
> + "movntdqa 48(%0), %%xmm3\n"
> + "movaps %%xmm0, (%1)\n"
> + "movaps %%xmm1, 16(%1)\n"
> + "movaps %%xmm2, 32(%1)\n"
> + "movaps %%xmm3, 48(%1)\n"
> + :: "r" (src), "r" (dst) : "memory");
>
> - doesn't need "volatile" because asm statements that have no output
> operands are implicitly volatile.
>
> - makes the compiler give us the source and destination *addresses* in a
> register each; beyond that, it doesn't know what we're doing with them,
> so the third ("clobbers") parameter has to say "memory" i.e. treat *all*
> memory contents as unknown after this.
>
> [[From GCC docs: The "memory" clobber tells the compiler that the
> assembly code performs memory reads or writes to items other than those
> listed in the input and output operands (for example, accessing the
> memory pointed to by one of the input parameters). To ensure memory
> contains correct values, GCC may need to flush specific register values
> to memory before executing the asm. Further, the compiler does not
> assume that any values read from memory before an asm remain unchanged
> after that asm; it reloads them as needed. Using the "memory" clobber
> effectively forms a read/write memory barrier for the compiler.]]
>
> BTW, should we not tell it we've *also* clobbered %xmm[0-3]?
>
> So they're both correct, just taking different approaches. I don't know
> which would give the best performance for this specific case.
Cool, learn something new every day. :)
I've tried writing it as:
struct qw2 {
u64 q[2];
} __attribute__((packed));
static void __memcpy_ntdqa(struct qw2 *dst, const struct qw2 *src, unsigned long len)
{
kernel_fpu_begin();
len >>= 4;
while (len >= 4) {
asm("movntdqa (%0), %%xmm0" :: "r" (src), "m" (src[0]));
asm("movntdqa 16(%0), %%xmm1" :: "r" (src), "m" (src[1]));
asm("movntdqa 32(%0), %%xmm2" :: "r" (src), "m" (src[2]));
asm("movntdqa 48(%0), %%xmm3" :: "r" (src), "m" (src[3]));
asm("movaps %%xmm0, (%1)" : "=m" (dst[0]) : "r" (dst));
asm("movaps %%xmm1, 16(%1)" : "=m" (dst[1]) : "r" (dst));
asm("movaps %%xmm2, 32(%1)" : "=m" (dst[2]) : "r" (dst));
asm("movaps %%xmm3, 48(%1)" : "=m" (dst[3]) : "r" (dst));
src += 4;
dst += 4;
len -= 4;
}
while (len--) {
asm("movntdqa (%0), %%xmm0" :: "r" (src), "m" (src[0]));
asm("movaps %%xmm0, (%1)" : "=m" (dst[0]) : "r" (dst));
src++;
dst++;
}
kernel_fpu_end();
}
That appears to allow GCC to interleave SSE and normal instructions,
presumably that means it is trying to utilize the execution units better?
I wonder if it makes a difference in speed?
Old code main loop assembly looks like:
58: 66 0f 38 2a 00 movntdqa (%rax),%xmm0
5d: 66 0f 38 2a 48 10 movntdqa 0x10(%rax),%xmm1
63: 66 0f 38 2a 50 20 movntdqa 0x20(%rax),%xmm2
69: 66 0f 38 2a 58 30 movntdqa 0x30(%rax),%xmm3
6f: 0f 29 01 movaps %xmm0,(%rcx)
72: 0f 29 49 10 movaps %xmm1,0x10(%rcx)
76: 0f 29 51 20 movaps %xmm2,0x20(%rcx)
7a: 0f 29 59 30 movaps %xmm3,0x30(%rcx)
7e: 49 83 e8 04 sub $0x4,%r8
82: 48 83 c0 40 add $0x40,%rax
86: 48 83 c1 40 add $0x40,%rcx
8a: 49 83 f8 03 cmp $0x3,%r8
8e: 77 c8 ja 58 <i915_memcpy_from_wc+0x58>
While the above version generates:
58: 66 0f 38 2a 00 movntdqa (%rax),%xmm0
5d: 66 0f 38 2a 48 10 movntdqa 0x10(%rax),%xmm1
63: 66 0f 38 2a 50 20 movntdqa 0x20(%rax),%xmm2
69: 66 0f 38 2a 58 30 movntdqa 0x30(%rax),%xmm3
6f: 49 83 e8 04 sub $0x4,%r8
73: 48 83 c0 40 add $0x40,%rax
77: 0f 29 01 movaps %xmm0,(%rcx)
7a: 0f 29 49 10 movaps %xmm1,0x10(%rcx)
7e: 0f 29 51 20 movaps %xmm2,0x20(%rcx)
82: 0f 29 59 30 movaps %xmm3,0x30(%rcx)
86: 48 83 c1 40 add $0x40,%rcx
8a: 49 83 f8 03 cmp $0x3,%r8
8e: 77 c8 ja 58 <i915_memcpy_from_wc+0x58>
Interestingly, in both cases GCC does some in my mind futile
shuffling aroung between the two loops. Instead of just
carrying on with src and dst and len how they are, it goes
to use a different register set for the second loop:
So this reshuffling:
90: 48 8d 42 fc lea -0x4(%rdx),%rax
94: 83 e2 03 and $0x3,%edx
97: 48 c1 e8 02 shr $0x2,%rax
9b: 48 83 c0 01 add $0x1,%rax
9f: 48 c1 e0 06 shl $0x6,%rax
a3: 48 01 c6 add %rax,%rsi
a6: 48 01 c7 add %rax,%rdi
a9: 48 8d 42 ff lea -0x1(%rdx),%rax
ad: 48 85 d2 test %rdx,%rdx
b0: 74 1a je cc <i915_memcpy_from_wc+0xcc>
And then the second loop:
b2: 66 0f 38 2a 06 movntdqa (%rsi),%xmm0
b7: 48 83 e8 01 sub $0x1,%rax
bb: 48 83 c6 10 add $0x10,%rsi
bf: 0f 29 07 movaps %xmm0,(%rdi)
c2: 48 83 c7 10 add $0x10,%rdi
c6: 48 83 f8 ff cmp $0xffffffffffffffff,%rax
ca: 75 e6 jne b2 <i915_memcpy_from_wc+0xb2>
Any thoughts on this?
Regards,
Tvrtko
More information about the Intel-gfx
mailing list