Blame view

kernel/linux-rt-4.4.41/arch/x86/crypto/chacha20-avx2-x86_64.S 12.5 KB
5113f6f70   김현기   kernel add
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
  /*
   * ChaCha20 256-bit cipher algorithm, RFC7539, x64 AVX2 functions
   *
   * Copyright (C) 2015 Martin Willi
   *
   * This program is free software; you can redistribute it and/or modify
   * it under the terms of the GNU General Public License as published by
   * the Free Software Foundation; either version 2 of the License, or
   * (at your option) any later version.
   */
  
  #include <linux/linkage.h>
  
  .data
  .align 32
  
  ROT8:	.octa 0x0e0d0c0f0a09080b0605040702010003
  	.octa 0x0e0d0c0f0a09080b0605040702010003
  ROT16:	.octa 0x0d0c0f0e09080b0a0504070601000302
  	.octa 0x0d0c0f0e09080b0a0504070601000302
  CTRINC:	.octa 0x00000003000000020000000100000000
  	.octa 0x00000007000000060000000500000004
  
  .text
  
  ENTRY(chacha20_8block_xor_avx2)
  	# %rdi: Input state matrix, s
  	# %rsi: 8 data blocks output, o
  	# %rdx: 8 data blocks input, i
  
  	# This function encrypts eight consecutive ChaCha20 blocks by loading
  	# the state matrix in AVX registers eight times. As we need some
  	# scratch registers, we save the first four registers on the stack. The
  	# algorithm performs each operation on the corresponding word of each
  	# state matrix, hence requires no word shuffling. For final XORing step
  	# we transpose the matrix by interleaving 32-, 64- and then 128-bit
  	# words, which allows us to do XOR in AVX registers. 8/16-bit word
  	# rotation is done with the slightly better performing byte shuffling,
  	# 7/12-bit word rotation uses traditional shift+OR.
  
  	vzeroupper
  	# 4 * 32 byte stack, 32-byte aligned
  	mov		%rsp, %r8
  	and		$~31, %rsp
  	sub		$0x80, %rsp
  
  	# x0..15[0-7] = s[0..15]
  	vpbroadcastd	0x00(%rdi),%ymm0
  	vpbroadcastd	0x04(%rdi),%ymm1
  	vpbroadcastd	0x08(%rdi),%ymm2
  	vpbroadcastd	0x0c(%rdi),%ymm3
  	vpbroadcastd	0x10(%rdi),%ymm4
  	vpbroadcastd	0x14(%rdi),%ymm5
  	vpbroadcastd	0x18(%rdi),%ymm6
  	vpbroadcastd	0x1c(%rdi),%ymm7
  	vpbroadcastd	0x20(%rdi),%ymm8
  	vpbroadcastd	0x24(%rdi),%ymm9
  	vpbroadcastd	0x28(%rdi),%ymm10
  	vpbroadcastd	0x2c(%rdi),%ymm11
  	vpbroadcastd	0x30(%rdi),%ymm12
  	vpbroadcastd	0x34(%rdi),%ymm13
  	vpbroadcastd	0x38(%rdi),%ymm14
  	vpbroadcastd	0x3c(%rdi),%ymm15
  	# x0..3 on stack
  	vmovdqa		%ymm0,0x00(%rsp)
  	vmovdqa		%ymm1,0x20(%rsp)
  	vmovdqa		%ymm2,0x40(%rsp)
  	vmovdqa		%ymm3,0x60(%rsp)
  
  	vmovdqa		CTRINC(%rip),%ymm1
  	vmovdqa		ROT8(%rip),%ymm2
  	vmovdqa		ROT16(%rip),%ymm3
  
  	# x12 += counter values 0-3
  	vpaddd		%ymm1,%ymm12,%ymm12
  
  	mov		$10,%ecx
  
  .Ldoubleround8:
  	# x0 += x4, x12 = rotl32(x12 ^ x0, 16)
  	vpaddd		0x00(%rsp),%ymm4,%ymm0
  	vmovdqa		%ymm0,0x00(%rsp)
  	vpxor		%ymm0,%ymm12,%ymm12
  	vpshufb		%ymm3,%ymm12,%ymm12
  	# x1 += x5, x13 = rotl32(x13 ^ x1, 16)
  	vpaddd		0x20(%rsp),%ymm5,%ymm0
  	vmovdqa		%ymm0,0x20(%rsp)
  	vpxor		%ymm0,%ymm13,%ymm13
  	vpshufb		%ymm3,%ymm13,%ymm13
  	# x2 += x6, x14 = rotl32(x14 ^ x2, 16)
  	vpaddd		0x40(%rsp),%ymm6,%ymm0
  	vmovdqa		%ymm0,0x40(%rsp)
  	vpxor		%ymm0,%ymm14,%ymm14
  	vpshufb		%ymm3,%ymm14,%ymm14
  	# x3 += x7, x15 = rotl32(x15 ^ x3, 16)
  	vpaddd		0x60(%rsp),%ymm7,%ymm0
  	vmovdqa		%ymm0,0x60(%rsp)
  	vpxor		%ymm0,%ymm15,%ymm15
  	vpshufb		%ymm3,%ymm15,%ymm15
  
  	# x8 += x12, x4 = rotl32(x4 ^ x8, 12)
  	vpaddd		%ymm12,%ymm8,%ymm8
  	vpxor		%ymm8,%ymm4,%ymm4
  	vpslld		$12,%ymm4,%ymm0
  	vpsrld		$20,%ymm4,%ymm4
  	vpor		%ymm0,%ymm4,%ymm4
  	# x9 += x13, x5 = rotl32(x5 ^ x9, 12)
  	vpaddd		%ymm13,%ymm9,%ymm9
  	vpxor		%ymm9,%ymm5,%ymm5
  	vpslld		$12,%ymm5,%ymm0
  	vpsrld		$20,%ymm5,%ymm5
  	vpor		%ymm0,%ymm5,%ymm5
  	# x10 += x14, x6 = rotl32(x6 ^ x10, 12)
  	vpaddd		%ymm14,%ymm10,%ymm10
  	vpxor		%ymm10,%ymm6,%ymm6
  	vpslld		$12,%ymm6,%ymm0
  	vpsrld		$20,%ymm6,%ymm6
  	vpor		%ymm0,%ymm6,%ymm6
  	# x11 += x15, x7 = rotl32(x7 ^ x11, 12)
  	vpaddd		%ymm15,%ymm11,%ymm11
  	vpxor		%ymm11,%ymm7,%ymm7
  	vpslld		$12,%ymm7,%ymm0
  	vpsrld		$20,%ymm7,%ymm7
  	vpor		%ymm0,%ymm7,%ymm7
  
  	# x0 += x4, x12 = rotl32(x12 ^ x0, 8)
  	vpaddd		0x00(%rsp),%ymm4,%ymm0
  	vmovdqa		%ymm0,0x00(%rsp)
  	vpxor		%ymm0,%ymm12,%ymm12
  	vpshufb		%ymm2,%ymm12,%ymm12
  	# x1 += x5, x13 = rotl32(x13 ^ x1, 8)
  	vpaddd		0x20(%rsp),%ymm5,%ymm0
  	vmovdqa		%ymm0,0x20(%rsp)
  	vpxor		%ymm0,%ymm13,%ymm13
  	vpshufb		%ymm2,%ymm13,%ymm13
  	# x2 += x6, x14 = rotl32(x14 ^ x2, 8)
  	vpaddd		0x40(%rsp),%ymm6,%ymm0
  	vmovdqa		%ymm0,0x40(%rsp)
  	vpxor		%ymm0,%ymm14,%ymm14
  	vpshufb		%ymm2,%ymm14,%ymm14
  	# x3 += x7, x15 = rotl32(x15 ^ x3, 8)
  	vpaddd		0x60(%rsp),%ymm7,%ymm0
  	vmovdqa		%ymm0,0x60(%rsp)
  	vpxor		%ymm0,%ymm15,%ymm15
  	vpshufb		%ymm2,%ymm15,%ymm15
  
  	# x8 += x12, x4 = rotl32(x4 ^ x8, 7)
  	vpaddd		%ymm12,%ymm8,%ymm8
  	vpxor		%ymm8,%ymm4,%ymm4
  	vpslld		$7,%ymm4,%ymm0
  	vpsrld		$25,%ymm4,%ymm4
  	vpor		%ymm0,%ymm4,%ymm4
  	# x9 += x13, x5 = rotl32(x5 ^ x9, 7)
  	vpaddd		%ymm13,%ymm9,%ymm9
  	vpxor		%ymm9,%ymm5,%ymm5
  	vpslld		$7,%ymm5,%ymm0
  	vpsrld		$25,%ymm5,%ymm5
  	vpor		%ymm0,%ymm5,%ymm5
  	# x10 += x14, x6 = rotl32(x6 ^ x10, 7)
  	vpaddd		%ymm14,%ymm10,%ymm10
  	vpxor		%ymm10,%ymm6,%ymm6
  	vpslld		$7,%ymm6,%ymm0
  	vpsrld		$25,%ymm6,%ymm6
  	vpor		%ymm0,%ymm6,%ymm6
  	# x11 += x15, x7 = rotl32(x7 ^ x11, 7)
  	vpaddd		%ymm15,%ymm11,%ymm11
  	vpxor		%ymm11,%ymm7,%ymm7
  	vpslld		$7,%ymm7,%ymm0
  	vpsrld		$25,%ymm7,%ymm7
  	vpor		%ymm0,%ymm7,%ymm7
  
  	# x0 += x5, x15 = rotl32(x15 ^ x0, 16)
  	vpaddd		0x00(%rsp),%ymm5,%ymm0
  	vmovdqa		%ymm0,0x00(%rsp)
  	vpxor		%ymm0,%ymm15,%ymm15
  	vpshufb		%ymm3,%ymm15,%ymm15
  	# x1 += x6, x12 = rotl32(x12 ^ x1, 16)%ymm0
  	vpaddd		0x20(%rsp),%ymm6,%ymm0
  	vmovdqa		%ymm0,0x20(%rsp)
  	vpxor		%ymm0,%ymm12,%ymm12
  	vpshufb		%ymm3,%ymm12,%ymm12
  	# x2 += x7, x13 = rotl32(x13 ^ x2, 16)
  	vpaddd		0x40(%rsp),%ymm7,%ymm0
  	vmovdqa		%ymm0,0x40(%rsp)
  	vpxor		%ymm0,%ymm13,%ymm13
  	vpshufb		%ymm3,%ymm13,%ymm13
  	# x3 += x4, x14 = rotl32(x14 ^ x3, 16)
  	vpaddd		0x60(%rsp),%ymm4,%ymm0
  	vmovdqa		%ymm0,0x60(%rsp)
  	vpxor		%ymm0,%ymm14,%ymm14
  	vpshufb		%ymm3,%ymm14,%ymm14
  
  	# x10 += x15, x5 = rotl32(x5 ^ x10, 12)
  	vpaddd		%ymm15,%ymm10,%ymm10
  	vpxor		%ymm10,%ymm5,%ymm5
  	vpslld		$12,%ymm5,%ymm0
  	vpsrld		$20,%ymm5,%ymm5
  	vpor		%ymm0,%ymm5,%ymm5
  	# x11 += x12, x6 = rotl32(x6 ^ x11, 12)
  	vpaddd		%ymm12,%ymm11,%ymm11
  	vpxor		%ymm11,%ymm6,%ymm6
  	vpslld		$12,%ymm6,%ymm0
  	vpsrld		$20,%ymm6,%ymm6
  	vpor		%ymm0,%ymm6,%ymm6
  	# x8 += x13, x7 = rotl32(x7 ^ x8, 12)
  	vpaddd		%ymm13,%ymm8,%ymm8
  	vpxor		%ymm8,%ymm7,%ymm7
  	vpslld		$12,%ymm7,%ymm0
  	vpsrld		$20,%ymm7,%ymm7
  	vpor		%ymm0,%ymm7,%ymm7
  	# x9 += x14, x4 = rotl32(x4 ^ x9, 12)
  	vpaddd		%ymm14,%ymm9,%ymm9
  	vpxor		%ymm9,%ymm4,%ymm4
  	vpslld		$12,%ymm4,%ymm0
  	vpsrld		$20,%ymm4,%ymm4
  	vpor		%ymm0,%ymm4,%ymm4
  
  	# x0 += x5, x15 = rotl32(x15 ^ x0, 8)
  	vpaddd		0x00(%rsp),%ymm5,%ymm0
  	vmovdqa		%ymm0,0x00(%rsp)
  	vpxor		%ymm0,%ymm15,%ymm15
  	vpshufb		%ymm2,%ymm15,%ymm15
  	# x1 += x6, x12 = rotl32(x12 ^ x1, 8)
  	vpaddd		0x20(%rsp),%ymm6,%ymm0
  	vmovdqa		%ymm0,0x20(%rsp)
  	vpxor		%ymm0,%ymm12,%ymm12
  	vpshufb		%ymm2,%ymm12,%ymm12
  	# x2 += x7, x13 = rotl32(x13 ^ x2, 8)
  	vpaddd		0x40(%rsp),%ymm7,%ymm0
  	vmovdqa		%ymm0,0x40(%rsp)
  	vpxor		%ymm0,%ymm13,%ymm13
  	vpshufb		%ymm2,%ymm13,%ymm13
  	# x3 += x4, x14 = rotl32(x14 ^ x3, 8)
  	vpaddd		0x60(%rsp),%ymm4,%ymm0
  	vmovdqa		%ymm0,0x60(%rsp)
  	vpxor		%ymm0,%ymm14,%ymm14
  	vpshufb		%ymm2,%ymm14,%ymm14
  
  	# x10 += x15, x5 = rotl32(x5 ^ x10, 7)
  	vpaddd		%ymm15,%ymm10,%ymm10
  	vpxor		%ymm10,%ymm5,%ymm5
  	vpslld		$7,%ymm5,%ymm0
  	vpsrld		$25,%ymm5,%ymm5
  	vpor		%ymm0,%ymm5,%ymm5
  	# x11 += x12, x6 = rotl32(x6 ^ x11, 7)
  	vpaddd		%ymm12,%ymm11,%ymm11
  	vpxor		%ymm11,%ymm6,%ymm6
  	vpslld		$7,%ymm6,%ymm0
  	vpsrld		$25,%ymm6,%ymm6
  	vpor		%ymm0,%ymm6,%ymm6
  	# x8 += x13, x7 = rotl32(x7 ^ x8, 7)
  	vpaddd		%ymm13,%ymm8,%ymm8
  	vpxor		%ymm8,%ymm7,%ymm7
  	vpslld		$7,%ymm7,%ymm0
  	vpsrld		$25,%ymm7,%ymm7
  	vpor		%ymm0,%ymm7,%ymm7
  	# x9 += x14, x4 = rotl32(x4 ^ x9, 7)
  	vpaddd		%ymm14,%ymm9,%ymm9
  	vpxor		%ymm9,%ymm4,%ymm4
  	vpslld		$7,%ymm4,%ymm0
  	vpsrld		$25,%ymm4,%ymm4
  	vpor		%ymm0,%ymm4,%ymm4
  
  	dec		%ecx
  	jnz		.Ldoubleround8
  
  	# x0..15[0-3] += s[0..15]
  	vpbroadcastd	0x00(%rdi),%ymm0
  	vpaddd		0x00(%rsp),%ymm0,%ymm0
  	vmovdqa		%ymm0,0x00(%rsp)
  	vpbroadcastd	0x04(%rdi),%ymm0
  	vpaddd		0x20(%rsp),%ymm0,%ymm0
  	vmovdqa		%ymm0,0x20(%rsp)
  	vpbroadcastd	0x08(%rdi),%ymm0
  	vpaddd		0x40(%rsp),%ymm0,%ymm0
  	vmovdqa		%ymm0,0x40(%rsp)
  	vpbroadcastd	0x0c(%rdi),%ymm0
  	vpaddd		0x60(%rsp),%ymm0,%ymm0
  	vmovdqa		%ymm0,0x60(%rsp)
  	vpbroadcastd	0x10(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm4,%ymm4
  	vpbroadcastd	0x14(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm5,%ymm5
  	vpbroadcastd	0x18(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm6,%ymm6
  	vpbroadcastd	0x1c(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm7,%ymm7
  	vpbroadcastd	0x20(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm8,%ymm8
  	vpbroadcastd	0x24(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm9,%ymm9
  	vpbroadcastd	0x28(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm10,%ymm10
  	vpbroadcastd	0x2c(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm11,%ymm11
  	vpbroadcastd	0x30(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm12,%ymm12
  	vpbroadcastd	0x34(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm13,%ymm13
  	vpbroadcastd	0x38(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm14,%ymm14
  	vpbroadcastd	0x3c(%rdi),%ymm0
  	vpaddd		%ymm0,%ymm15,%ymm15
  
  	# x12 += counter values 0-3
  	vpaddd		%ymm1,%ymm12,%ymm12
  
  	# interleave 32-bit words in state n, n+1
  	vmovdqa		0x00(%rsp),%ymm0
  	vmovdqa		0x20(%rsp),%ymm1
  	vpunpckldq	%ymm1,%ymm0,%ymm2
  	vpunpckhdq	%ymm1,%ymm0,%ymm1
  	vmovdqa		%ymm2,0x00(%rsp)
  	vmovdqa		%ymm1,0x20(%rsp)
  	vmovdqa		0x40(%rsp),%ymm0
  	vmovdqa		0x60(%rsp),%ymm1
  	vpunpckldq	%ymm1,%ymm0,%ymm2
  	vpunpckhdq	%ymm1,%ymm0,%ymm1
  	vmovdqa		%ymm2,0x40(%rsp)
  	vmovdqa		%ymm1,0x60(%rsp)
  	vmovdqa		%ymm4,%ymm0
  	vpunpckldq	%ymm5,%ymm0,%ymm4
  	vpunpckhdq	%ymm5,%ymm0,%ymm5
  	vmovdqa		%ymm6,%ymm0
  	vpunpckldq	%ymm7,%ymm0,%ymm6
  	vpunpckhdq	%ymm7,%ymm0,%ymm7
  	vmovdqa		%ymm8,%ymm0
  	vpunpckldq	%ymm9,%ymm0,%ymm8
  	vpunpckhdq	%ymm9,%ymm0,%ymm9
  	vmovdqa		%ymm10,%ymm0
  	vpunpckldq	%ymm11,%ymm0,%ymm10
  	vpunpckhdq	%ymm11,%ymm0,%ymm11
  	vmovdqa		%ymm12,%ymm0
  	vpunpckldq	%ymm13,%ymm0,%ymm12
  	vpunpckhdq	%ymm13,%ymm0,%ymm13
  	vmovdqa		%ymm14,%ymm0
  	vpunpckldq	%ymm15,%ymm0,%ymm14
  	vpunpckhdq	%ymm15,%ymm0,%ymm15
  
  	# interleave 64-bit words in state n, n+2
  	vmovdqa		0x00(%rsp),%ymm0
  	vmovdqa		0x40(%rsp),%ymm2
  	vpunpcklqdq	%ymm2,%ymm0,%ymm1
  	vpunpckhqdq	%ymm2,%ymm0,%ymm2
  	vmovdqa		%ymm1,0x00(%rsp)
  	vmovdqa		%ymm2,0x40(%rsp)
  	vmovdqa		0x20(%rsp),%ymm0
  	vmovdqa		0x60(%rsp),%ymm2
  	vpunpcklqdq	%ymm2,%ymm0,%ymm1
  	vpunpckhqdq	%ymm2,%ymm0,%ymm2
  	vmovdqa		%ymm1,0x20(%rsp)
  	vmovdqa		%ymm2,0x60(%rsp)
  	vmovdqa		%ymm4,%ymm0
  	vpunpcklqdq	%ymm6,%ymm0,%ymm4
  	vpunpckhqdq	%ymm6,%ymm0,%ymm6
  	vmovdqa		%ymm5,%ymm0
  	vpunpcklqdq	%ymm7,%ymm0,%ymm5
  	vpunpckhqdq	%ymm7,%ymm0,%ymm7
  	vmovdqa		%ymm8,%ymm0
  	vpunpcklqdq	%ymm10,%ymm0,%ymm8
  	vpunpckhqdq	%ymm10,%ymm0,%ymm10
  	vmovdqa		%ymm9,%ymm0
  	vpunpcklqdq	%ymm11,%ymm0,%ymm9
  	vpunpckhqdq	%ymm11,%ymm0,%ymm11
  	vmovdqa		%ymm12,%ymm0
  	vpunpcklqdq	%ymm14,%ymm0,%ymm12
  	vpunpckhqdq	%ymm14,%ymm0,%ymm14
  	vmovdqa		%ymm13,%ymm0
  	vpunpcklqdq	%ymm15,%ymm0,%ymm13
  	vpunpckhqdq	%ymm15,%ymm0,%ymm15
  
  	# interleave 128-bit words in state n, n+4
  	vmovdqa		0x00(%rsp),%ymm0
  	vperm2i128	$0x20,%ymm4,%ymm0,%ymm1
  	vperm2i128	$0x31,%ymm4,%ymm0,%ymm4
  	vmovdqa		%ymm1,0x00(%rsp)
  	vmovdqa		0x20(%rsp),%ymm0
  	vperm2i128	$0x20,%ymm5,%ymm0,%ymm1
  	vperm2i128	$0x31,%ymm5,%ymm0,%ymm5
  	vmovdqa		%ymm1,0x20(%rsp)
  	vmovdqa		0x40(%rsp),%ymm0
  	vperm2i128	$0x20,%ymm6,%ymm0,%ymm1
  	vperm2i128	$0x31,%ymm6,%ymm0,%ymm6
  	vmovdqa		%ymm1,0x40(%rsp)
  	vmovdqa		0x60(%rsp),%ymm0
  	vperm2i128	$0x20,%ymm7,%ymm0,%ymm1
  	vperm2i128	$0x31,%ymm7,%ymm0,%ymm7
  	vmovdqa		%ymm1,0x60(%rsp)
  	vperm2i128	$0x20,%ymm12,%ymm8,%ymm0
  	vperm2i128	$0x31,%ymm12,%ymm8,%ymm12
  	vmovdqa		%ymm0,%ymm8
  	vperm2i128	$0x20,%ymm13,%ymm9,%ymm0
  	vperm2i128	$0x31,%ymm13,%ymm9,%ymm13
  	vmovdqa		%ymm0,%ymm9
  	vperm2i128	$0x20,%ymm14,%ymm10,%ymm0
  	vperm2i128	$0x31,%ymm14,%ymm10,%ymm14
  	vmovdqa		%ymm0,%ymm10
  	vperm2i128	$0x20,%ymm15,%ymm11,%ymm0
  	vperm2i128	$0x31,%ymm15,%ymm11,%ymm15
  	vmovdqa		%ymm0,%ymm11
  
  	# xor with corresponding input, write to output
  	vmovdqa		0x00(%rsp),%ymm0
  	vpxor		0x0000(%rdx),%ymm0,%ymm0
  	vmovdqu		%ymm0,0x0000(%rsi)
  	vmovdqa		0x20(%rsp),%ymm0
  	vpxor		0x0080(%rdx),%ymm0,%ymm0
  	vmovdqu		%ymm0,0x0080(%rsi)
  	vmovdqa		0x40(%rsp),%ymm0
  	vpxor		0x0040(%rdx),%ymm0,%ymm0
  	vmovdqu		%ymm0,0x0040(%rsi)
  	vmovdqa		0x60(%rsp),%ymm0
  	vpxor		0x00c0(%rdx),%ymm0,%ymm0
  	vmovdqu		%ymm0,0x00c0(%rsi)
  	vpxor		0x0100(%rdx),%ymm4,%ymm4
  	vmovdqu		%ymm4,0x0100(%rsi)
  	vpxor		0x0180(%rdx),%ymm5,%ymm5
  	vmovdqu		%ymm5,0x00180(%rsi)
  	vpxor		0x0140(%rdx),%ymm6,%ymm6
  	vmovdqu		%ymm6,0x0140(%rsi)
  	vpxor		0x01c0(%rdx),%ymm7,%ymm7
  	vmovdqu		%ymm7,0x01c0(%rsi)
  	vpxor		0x0020(%rdx),%ymm8,%ymm8
  	vmovdqu		%ymm8,0x0020(%rsi)
  	vpxor		0x00a0(%rdx),%ymm9,%ymm9
  	vmovdqu		%ymm9,0x00a0(%rsi)
  	vpxor		0x0060(%rdx),%ymm10,%ymm10
  	vmovdqu		%ymm10,0x0060(%rsi)
  	vpxor		0x00e0(%rdx),%ymm11,%ymm11
  	vmovdqu		%ymm11,0x00e0(%rsi)
  	vpxor		0x0120(%rdx),%ymm12,%ymm12
  	vmovdqu		%ymm12,0x0120(%rsi)
  	vpxor		0x01a0(%rdx),%ymm13,%ymm13
  	vmovdqu		%ymm13,0x01a0(%rsi)
  	vpxor		0x0160(%rdx),%ymm14,%ymm14
  	vmovdqu		%ymm14,0x0160(%rsi)
  	vpxor		0x01e0(%rdx),%ymm15,%ymm15
  	vmovdqu		%ymm15,0x01e0(%rsi)
  
  	vzeroupper
  	mov		%r8,%rsp
  	ret
  ENDPROC(chacha20_8block_xor_avx2)