File: x86-64-avx512vbmi2_vl.s

package info (click to toggle)
binutils 2.31.1-16
  • links: PTS, VCS
  • area: main
  • in suites: buster
  • size: 309,412 kB
  • sloc: ansic: 1,161,194; asm: 638,508; cpp: 128,829; exp: 68,580; makefile: 55,828; sh: 22,360; yacc: 14,238; lisp: 13,272; perl: 2,111; ada: 1,681; lex: 1,652; pascal: 1,446; cs: 879; sed: 195; python: 154; xml: 95; awk: 25
file content (412 lines) | stat: -rw-r--r-- 23,380 bytes parent folder | download | duplicates (13)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
# Check 64bit AVX512{VBMI2,VL} instructions

	.allow_index_reg
	.text
_start:
	vpcompressb	%xmm30, (%rcx){%k7}	 # AVX512{VBMI2,VL}
	vpcompressb	%xmm30, 0x123(%rax,%r14,8)	 # AVX512{VBMI2,VL}
	vpcompressb	%xmm30, 127(%rdx)	 # AVX512{VBMI2,VL} Disp8
	vpcompressb	%ymm30, (%rcx){%k7}	 # AVX512{VBMI2,VL}
	vpcompressb	%ymm30, 0x123(%rax,%r14,8)	 # AVX512{VBMI2,VL}
	vpcompressb	%ymm30, 127(%rdx)	 # AVX512{VBMI2,VL} Disp8
	vpcompressb	%xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpcompressb	%xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpcompressb	%xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpcompressb	%ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpcompressb	%ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpcompressb	%ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}

	vpcompressw	%xmm30, (%rcx){%k7}	 # AVX512{VBMI2,VL}
	vpcompressw	%xmm30, 0x123(%rax,%r14,8)	 # AVX512{VBMI2,VL}
	vpcompressw	%xmm30, 254(%rdx)	 # AVX512{VBMI2,VL} Disp8
	vpcompressw	%ymm30, (%rcx){%k7}	 # AVX512{VBMI2,VL}
	vpcompressw	%ymm30, 0x123(%rax,%r14,8)	 # AVX512{VBMI2,VL}
	vpcompressw	%ymm30, 254(%rdx)	 # AVX512{VBMI2,VL} Disp8
	vpcompressw	%xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpcompressw	%xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpcompressw	%xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpcompressw	%ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpcompressw	%ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpcompressw	%ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}

	vpexpandb	(%rcx), %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandb	(%rcx), %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandb	0x123(%rax,%r14,8), %xmm30	 # AVX512{VBMI2,VL}
	vpexpandb	127(%rdx), %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpexpandb	(%rcx), %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandb	(%rcx), %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandb	0x123(%rax,%r14,8), %ymm30	 # AVX512{VBMI2,VL}
	vpexpandb	127(%rdx), %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpexpandb	%xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpexpandb	%xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandb	%xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandb	%ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpexpandb	%ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandb	%ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}

	vpexpandw	(%rcx), %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandw	(%rcx), %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandw	0x123(%rax,%r14,8), %xmm30	 # AVX512{VBMI2,VL}
	vpexpandw	254(%rdx), %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpexpandw	(%rcx), %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandw	(%rcx), %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandw	0x123(%rax,%r14,8), %ymm30	 # AVX512{VBMI2,VL}
	vpexpandw	254(%rdx), %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpexpandw	%xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpexpandw	%xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandw	%xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpexpandw	%ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpexpandw	%ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpexpandw	%ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}

	vpshldvw	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvw	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvw	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvw	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvw	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvw	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvw	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvw	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvw	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvw	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshldvd	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvd	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvd	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvd	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvd	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	508(%rdx){1to4}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvd	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvd	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvd	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvd	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	508(%rdx){1to8}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshldvq	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvq	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvq	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvq	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldvq	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	1016(%rdx){1to2}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvq	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldvq	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldvq	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldvq	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	1016(%rdx){1to4}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdvw	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvw	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvw	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvw	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvw	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvw	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvw	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvw	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvw	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvw	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdvd	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvd	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvd	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvd	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvd	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	508(%rdx){1to4}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvd	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvd	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvd	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvd	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	508(%rdx){1to8}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdvq	%xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvq	%xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvq	%xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvq	0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdvq	2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	1016(%rdx){1to2}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	%ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvq	%ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdvq	%ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdvq	0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdvq	4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	1016(%rdx){1to4}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshldw	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldw	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldw	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldw	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldw	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldw	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldw	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldw	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldw	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldw	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshldd	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldd	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldd	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldd	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldd	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldd	$123, 508(%rdx){1to4}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldd	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldd	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldd	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldd	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldd	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshldd	$123, 508(%rdx){1to8}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshldq	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldq	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldq	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldq	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshldq	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldq	$123, 1016(%rdx){1to2}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshldq	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldq	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshldq	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshldq	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshldq	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshldq	$123, 1016(%rdx){1to4}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdw	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdw	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdw	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdw	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdw	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdw	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdw	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdw	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdw	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdw	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdd	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdd	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdd	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdd	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdd	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	$123, 508(%rdx){1to4}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdd	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdd	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdd	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdd	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	$123, 508(%rdx){1to8}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	vpshrdq	$0xab, %xmm28, %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdq	$0xab, %xmm28, %xmm29, %xmm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdq	$0xab, %xmm28, %xmm29, %xmm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdq	$123, 0x123(%rax,%r14,8), %xmm29, %xmm30	 # AVX512{VBMI2,VL}
	vpshrdq	$123, 2032(%rdx), %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	$123, 1016(%rdx){1to2}, %xmm29, %xmm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	$0xab, %ymm28, %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdq	$0xab, %ymm28, %ymm29, %ymm30{%k7}	 # AVX512{VBMI2,VL}
	vpshrdq	$0xab, %ymm28, %ymm29, %ymm30{%k7}{z}	 # AVX512{VBMI2,VL}
	vpshrdq	$123, 0x123(%rax,%r14,8), %ymm29, %ymm30	 # AVX512{VBMI2,VL}
	vpshrdq	$123, 4064(%rdx), %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	$123, 1016(%rdx){1to4}, %ymm29, %ymm30	 # AVX512{VBMI2,VL} Disp8

	.intel_syntax noprefix
	vpcompressb	XMMWORD PTR [rcx]{k7}, xmm30	 # AVX512{VBMI2,VL}
	vpcompressb	XMMWORD PTR [rax+r14*8+0x1234], xmm30	 # AVX512{VBMI2,VL}
	vpcompressb	XMMWORD PTR [rdx+127], xmm30	 # AVX512{VBMI2,VL} Disp8
	vpcompressb	YMMWORD PTR [rcx]{k7}, ymm30	 # AVX512{VBMI2,VL}
	vpcompressb	YMMWORD PTR [rax+r14*8+0x1234], ymm30	 # AVX512{VBMI2,VL}
	vpcompressb	YMMWORD PTR [rdx+127], ymm30	 # AVX512{VBMI2,VL} Disp8
	vpcompressb	xmm30, xmm29	 # AVX512{VBMI2,VL}
	vpcompressb	xmm30{k7}, xmm29	 # AVX512{VBMI2,VL}
	vpcompressb	xmm30{k7}{z}, xmm29	 # AVX512{VBMI2,VL}
	vpcompressb	ymm30, ymm29	 # AVX512{VBMI2,VL}
	vpcompressb	ymm30{k7}, ymm29	 # AVX512{VBMI2,VL}
	vpcompressb	ymm30{k7}{z}, ymm29	 # AVX512{VBMI2,VL}

	vpcompressw	XMMWORD PTR [rcx]{k7}, xmm30	 # AVX512{VBMI2,VL}
	vpcompressw	XMMWORD PTR [rax+r14*8+0x1234], xmm30	 # AVX512{VBMI2,VL}
	vpcompressw	XMMWORD PTR [rdx+254], xmm30	 # AVX512{VBMI2,VL} Disp8
	vpcompressw	YMMWORD PTR [rcx]{k7}, ymm30	 # AVX512{VBMI2,VL}
	vpcompressw	YMMWORD PTR [rax+r14*8+0x1234], ymm30	 # AVX512{VBMI2,VL}
	vpcompressw	YMMWORD PTR [rdx+254], ymm30	 # AVX512{VBMI2,VL} Disp8
	vpcompressw	xmm30, xmm29	 # AVX512{VBMI2,VL}
	vpcompressw	xmm30{k7}, xmm29	 # AVX512{VBMI2,VL}
	vpcompressw	xmm30{k7}{z}, xmm29	 # AVX512{VBMI2,VL}
	vpcompressw	ymm30, ymm29	 # AVX512{VBMI2,VL}
	vpcompressw	ymm30{k7}, ymm29	 # AVX512{VBMI2,VL}
	vpcompressw	ymm30{k7}{z}, ymm29	 # AVX512{VBMI2,VL}

	vpexpandb	xmm30{k7}, XMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandb	xmm30{k7}{z}, XMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandb	xmm30, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpexpandb	xmm30, XMMWORD PTR [rdx+127]	 # AVX512{VBMI2,VL} Disp8
	vpexpandb	ymm30{k7}, YMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30{k7}{z}, YMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30, YMMWORD PTR [rdx+127]	 # AVX512{VBMI2,VL} Disp8
	vpexpandb	xmm30, xmm29	 # AVX512{VBMI2,VL}
	vpexpandb	xmm30{k7}, xmm29	 # AVX512{VBMI2,VL}
	vpexpandb	xmm30{k7}{z}, xmm29	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30, ymm29	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30{k7}, ymm29	 # AVX512{VBMI2,VL}
	vpexpandb	ymm30{k7}{z}, ymm29	 # AVX512{VBMI2,VL}

	vpexpandw	xmm30{k7}, XMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandw	xmm30{k7}{z}, XMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandw	xmm30, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpexpandw	xmm30, XMMWORD PTR [rdx+254]	 # AVX512{VBMI2,VL} Disp8
	vpexpandw	ymm30{k7}, YMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30{k7}{z}, YMMWORD PTR [rcx]	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30, YMMWORD PTR [rdx+254]	 # AVX512{VBMI2,VL} Disp8
	vpexpandw	xmm30, xmm29	 # AVX512{VBMI2,VL}
	vpexpandw	xmm30{k7}, xmm29	 # AVX512{VBMI2,VL}
	vpexpandw	xmm30{k7}{z}, xmm29	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30, ymm29	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30{k7}, ymm29	 # AVX512{VBMI2,VL}
	vpexpandw	ymm30{k7}{z}, ymm29	 # AVX512{VBMI2,VL}

	vpshldvw	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvw	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvw	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvw	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvw	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshldvw	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvw	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvw	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvw	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvw	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8

	vpshldvd	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvd	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvd	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvd	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvd	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	xmm30, xmm29, [rdx+508]{1to4}	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvd	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvd	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvd	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvd	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8
	vpshldvd	ymm30, ymm29, [rdx+508]{1to8}	 # AVX512{VBMI2,VL} Disp8

	vpshldvq	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvq	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvq	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshldvq	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvq	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	xmm30, xmm29, [rdx+1016]{1to2}	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvq	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvq	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshldvq	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshldvq	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8
	vpshldvq	ymm30, ymm29, [rdx+1016]{1to4}	 # AVX512{VBMI2,VL} Disp8

	vpshrdvw	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvw	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvw	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvw	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvw	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshrdvw	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvw	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvw	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvw	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvw	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8

	vpshrdvd	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvd	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvd	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvd	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvd	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	xmm30, xmm29, [rdx+508]{1to4}	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvd	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvd	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvd	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvd	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8
	vpshrdvd	ymm30, ymm29, [rdx+508]{1to8}	 # AVX512{VBMI2,VL} Disp8

	vpshrdvq	xmm30, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvq	xmm30{k7}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvq	xmm30{k7}{z}, xmm29, xmm28	 # AVX512{VBMI2,VL}
	vpshrdvq	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvq	xmm30, xmm29, XMMWORD PTR [rdx+2032]	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	xmm30, xmm29, [rdx+1016]{1to2}	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	ymm30, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvq	ymm30{k7}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvq	ymm30{k7}{z}, ymm29, ymm28	 # AVX512{VBMI2,VL}
	vpshrdvq	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234]	 # AVX512{VBMI2,VL}
	vpshrdvq	ymm30, ymm29, YMMWORD PTR [rdx+4064]	 # AVX512{VBMI2,VL} Disp8
	vpshrdvq	ymm30, ymm29, [rdx+1016]{1to4}	 # AVX512{VBMI2,VL} Disp8

	vpshldw	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldw	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshldw	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldw	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldw	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8

	vpshldd	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldd	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshldd	xmm30, xmm29, [rdx+508]{1to4}, 123	 # AVX512{VBMI2,VL} Disp8
	vpshldd	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldd	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldd	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8
	vpshldd	ymm30, ymm29, [rdx+508]{1to8}, 123	 # AVX512{VBMI2,VL} Disp8

	vpshldq	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldq	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshldq	xmm30, xmm29, [rdx+1016]{1to2}, 123	 # AVX512{VBMI2,VL} Disp8
	vpshldq	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshldq	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshldq	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8
	vpshldq	ymm30, ymm29, [rdx+1016]{1to4}, 123	 # AVX512{VBMI2,VL} Disp8

	vpshrdw	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdw	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdw	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdw	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdw	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8

	vpshrdd	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdd	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	xmm30, xmm29, [rdx+508]{1to4}, 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdd	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdd	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdd	ymm30, ymm29, [rdx+508]{1to8}, 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	xmm30, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	xmm30{k7}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	xmm30{k7}{z}, xmm29, xmm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdq	xmm30, xmm29, XMMWORD PTR [rdx+2032], 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	xmm30, xmm29, [rdx+1016]{1to2}, 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	ymm30, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	ymm30{k7}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	ymm30{k7}{z}, ymm29, ymm28, 0xab	 # AVX512{VBMI2,VL}
	vpshrdq	ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123	 # AVX512{VBMI2,VL}
	vpshrdq	ymm30, ymm29, YMMWORD PTR [rdx+4064], 123	 # AVX512{VBMI2,VL} Disp8
	vpshrdq	ymm30, ymm29, [rdx+1016]{1to4}, 123	 # AVX512{VBMI2,VL} Disp8