| # Check 64bit AVX512{VBMI2,VL} instructions |
| |
| .allow_index_reg |
| .text |
| _start: |
| vpcompressb %xmm30, (%rcx){%k7} # AVX512{VBMI2,VL} |
| vpcompressb %xmm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL} |
| vpcompressb %xmm30, 127(%rdx) # AVX512{VBMI2,VL} Disp8 |
| vpcompressb %ymm30, (%rcx){%k7} # AVX512{VBMI2,VL} |
| vpcompressb %ymm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL} |
| vpcompressb %ymm30, 127(%rdx) # AVX512{VBMI2,VL} Disp8 |
| vpcompressb %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpcompressb %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpcompressb %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpcompressb %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpcompressb %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpcompressb %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| |
| vpcompressw %xmm30, (%rcx){%k7} # AVX512{VBMI2,VL} |
| vpcompressw %xmm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL} |
| vpcompressw %xmm30, 254(%rdx) # AVX512{VBMI2,VL} Disp8 |
| vpcompressw %ymm30, (%rcx){%k7} # AVX512{VBMI2,VL} |
| vpcompressw %ymm30, 0x123(%rax,%r14,8) # AVX512{VBMI2,VL} |
| vpcompressw %ymm30, 254(%rdx) # AVX512{VBMI2,VL} Disp8 |
| vpcompressw %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpcompressw %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpcompressw %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpcompressw %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpcompressw %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpcompressw %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| |
| vpexpandb (%rcx), %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandb (%rcx), %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandb 0x123(%rax,%r14,8), %xmm30 # AVX512{VBMI2,VL} |
| vpexpandb 127(%rdx), %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpexpandb (%rcx), %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandb (%rcx), %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandb 0x123(%rax,%r14,8), %ymm30 # AVX512{VBMI2,VL} |
| vpexpandb 127(%rdx), %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpexpandb %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpexpandb %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandb %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandb %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpexpandb %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandb %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| |
| vpexpandw (%rcx), %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandw (%rcx), %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandw 0x123(%rax,%r14,8), %xmm30 # AVX512{VBMI2,VL} |
| vpexpandw 254(%rdx), %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpexpandw (%rcx), %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandw (%rcx), %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandw 0x123(%rax,%r14,8), %ymm30 # AVX512{VBMI2,VL} |
| vpexpandw 254(%rdx), %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpexpandw %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpexpandw %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandw %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpexpandw %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpexpandw %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpexpandw %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| |
| vpshldvw %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvw %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvw %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvw 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvw 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvw %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvw %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvw %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvw 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvw 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldvd %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvd %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvd %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvd 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvd 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvd 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvd %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvd %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvd %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvd 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvd 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvd 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldvq %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvq %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvq %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvq 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldvq 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvq 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvq %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvq %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldvq %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldvq 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldvq 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldvq 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvw %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvw %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvw %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvw 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvw 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvw %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvw %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvw %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvw 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvw 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvd %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvd %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvd %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvd 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvd 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvd %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvd %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvd 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvd 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvq %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvq %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvq %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvq 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdvq 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvq %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdvq %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdvq 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdvq 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldw $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldw $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldw $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldw $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldw $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldw $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldw $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldw $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldw $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldw $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldd $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldd $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldd $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldd $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldd $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldd $123, 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldd $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldd $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldd $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldd $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldd $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldd $123, 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldq $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldq $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshldq $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldq $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshldq $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldq $123, 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldq $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldq $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshldq $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshldq $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshldq $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshldq $123, 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdw $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdw $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdw $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdw $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdw $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdw $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdw $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdw $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdw $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdw $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdd $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdd $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdd $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdd $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdd $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd $123, 508(%rdx){1to4}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdd $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdd $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdd $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdd $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd $123, 508(%rdx){1to8}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdq $0xab, %xmm28, %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdq $0xab, %xmm28, %xmm29, %xmm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdq $0xab, %xmm28, %xmm29, %xmm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdq $123, 0x123(%rax,%r14,8), %xmm29, %xmm30 # AVX512{VBMI2,VL} |
| vpshrdq $123, 2032(%rdx), %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq $123, 1016(%rdx){1to2}, %xmm29, %xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq $0xab, %ymm28, %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdq $0xab, %ymm28, %ymm29, %ymm30{%k7} # AVX512{VBMI2,VL} |
| vpshrdq $0xab, %ymm28, %ymm29, %ymm30{%k7}{z} # AVX512{VBMI2,VL} |
| vpshrdq $123, 0x123(%rax,%r14,8), %ymm29, %ymm30 # AVX512{VBMI2,VL} |
| vpshrdq $123, 4064(%rdx), %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq $123, 1016(%rdx){1to4}, %ymm29, %ymm30 # AVX512{VBMI2,VL} Disp8 |
| |
| .intel_syntax noprefix |
| vpcompressb XMMWORD PTR [rcx]{k7}, xmm30 # AVX512{VBMI2,VL} |
| vpcompressb XMMWORD PTR [rax+r14*8+0x1234], xmm30 # AVX512{VBMI2,VL} |
| vpcompressb XMMWORD PTR [rdx+127], xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpcompressb YMMWORD PTR [rcx]{k7}, ymm30 # AVX512{VBMI2,VL} |
| vpcompressb YMMWORD PTR [rax+r14*8+0x1234], ymm30 # AVX512{VBMI2,VL} |
| vpcompressb YMMWORD PTR [rdx+127], ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpcompressb xmm30, xmm29 # AVX512{VBMI2,VL} |
| vpcompressb xmm30{k7}, xmm29 # AVX512{VBMI2,VL} |
| vpcompressb xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL} |
| vpcompressb ymm30, ymm29 # AVX512{VBMI2,VL} |
| vpcompressb ymm30{k7}, ymm29 # AVX512{VBMI2,VL} |
| vpcompressb ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL} |
| |
| vpcompressw XMMWORD PTR [rcx]{k7}, xmm30 # AVX512{VBMI2,VL} |
| vpcompressw XMMWORD PTR [rax+r14*8+0x1234], xmm30 # AVX512{VBMI2,VL} |
| vpcompressw XMMWORD PTR [rdx+254], xmm30 # AVX512{VBMI2,VL} Disp8 |
| vpcompressw YMMWORD PTR [rcx]{k7}, ymm30 # AVX512{VBMI2,VL} |
| vpcompressw YMMWORD PTR [rax+r14*8+0x1234], ymm30 # AVX512{VBMI2,VL} |
| vpcompressw YMMWORD PTR [rdx+254], ymm30 # AVX512{VBMI2,VL} Disp8 |
| vpcompressw xmm30, xmm29 # AVX512{VBMI2,VL} |
| vpcompressw xmm30{k7}, xmm29 # AVX512{VBMI2,VL} |
| vpcompressw xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL} |
| vpcompressw ymm30, ymm29 # AVX512{VBMI2,VL} |
| vpcompressw ymm30{k7}, ymm29 # AVX512{VBMI2,VL} |
| vpcompressw ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL} |
| |
| vpexpandb xmm30{k7}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandb xmm30{k7}{z}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandb xmm30, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpexpandb xmm30, XMMWORD PTR [rdx+127] # AVX512{VBMI2,VL} Disp8 |
| vpexpandb ymm30{k7}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandb ymm30{k7}{z}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandb ymm30, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpexpandb ymm30, YMMWORD PTR [rdx+127] # AVX512{VBMI2,VL} Disp8 |
| vpexpandb xmm30, xmm29 # AVX512{VBMI2,VL} |
| vpexpandb xmm30{k7}, xmm29 # AVX512{VBMI2,VL} |
| vpexpandb xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL} |
| vpexpandb ymm30, ymm29 # AVX512{VBMI2,VL} |
| vpexpandb ymm30{k7}, ymm29 # AVX512{VBMI2,VL} |
| vpexpandb ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL} |
| |
| vpexpandw xmm30{k7}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandw xmm30{k7}{z}, XMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandw xmm30, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpexpandw xmm30, XMMWORD PTR [rdx+254] # AVX512{VBMI2,VL} Disp8 |
| vpexpandw ymm30{k7}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandw ymm30{k7}{z}, YMMWORD PTR [rcx] # AVX512{VBMI2,VL} |
| vpexpandw ymm30, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpexpandw ymm30, YMMWORD PTR [rdx+254] # AVX512{VBMI2,VL} Disp8 |
| vpexpandw xmm30, xmm29 # AVX512{VBMI2,VL} |
| vpexpandw xmm30{k7}, xmm29 # AVX512{VBMI2,VL} |
| vpexpandw xmm30{k7}{z}, xmm29 # AVX512{VBMI2,VL} |
| vpexpandw ymm30, ymm29 # AVX512{VBMI2,VL} |
| vpexpandw ymm30{k7}, ymm29 # AVX512{VBMI2,VL} |
| vpexpandw ymm30{k7}{z}, ymm29 # AVX512{VBMI2,VL} |
| |
| vpshldvw xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvw xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvw xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvw xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshldvw ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvw ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvw ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvw ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldvd xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvd xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvd xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvd xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshldvd xmm30, xmm29, [rdx+508]{1to4} # AVX512{VBMI2,VL} Disp8 |
| vpshldvd ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvd ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvd ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvd ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| vpshldvd ymm30, ymm29, [rdx+508]{1to8} # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldvq xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvq xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvq xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshldvq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvq xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshldvq xmm30, xmm29, [rdx+1016]{1to2} # AVX512{VBMI2,VL} Disp8 |
| vpshldvq ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvq ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvq ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshldvq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshldvq ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| vpshldvq ymm30, ymm29, [rdx+1016]{1to4} # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvw xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvw xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvw xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvw xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshrdvw ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvw ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvw ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvw ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvd xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvd xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvd xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvd xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd xmm30, xmm29, [rdx+508]{1to4} # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvd ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvd ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvd ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| vpshrdvd ymm30, ymm29, [rdx+508]{1to8} # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdvq xmm30, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvq xmm30{k7}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvq xmm30{k7}{z}, xmm29, xmm28 # AVX512{VBMI2,VL} |
| vpshrdvq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvq xmm30, xmm29, XMMWORD PTR [rdx+2032] # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq xmm30, xmm29, [rdx+1016]{1to2} # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq ymm30, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvq ymm30{k7}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvq ymm30{k7}{z}, ymm29, ymm28 # AVX512{VBMI2,VL} |
| vpshrdvq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234] # AVX512{VBMI2,VL} |
| vpshrdvq ymm30, ymm29, YMMWORD PTR [rdx+4064] # AVX512{VBMI2,VL} Disp8 |
| vpshrdvq ymm30, ymm29, [rdx+1016]{1to4} # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldw xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldw xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldw ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldw ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldd xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldd xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldd xmm30, xmm29, [rdx+508]{1to4}, 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldd ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldd ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldd ymm30, ymm29, [rdx+508]{1to8}, 123 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshldq xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldq xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldq xmm30, xmm29, [rdx+1016]{1to2}, 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldq ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshldq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshldq ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshldq ymm30, ymm29, [rdx+1016]{1to4}, 123 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdw xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdw xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdw ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdw ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdw ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| |
| vpshrdd xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdd xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd xmm30, xmm29, [rdx+508]{1to4}, 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdd ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdd ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdd ymm30, ymm29, [rdx+508]{1to8}, 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq xmm30, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq xmm30{k7}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq xmm30{k7}{z}, xmm29, xmm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq xmm30, xmm29, XMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdq xmm30, xmm29, XMMWORD PTR [rdx+2032], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq xmm30, xmm29, [rdx+1016]{1to2}, 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq ymm30, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq ymm30{k7}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq ymm30{k7}{z}, ymm29, ymm28, 0xab # AVX512{VBMI2,VL} |
| vpshrdq ymm30, ymm29, YMMWORD PTR [rax+r14*8+0x1234], 123 # AVX512{VBMI2,VL} |
| vpshrdq ymm30, ymm29, YMMWORD PTR [rdx+4064], 123 # AVX512{VBMI2,VL} Disp8 |
| vpshrdq ymm30, ymm29, [rdx+1016]{1to4}, 123 # AVX512{VBMI2,VL} Disp8 |