You've already forked linux-packaging-mono
							
							
		
			
				
	
	
		
			372 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
			
		
		
	
	
			372 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			LLVM
		
	
	
	
	
	
| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
 | |
| ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X32
 | |
| ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X64
 | |
| 
 | |
| ; Splat patterns below
 | |
| 
 | |
| define <4 x i32> @shl4(<4 x i32> %A) nounwind {
 | |
| ; X32-LABEL: shl4:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    pslld $2, %xmm1
 | |
| ; X32-NEXT:    paddd %xmm0, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shl4:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    pslld $2, %xmm1
 | |
| ; X64-NEXT:    paddd %xmm0, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = shl <4 x i32> %A,  < i32 2, i32 2, i32 2, i32 2>
 | |
|   %C = shl <4 x i32> %A,  < i32 1, i32 1, i32 1, i32 1>
 | |
|   %K = xor <4 x i32> %B, %C
 | |
|   ret <4 x i32> %K
 | |
| }
 | |
| 
 | |
| define <4 x i32> @shr4(<4 x i32> %A) nounwind {
 | |
| ; X32-LABEL: shr4:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrld $2, %xmm1
 | |
| ; X32-NEXT:    psrld $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr4:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrld $2, %xmm1
 | |
| ; X64-NEXT:    psrld $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = lshr <4 x i32> %A,  < i32 2, i32 2, i32 2, i32 2>
 | |
|   %C = lshr <4 x i32> %A,  < i32 1, i32 1, i32 1, i32 1>
 | |
|   %K = xor <4 x i32> %B, %C
 | |
|   ret <4 x i32> %K
 | |
| }
 | |
| 
 | |
| define <4 x i32> @sra4(<4 x i32> %A) nounwind {
 | |
| ; X32-LABEL: sra4:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrad $2, %xmm1
 | |
| ; X32-NEXT:    psrad $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: sra4:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrad $2, %xmm1
 | |
| ; X64-NEXT:    psrad $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = ashr <4 x i32> %A,  < i32 2, i32 2, i32 2, i32 2>
 | |
|   %C = ashr <4 x i32> %A,  < i32 1, i32 1, i32 1, i32 1>
 | |
|   %K = xor <4 x i32> %B, %C
 | |
|   ret <4 x i32> %K
 | |
| }
 | |
| 
 | |
| define <2 x i64> @shl2(<2 x i64> %A) nounwind {
 | |
| ; X32-LABEL: shl2:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psllq $2, %xmm1
 | |
| ; X32-NEXT:    psllq $9, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shl2:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psllq $2, %xmm1
 | |
| ; X64-NEXT:    psllq $9, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = shl <2 x i64> %A,  < i64 2, i64 2>
 | |
|   %C = shl <2 x i64> %A,  < i64 9, i64 9>
 | |
|   %K = xor <2 x i64> %B, %C
 | |
|   ret <2 x i64> %K
 | |
| }
 | |
| 
 | |
| define <2 x i64> @shr2(<2 x i64> %A) nounwind {
 | |
| ; X32-LABEL: shr2:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrlq $8, %xmm1
 | |
| ; X32-NEXT:    psrlq $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr2:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrlq $8, %xmm1
 | |
| ; X64-NEXT:    psrlq $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = lshr <2 x i64> %A,  < i64 8, i64 8>
 | |
|   %C = lshr <2 x i64> %A,  < i64 1, i64 1>
 | |
|   %K = xor <2 x i64> %B, %C
 | |
|   ret <2 x i64> %K
 | |
| }
 | |
| 
 | |
| 
 | |
| define <8 x i16> @shl8(<8 x i16> %A) nounwind {
 | |
| ; X32-LABEL: shl8:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psllw $2, %xmm1
 | |
| ; X32-NEXT:    paddw %xmm0, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shl8:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psllw $2, %xmm1
 | |
| ; X64-NEXT:    paddw %xmm0, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = shl <8 x i16> %A,  < i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2>
 | |
|   %C = shl <8 x i16> %A,  < i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>
 | |
|   %K = xor <8 x i16> %B, %C
 | |
|   ret <8 x i16> %K
 | |
| }
 | |
| 
 | |
| define <8 x i16> @shr8(<8 x i16> %A) nounwind {
 | |
| ; X32-LABEL: shr8:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrlw $2, %xmm1
 | |
| ; X32-NEXT:    psrlw $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr8:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrlw $2, %xmm1
 | |
| ; X64-NEXT:    psrlw $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = lshr <8 x i16> %A,  < i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2>
 | |
|   %C = lshr <8 x i16> %A,  < i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>
 | |
|   %K = xor <8 x i16> %B, %C
 | |
|   ret <8 x i16> %K
 | |
| }
 | |
| 
 | |
| define <8 x i16> @sra8(<8 x i16> %A) nounwind {
 | |
| ; X32-LABEL: sra8:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psraw $2, %xmm1
 | |
| ; X32-NEXT:    psraw $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: sra8:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psraw $2, %xmm1
 | |
| ; X64-NEXT:    psraw $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = ashr <8 x i16> %A,  < i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2, i16 2>
 | |
|   %C = ashr <8 x i16> %A,  < i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1, i16 1>
 | |
|   %K = xor <8 x i16> %B, %C
 | |
|   ret <8 x i16> %K
 | |
| }
 | |
| 
 | |
| ; non-splat test
 | |
| 
 | |
| 
 | |
| define <8 x i16> @sll8_nosplat(<8 x i16> %A) nounwind {
 | |
| ; X32-LABEL: sll8_nosplat:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa {{.*#+}} xmm1 = [2,4,8,64,4,4,4,4]
 | |
| ; X32-NEXT:    pmullw %xmm0, %xmm1
 | |
| ; X32-NEXT:    pmullw {{\.LCPI.*}}, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: sll8_nosplat:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa {{.*#+}} xmm1 = [2,4,8,64,4,4,4,4]
 | |
| ; X64-NEXT:    pmullw %xmm0, %xmm1
 | |
| ; X64-NEXT:    pmullw {{.*}}(%rip), %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = shl <8 x i16> %A,  < i16 1, i16 2, i16 3, i16 6, i16 2, i16 2, i16 2, i16 2>
 | |
|   %C = shl <8 x i16> %A,  < i16 9, i16 7, i16 5, i16 1, i16 4, i16 1, i16 1, i16 1>
 | |
|   %K = xor <8 x i16> %B, %C
 | |
|   ret <8 x i16> %K
 | |
| }
 | |
| 
 | |
| 
 | |
| define <2 x i64> @shr2_nosplat(<2 x i64> %A) nounwind {
 | |
| ; X32-LABEL: shr2_nosplat:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm2
 | |
| ; X32-NEXT:    psrlq $8, %xmm2
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrlq $1, %xmm1
 | |
| ; X32-NEXT:    movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
 | |
| ; X32-NEXT:    movsd {{.*#+}} xmm1 = xmm2[0],xmm1[1]
 | |
| ; X32-NEXT:    xorpd %xmm0, %xmm1
 | |
| ; X32-NEXT:    movapd %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr2_nosplat:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm2
 | |
| ; X64-NEXT:    psrlq $8, %xmm2
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrlq $1, %xmm1
 | |
| ; X64-NEXT:    movsd {{.*#+}} xmm0 = xmm1[0],xmm0[1]
 | |
| ; X64-NEXT:    movsd {{.*#+}} xmm1 = xmm2[0],xmm1[1]
 | |
| ; X64-NEXT:    xorpd %xmm0, %xmm1
 | |
| ; X64-NEXT:    movapd %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = lshr <2 x i64> %A,  < i64 8, i64 1>
 | |
|   %C = lshr <2 x i64> %A,  < i64 1, i64 0>
 | |
|   %K = xor <2 x i64> %B, %C
 | |
|   ret <2 x i64> %K
 | |
| }
 | |
| 
 | |
| 
 | |
| ; Other shifts
 | |
| 
 | |
| define <2 x i32> @shl2_other(<2 x i32> %A) nounwind {
 | |
| ; X32-LABEL: shl2_other:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psllq $2, %xmm1
 | |
| ; X32-NEXT:    psllq $9, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shl2_other:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psllq $2, %xmm1
 | |
| ; X64-NEXT:    psllq $9, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = shl <2 x i32> %A,  < i32 2, i32 2>
 | |
|   %C = shl <2 x i32> %A,  < i32 9, i32 9>
 | |
|   %K = xor <2 x i32> %B, %C
 | |
|   ret <2 x i32> %K
 | |
| }
 | |
| 
 | |
| define <2 x i32> @shr2_other(<2 x i32> %A) nounwind {
 | |
| ; X32-LABEL: shr2_other:
 | |
| ; X32:       # %bb.0: # %entry
 | |
| ; X32-NEXT:    pand {{\.LCPI.*}}, %xmm0
 | |
| ; X32-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X32-NEXT:    psrlq $8, %xmm1
 | |
| ; X32-NEXT:    psrlq $1, %xmm0
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr2_other:
 | |
| ; X64:       # %bb.0: # %entry
 | |
| ; X64-NEXT:    pand {{.*}}(%rip), %xmm0
 | |
| ; X64-NEXT:    movdqa %xmm0, %xmm1
 | |
| ; X64-NEXT:    psrlq $8, %xmm1
 | |
| ; X64-NEXT:    psrlq $1, %xmm0
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
| entry:
 | |
|   %B = lshr <2 x i32> %A,  < i32 8, i32 8>
 | |
|   %C = lshr <2 x i32> %A,  < i32 1, i32 1>
 | |
|   %K = xor <2 x i32> %B, %C
 | |
|   ret <2 x i32> %K
 | |
| }
 | |
| 
 | |
| define <16 x i8> @shl9(<16 x i8> %A) nounwind {
 | |
| ; X32-LABEL: shl9:
 | |
| ; X32:       # %bb.0:
 | |
| ; X32-NEXT:    psllw $3, %xmm0
 | |
| ; X32-NEXT:    pand {{\.LCPI.*}}, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shl9:
 | |
| ; X64:       # %bb.0:
 | |
| ; X64-NEXT:    psllw $3, %xmm0
 | |
| ; X64-NEXT:    pand {{.*}}(%rip), %xmm0
 | |
| ; X64-NEXT:    retq
 | |
|   %B = shl <16 x i8> %A, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
 | |
|   ret <16 x i8> %B
 | |
| }
 | |
| 
 | |
| define <16 x i8> @shr9(<16 x i8> %A) nounwind {
 | |
| ; X32-LABEL: shr9:
 | |
| ; X32:       # %bb.0:
 | |
| ; X32-NEXT:    psrlw $3, %xmm0
 | |
| ; X32-NEXT:    pand {{\.LCPI.*}}, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: shr9:
 | |
| ; X64:       # %bb.0:
 | |
| ; X64-NEXT:    psrlw $3, %xmm0
 | |
| ; X64-NEXT:    pand {{.*}}(%rip), %xmm0
 | |
| ; X64-NEXT:    retq
 | |
|   %B = lshr <16 x i8> %A, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
 | |
|   ret <16 x i8> %B
 | |
| }
 | |
| 
 | |
| define <16 x i8> @sra_v16i8_7(<16 x i8> %A) nounwind {
 | |
| ; X32-LABEL: sra_v16i8_7:
 | |
| ; X32:       # %bb.0:
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm1
 | |
| ; X32-NEXT:    pcmpgtb %xmm0, %xmm1
 | |
| ; X32-NEXT:    movdqa %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: sra_v16i8_7:
 | |
| ; X64:       # %bb.0:
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm1
 | |
| ; X64-NEXT:    pcmpgtb %xmm0, %xmm1
 | |
| ; X64-NEXT:    movdqa %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
|   %B = ashr <16 x i8> %A, <i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7>
 | |
|   ret <16 x i8> %B
 | |
| }
 | |
| 
 | |
| define <16 x i8> @sra_v16i8(<16 x i8> %A) nounwind {
 | |
| ; X32-LABEL: sra_v16i8:
 | |
| ; X32:       # %bb.0:
 | |
| ; X32-NEXT:    psrlw $3, %xmm0
 | |
| ; X32-NEXT:    pand {{\.LCPI.*}}, %xmm0
 | |
| ; X32-NEXT:    movdqa {{.*#+}} xmm1 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
 | |
| ; X32-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X32-NEXT:    psubb %xmm1, %xmm0
 | |
| ; X32-NEXT:    retl
 | |
| ;
 | |
| ; X64-LABEL: sra_v16i8:
 | |
| ; X64:       # %bb.0:
 | |
| ; X64-NEXT:    psrlw $3, %xmm0
 | |
| ; X64-NEXT:    pand {{.*}}(%rip), %xmm0
 | |
| ; X64-NEXT:    movdqa {{.*#+}} xmm1 = [16,16,16,16,16,16,16,16,16,16,16,16,16,16,16,16]
 | |
| ; X64-NEXT:    pxor %xmm1, %xmm0
 | |
| ; X64-NEXT:    psubb %xmm1, %xmm0
 | |
| ; X64-NEXT:    retq
 | |
|   %B = ashr <16 x i8> %A, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3>
 | |
|   ret <16 x i8> %B
 | |
| }
 |