91 lines
		
	
	
		
			3.4 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			91 lines
		
	
	
		
			3.4 KiB
		
	
	
	
		
			LLVM
		
	
	
	
; RUN: llc -mtriple=x86_64-linux -mcpu=nehalem < %s | FileCheck %s --check-prefix=LIN
 | 
						|
; RUN: llc -mtriple=x86_64-win32 -mcpu=nehalem < %s | FileCheck %s --check-prefix=WIN
 | 
						|
; RUN: llc -mtriple=i686-win32 -mcpu=nehalem < %s | FileCheck %s --check-prefix=LIN32
 | 
						|
; rdar://7398554
 | 
						|
 | 
						|
; When doing vector gather-scatter index calculation with 32-bit indices,
 | 
						|
; use an efficient mov/shift sequence rather than shuffling each individual
 | 
						|
; element out of the index vector.
 | 
						|
 | 
						|
; CHECK-LABEL: foo:
 | 
						|
; LIN: movdqa	(%rsi), %xmm0
 | 
						|
; LIN: pand 	(%rdx), %xmm0
 | 
						|
; LIN: pextrq	$1, %xmm0, %r[[REG4:.+]]
 | 
						|
; LIN: movd 	%xmm0, %r[[REG2:.+]]
 | 
						|
; LIN: movslq	%e[[REG2]], %r[[REG1:.+]]
 | 
						|
; LIN: sarq    $32, %r[[REG2]]
 | 
						|
; LIN: movslq	%e[[REG4]], %r[[REG3:.+]]
 | 
						|
; LIN: sarq    $32, %r[[REG4]]
 | 
						|
; LIN: movsd	(%rdi,%r[[REG1]],8), %xmm0
 | 
						|
; LIN: movhpd	(%rdi,%r[[REG2]],8), %xmm0
 | 
						|
; LIN: movsd	(%rdi,%r[[REG3]],8), %xmm1
 | 
						|
; LIN: movhpd	(%rdi,%r[[REG4]],8), %xmm1
 | 
						|
 | 
						|
; WIN: movdqa	(%rdx), %xmm0
 | 
						|
; WIN: pand 	(%r8), %xmm0
 | 
						|
; WIN: pextrq	$1, %xmm0, %r[[REG4:.+]]
 | 
						|
; WIN: movd 	%xmm0, %r[[REG2:.+]]
 | 
						|
; WIN: movslq	%e[[REG2]], %r[[REG1:.+]]
 | 
						|
; WIN: sarq    $32, %r[[REG2]]
 | 
						|
; WIN: movslq	%e[[REG4]], %r[[REG3:.+]]
 | 
						|
; WIN: sarq    $32, %r[[REG4]]
 | 
						|
; WIN: movsd	(%rcx,%r[[REG1]],8), %xmm0
 | 
						|
; WIN: movhpd	(%rcx,%r[[REG2]],8), %xmm0
 | 
						|
; WIN: movsd	(%rcx,%r[[REG3]],8), %xmm1
 | 
						|
; WIN: movhpd	(%rcx,%r[[REG4]],8), %xmm1
 | 
						|
 | 
						|
define <4 x double> @foo(double* %p, <4 x i32>* %i, <4 x i32>* %h) nounwind {
 | 
						|
  %a = load <4 x i32>, <4 x i32>* %i
 | 
						|
  %b = load <4 x i32>, <4 x i32>* %h
 | 
						|
  %j = and <4 x i32> %a, %b
 | 
						|
  %d0 = extractelement <4 x i32> %j, i32 0
 | 
						|
  %d1 = extractelement <4 x i32> %j, i32 1
 | 
						|
  %d2 = extractelement <4 x i32> %j, i32 2
 | 
						|
  %d3 = extractelement <4 x i32> %j, i32 3
 | 
						|
  %q0 = getelementptr double, double* %p, i32 %d0
 | 
						|
  %q1 = getelementptr double, double* %p, i32 %d1
 | 
						|
  %q2 = getelementptr double, double* %p, i32 %d2
 | 
						|
  %q3 = getelementptr double, double* %p, i32 %d3
 | 
						|
  %r0 = load double, double* %q0
 | 
						|
  %r1 = load double, double* %q1
 | 
						|
  %r2 = load double, double* %q2
 | 
						|
  %r3 = load double, double* %q3
 | 
						|
  %v0 = insertelement <4 x double> undef, double %r0, i32 0
 | 
						|
  %v1 = insertelement <4 x double> %v0, double %r1, i32 1
 | 
						|
  %v2 = insertelement <4 x double> %v1, double %r2, i32 2
 | 
						|
  %v3 = insertelement <4 x double> %v2, double %r3, i32 3
 | 
						|
  ret <4 x double> %v3
 | 
						|
}
 | 
						|
 | 
						|
; Check that the sequence previously used above, which bounces the vector off the
 | 
						|
; cache works for x86-32. Note that in this case it will not be used for index
 | 
						|
; calculation, since indexes are 32-bit, not 64.
 | 
						|
; CHECK-LABEL: old:
 | 
						|
; LIN32: movaps	%xmm0, (%esp)
 | 
						|
; LIN32-DAG: {{(mov|and)}}l	(%esp),
 | 
						|
; LIN32-DAG: {{(mov|and)}}l	4(%esp),
 | 
						|
; LIN32-DAG: {{(mov|and)}}l	8(%esp),
 | 
						|
; LIN32-DAG: {{(mov|and)}}l	12(%esp),
 | 
						|
define <4 x i64> @old(double* %p, <4 x i32>* %i, <4 x i32>* %h, i64 %f) nounwind {
 | 
						|
  %a = load <4 x i32>, <4 x i32>* %i
 | 
						|
  %b = load <4 x i32>, <4 x i32>* %h
 | 
						|
  %j = and <4 x i32> %a, %b
 | 
						|
  %d0 = extractelement <4 x i32> %j, i32 0
 | 
						|
  %d1 = extractelement <4 x i32> %j, i32 1
 | 
						|
  %d2 = extractelement <4 x i32> %j, i32 2
 | 
						|
  %d3 = extractelement <4 x i32> %j, i32 3
 | 
						|
  %q0 = zext i32 %d0 to i64
 | 
						|
  %q1 = zext i32 %d1 to i64
 | 
						|
  %q2 = zext i32 %d2 to i64
 | 
						|
  %q3 = zext i32 %d3 to i64  
 | 
						|
  %r0 = and i64 %q0, %f
 | 
						|
  %r1 = and i64 %q1, %f
 | 
						|
  %r2 = and i64 %q2, %f
 | 
						|
  %r3 = and i64 %q3, %f
 | 
						|
  %v0 = insertelement <4 x i64> undef, i64 %r0, i32 0
 | 
						|
  %v1 = insertelement <4 x i64> %v0, i64 %r1, i32 1
 | 
						|
  %v2 = insertelement <4 x i64> %v1, i64 %r2, i32 2
 | 
						|
  %v3 = insertelement <4 x i64> %v2, i64 %r3, i32 3
 | 
						|
  ret <4 x i64> %v3
 | 
						|
}
 |