diff options
author | Aaron Watry <awatry@gmail.com> | 2014-08-20 13:58:57 +0000 |
---|---|---|
committer | Aaron Watry <awatry@gmail.com> | 2014-08-20 13:58:57 +0000 |
commit | f991505d02dae4c2e9cd9ae02372a3116da2a14b (patch) | |
tree | 99138f8a26c76632c9db66bf8622e97c38a6ddda /libclc/generic | |
parent | 474972585a1f4328c9f3138c88365a43c11556c0 (diff) | |
download | bcm5719-llvm-f991505d02dae4c2e9cd9ae02372a3116da2a14b.tar.gz bcm5719-llvm-f991505d02dae4c2e9cd9ae02372a3116da2a14b.zip |
vload/vstore: Use casts instead of scalarizing everything in CLC version
This generates bitcode which is indistinguishable from what was
hand-written for int32 types in v[load|store]_impl.ll.
v4: Use vec2+scalar for vec3 load/stores to prevent corruption (per Tom)
v3: Also remove unused generic/lib/shared/v[load|store]_impl.ll
v2: (Per Matt Arsenault) Fix alignment issues with vector load stores
Signed-off-by: Aaron Watry <awatry@gmail.com>
Reviewed-by: Tom Stellard <thomas.stellard@amd.com>
CC: Matt Arsenault <Matthew.Arsenault@amd.com>
CC: Tom Stellard <thomas.stellard@amd.com>
llvm-svn: 216069
Diffstat (limited to 'libclc/generic')
-rw-r--r-- | libclc/generic/lib/SOURCES | 2 | ||||
-rw-r--r-- | libclc/generic/lib/shared/vload.cl | 16 | ||||
-rw-r--r-- | libclc/generic/lib/shared/vload_impl.ll | 130 | ||||
-rw-r--r-- | libclc/generic/lib/shared/vstore.cl | 21 | ||||
-rw-r--r-- | libclc/generic/lib/shared/vstore_impl.ll | 40 |
5 files changed, 21 insertions, 188 deletions
diff --git a/libclc/generic/lib/SOURCES b/libclc/generic/lib/SOURCES index bfdec7b7cd4..13076616d44 100644 --- a/libclc/generic/lib/SOURCES +++ b/libclc/generic/lib/SOURCES @@ -57,8 +57,6 @@ shared/clamp.cl shared/max.cl shared/min.cl shared/vload.cl -shared/vload_impl.ll shared/vstore.cl -shared/vstore_impl.ll workitem/get_global_id.cl workitem/get_global_size.cl diff --git a/libclc/generic/lib/shared/vload.cl b/libclc/generic/lib/shared/vload.cl index 6793072cc6e..88972005cfa 100644 --- a/libclc/generic/lib/shared/vload.cl +++ b/libclc/generic/lib/shared/vload.cl @@ -1,24 +1,30 @@ #include <clc/clc.h> #define VLOAD_VECTORIZE(PRIM_TYPE, ADDR_SPACE) \ + typedef PRIM_TYPE##2 less_aligned_##ADDR_SPACE##PRIM_TYPE##2 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF PRIM_TYPE##2 vload2(size_t offset, const ADDR_SPACE PRIM_TYPE *x) { \ - return (PRIM_TYPE##2)(x[2*offset] , x[2*offset+1]); \ + return *((const ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##2*) (&x[2*offset])); \ } \ \ + typedef PRIM_TYPE##3 less_aligned_##ADDR_SPACE##PRIM_TYPE##3 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF PRIM_TYPE##3 vload3(size_t offset, const ADDR_SPACE PRIM_TYPE *x) { \ - return (PRIM_TYPE##3)(x[3*offset] , x[3*offset+1], x[3*offset+2]); \ + PRIM_TYPE##2 vec = *((const ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##2*) (&x[3*offset])); \ + return (PRIM_TYPE##3)(vec.s0, vec.s1, x[offset*3+2]); \ } \ \ + typedef PRIM_TYPE##4 less_aligned_##ADDR_SPACE##PRIM_TYPE##4 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF PRIM_TYPE##4 vload4(size_t offset, const ADDR_SPACE PRIM_TYPE *x) { \ - return (PRIM_TYPE##4)(x[4*offset], x[4*offset+1], x[4*offset+2], x[4*offset+3]); \ + return *((const ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##4*) (&x[4*offset])); \ } \ \ + typedef PRIM_TYPE##8 less_aligned_##ADDR_SPACE##PRIM_TYPE##8 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF PRIM_TYPE##8 vload8(size_t offset, const ADDR_SPACE PRIM_TYPE *x) { \ - return (PRIM_TYPE##8)(vload4(0, &x[8*offset]), vload4(1, &x[8*offset])); \ + return *((const ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##8*) (&x[8*offset])); \ } \ \ + typedef PRIM_TYPE##16 less_aligned_##ADDR_SPACE##PRIM_TYPE##16 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF PRIM_TYPE##16 vload16(size_t offset, const ADDR_SPACE PRIM_TYPE *x) { \ - return (PRIM_TYPE##16)(vload8(0, &x[16*offset]), vload8(1, &x[16*offset])); \ + return *((const ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##16*) (&x[16*offset])); \ } \ #define VLOAD_ADDR_SPACES(__CLC_SCALAR_GENTYPE) \ diff --git a/libclc/generic/lib/shared/vload_impl.ll b/libclc/generic/lib/shared/vload_impl.ll deleted file mode 100644 index 33ba996b25c..00000000000 --- a/libclc/generic/lib/shared/vload_impl.ll +++ /dev/null @@ -1,130 +0,0 @@ -; This provides optimized implementations of vload2/3/4/8/16 for 32-bit int/uint -; The address spaces get mapped to data types in target-specific usages - -define <2 x i32> @__clc_vload2_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <2 x i32> addrspace(1)* - %2 = load <2 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret <2 x i32> %2 -} - -define <3 x i32> @__clc_vload3_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <3 x i32> addrspace(1)* - %2 = load <3 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret <3 x i32> %2 -} - -define <4 x i32> @__clc_vload4_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <4 x i32> addrspace(1)* - %2 = load <4 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret <4 x i32> %2 -} - -define <8 x i32> @__clc_vload8_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <8 x i32> addrspace(1)* - %2 = load <8 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret <8 x i32> %2 -} - -define <16 x i32> @__clc_vload16_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <16 x i32> addrspace(1)* - %2 = load <16 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret <16 x i32> %2 -} - -define <2 x i32> @__clc_vload2_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(2)* %addr to <2 x i32> addrspace(2)* - %2 = load <2 x i32> addrspace(2)* %1, align 4, !tbaa !3 - ret <2 x i32> %2 -} - -define <3 x i32> @__clc_vload3_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(2)* %addr to <3 x i32> addrspace(2)* - %2 = load <3 x i32> addrspace(2)* %1, align 4, !tbaa !3 - ret <3 x i32> %2 -} - -define <4 x i32> @__clc_vload4_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(2)* %addr to <4 x i32> addrspace(2)* - %2 = load <4 x i32> addrspace(2)* %1, align 4, !tbaa !3 - ret <4 x i32> %2 -} - -define <8 x i32> @__clc_vload8_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(2)* %addr to <8 x i32> addrspace(2)* - %2 = load <8 x i32> addrspace(2)* %1, align 4, !tbaa !3 - ret <8 x i32> %2 -} - -define <16 x i32> @__clc_vload16_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(2)* %addr to <16 x i32> addrspace(2)* - %2 = load <16 x i32> addrspace(2)* %1, align 4, !tbaa !3 - ret <16 x i32> %2 -} - -define <2 x i32> @__clc_vload2_i32__addr3(i32 addrspace(3)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(3)* %addr to <2 x i32> addrspace(3)* - %2 = load <2 x i32> addrspace(3)* %1, align 4, !tbaa !3 - ret <2 x i32> %2 -} - -define <3 x i32> @__clc_vload3_i32__addr3(i32 addrspace(3)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(3)* %addr to <3 x i32> addrspace(3)* - %2 = load <3 x i32> addrspace(3)* %1, align 4, !tbaa !3 - ret <3 x i32> %2 -} - -define <4 x i32> @__clc_vload4_i32__addr3(i32 addrspace(3)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(3)* %addr to <4 x i32> addrspace(3)* - %2 = load <4 x i32> addrspace(3)* %1, align 4, !tbaa !3 - ret <4 x i32> %2 -} - -define <8 x i32> @__clc_vload8_i32__addr3(i32 addrspace(3)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(3)* %addr to <8 x i32> addrspace(3)* - %2 = load <8 x i32> addrspace(3)* %1, align 4, !tbaa !3 - ret <8 x i32> %2 -} - -define <16 x i32> @__clc_vload16_i32__addr3(i32 addrspace(3)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(3)* %addr to <16 x i32> addrspace(3)* - %2 = load <16 x i32> addrspace(3)* %1, align 4, !tbaa !3 - ret <16 x i32> %2 -} - -define <2 x i32> @__clc_vload2_i32__addr4(i32 addrspace(4)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(4)* %addr to <2 x i32> addrspace(4)* - %2 = load <2 x i32> addrspace(4)* %1, align 4, !tbaa !3 - ret <2 x i32> %2 -} - -define <3 x i32> @__clc_vload3_i32__addr4(i32 addrspace(4)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(4)* %addr to <3 x i32> addrspace(4)* - %2 = load <3 x i32> addrspace(4)* %1, align 4, !tbaa !3 - ret <3 x i32> %2 -} - -define <4 x i32> @__clc_vload4_i32__addr4(i32 addrspace(4)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(4)* %addr to <4 x i32> addrspace(4)* - %2 = load <4 x i32> addrspace(4)* %1, align 4, !tbaa !3 - ret <4 x i32> %2 -} - -define <8 x i32> @__clc_vload8_i32__addr4(i32 addrspace(4)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(4)* %addr to <8 x i32> addrspace(4)* - %2 = load <8 x i32> addrspace(4)* %1, align 4, !tbaa !3 - ret <8 x i32> %2 -} - -define <16 x i32> @__clc_vload16_i32__addr4(i32 addrspace(4)* nocapture %addr) nounwind readonly alwaysinline { - %1 = bitcast i32 addrspace(4)* %addr to <16 x i32> addrspace(4)* - %2 = load <16 x i32> addrspace(4)* %1, align 4, !tbaa !3 - ret <16 x i32> %2 -} - -!1 = metadata !{metadata !"char", metadata !5} -!2 = metadata !{metadata !"short", metadata !5} -!3 = metadata !{metadata !"int", metadata !5} -!4 = metadata !{metadata !"long", metadata !5} -!5 = metadata !{metadata !"omnipotent char", metadata !6} -!6 = metadata !{metadata !"Simple C/C++ TBAA"} - diff --git a/libclc/generic/lib/shared/vstore.cl b/libclc/generic/lib/shared/vstore.cl index f6d360e0d1d..4777b7ea76a 100644 --- a/libclc/generic/lib/shared/vstore.cl +++ b/libclc/generic/lib/shared/vstore.cl @@ -3,30 +3,29 @@ #pragma OPENCL EXTENSION cl_khr_byte_addressable_store : enable #define VSTORE_VECTORIZE(PRIM_TYPE, ADDR_SPACE) \ + typedef PRIM_TYPE##2 less_aligned_##ADDR_SPACE##PRIM_TYPE##2 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF void vstore2(PRIM_TYPE##2 vec, size_t offset, ADDR_SPACE PRIM_TYPE *mem) { \ - mem[2*offset] = vec.s0; \ - mem[2*offset+1] = vec.s1; \ + *((ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##2*) (&mem[2*offset])) = vec; \ } \ \ _CLC_OVERLOAD _CLC_DEF void vstore3(PRIM_TYPE##3 vec, size_t offset, ADDR_SPACE PRIM_TYPE *mem) { \ - mem[3*offset] = vec.s0; \ - mem[3*offset+1] = vec.s1; \ - mem[3*offset+2] = vec.s2; \ + *((ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##2*) (&mem[3*offset])) = (PRIM_TYPE##2)(vec.s0, vec.s1); \ + mem[3 * offset + 2] = vec.s2;\ } \ \ + typedef PRIM_TYPE##4 less_aligned_##ADDR_SPACE##PRIM_TYPE##4 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF void vstore4(PRIM_TYPE##4 vec, size_t offset, ADDR_SPACE PRIM_TYPE *mem) { \ - vstore2(vec.lo, 0, &mem[offset*4]); \ - vstore2(vec.hi, 1, &mem[offset*4]); \ + *((ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##4*) (&mem[4*offset])) = vec; \ } \ \ + typedef PRIM_TYPE##8 less_aligned_##ADDR_SPACE##PRIM_TYPE##8 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF void vstore8(PRIM_TYPE##8 vec, size_t offset, ADDR_SPACE PRIM_TYPE *mem) { \ - vstore4(vec.lo, 0, &mem[offset*8]); \ - vstore4(vec.hi, 1, &mem[offset*8]); \ + *((ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##8*) (&mem[8*offset])) = vec; \ } \ \ + typedef PRIM_TYPE##16 less_aligned_##ADDR_SPACE##PRIM_TYPE##16 __attribute__ ((aligned (sizeof(PRIM_TYPE))));\ _CLC_OVERLOAD _CLC_DEF void vstore16(PRIM_TYPE##16 vec, size_t offset, ADDR_SPACE PRIM_TYPE *mem) { \ - vstore8(vec.lo, 0, &mem[offset*16]); \ - vstore8(vec.hi, 1, &mem[offset*16]); \ + *((ADDR_SPACE less_aligned_##ADDR_SPACE##PRIM_TYPE##16*) (&mem[16*offset])) = vec; \ } \ #define VSTORE_ADDR_SPACES(__CLC_SCALAR___CLC_GENTYPE) \ diff --git a/libclc/generic/lib/shared/vstore_impl.ll b/libclc/generic/lib/shared/vstore_impl.ll deleted file mode 100644 index 9e2a37b2970..00000000000 --- a/libclc/generic/lib/shared/vstore_impl.ll +++ /dev/null @@ -1,40 +0,0 @@ -; This provides optimized implementations of vstore2/3/4/8/16 for 32-bit int/uint -; The address spaces get mapped to data types in target-specific usages - -define void @__clc_vstore2_i32__addr1(<2 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <2 x i32> addrspace(1)* - store <2 x i32> %vec, <2 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret void -} - -define void @__clc_vstore3_i32__addr1(<3 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <3 x i32> addrspace(1)* - store <3 x i32> %vec, <3 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret void -} - -define void @__clc_vstore4_i32__addr1(<4 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <4 x i32> addrspace(1)* - store <4 x i32> %vec, <4 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret void -} - -define void @__clc_vstore8_i32__addr1(<8 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <8 x i32> addrspace(1)* - store <8 x i32> %vec, <8 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret void -} - -define void @__clc_vstore16_i32__addr1(<16 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline { - %1 = bitcast i32 addrspace(1)* %addr to <16 x i32> addrspace(1)* - store <16 x i32> %vec, <16 x i32> addrspace(1)* %1, align 4, !tbaa !3 - ret void -} - -!1 = metadata !{metadata !"char", metadata !5} -!2 = metadata !{metadata !"short", metadata !5} -!3 = metadata !{metadata !"int", metadata !5} -!4 = metadata !{metadata !"long", metadata !5} -!5 = metadata !{metadata !"omnipotent char", metadata !6} -!6 = metadata !{metadata !"Simple C/C++ TBAA"} - |