summaryrefslogtreecommitdiffstats
path: root/llvm/test
diff options
context:
space:
mode:
authorSanjay Patel <spatel@rotateright.com>2016-05-28 14:58:37 +0000
committerSanjay Patel <spatel@rotateright.com>2016-05-28 14:58:37 +0000
commit97c2c108fd23a377ca21f89d25dc021ffa884b19 (patch)
treebba70d79cf628687cf784a8239e4034117f78314 /llvm/test
parenta3dc1ba142c5c1f8e4b53e501597904c23e5c61f (diff)
downloadbcm5719-llvm-97c2c108fd23a377ca21f89d25dc021ffa884b19.tar.gz
bcm5719-llvm-97c2c108fd23a377ca21f89d25dc021ffa884b19.zip
[x86] avoid printing unnecessary sign bits of hex immediates in asm comments (PR20347)
It would be better to check the valid/expected size of the immediate operand, but this is generally better than what we print right now. Differential Revision: http://reviews.llvm.org/D20385 llvm-svn: 271114
Diffstat (limited to 'llvm/test')
-rw-r--r--llvm/test/CodeGen/X86/2011-10-21-widen-cmp.ll2
-rw-r--r--llvm/test/CodeGen/X86/avx512-cmp.ll2
-rw-r--r--llvm/test/CodeGen/X86/bitreverse.ll2
-rw-r--r--llvm/test/CodeGen/X86/pr16360.ll2
-rw-r--r--llvm/test/CodeGen/X86/rem.ll4
-rw-r--r--llvm/test/CodeGen/X86/sad.ll32
-rw-r--r--llvm/test/CodeGen/X86/sse-intrinsics-fast-isel.ll12
-rw-r--r--llvm/test/CodeGen/X86/vec_fneg.ll4
-rw-r--r--llvm/test/CodeGen/X86/vector-bitreverse.ll4
-rw-r--r--llvm/test/CodeGen/X86/vector-idiv-sdiv-512.ll64
-rw-r--r--llvm/test/MC/X86/imm-comments.s4
11 files changed, 66 insertions, 66 deletions
diff --git a/llvm/test/CodeGen/X86/2011-10-21-widen-cmp.ll b/llvm/test/CodeGen/X86/2011-10-21-widen-cmp.ll
index cb4648c382f..420e843b52a 100644
--- a/llvm/test/CodeGen/X86/2011-10-21-widen-cmp.ll
+++ b/llvm/test/CodeGen/X86/2011-10-21-widen-cmp.ll
@@ -42,7 +42,7 @@ entry:
define void @mp_11193(<8 x float> * nocapture %aFOO, <8 x float>* nocapture %RET) nounwind {
; CHECK-LABEL: mp_11193:
; CHECK: # BB#0: # %allocas
-; CHECK-NEXT: movl $-1082130432, (%rsi) # imm = 0xFFFFFFFFBF800000
+; CHECK-NEXT: movl $-1082130432, (%rsi) # imm = 0xBF800000
; CHECK-NEXT: retq
allocas:
%bincmp = fcmp olt <8 x float> <float 1.000000e+00, float 1.000000e+00, float 1.000000e+00, float 1.000000e+00, float 9.000000e+00, float 1.000000e+00, float 9.000000e+00, float 1.000000e+00> , <float 3.000000e+00, float 3.000000e+00, float 3.000000e+00, float 3.000000e+00, float 3.000000e+00, float 3.000000e+00, float 3.000000e+00, float 3.000000e+00>
diff --git a/llvm/test/CodeGen/X86/avx512-cmp.ll b/llvm/test/CodeGen/X86/avx512-cmp.ll
index 68989b1bcea..2c0c0a5b8c7 100644
--- a/llvm/test/CodeGen/X86/avx512-cmp.ll
+++ b/llvm/test/CodeGen/X86/avx512-cmp.ll
@@ -122,7 +122,7 @@ define i32 @test8(i32 %a1, i32 %a2, i32 %a3) {
; ALL-NEXT: testl %edx, %edx
; ALL-NEXT: movl $1, %eax
; ALL-NEXT: cmovel %eax, %edx
-; ALL-NEXT: cmpl $-2147483648, %esi ## imm = 0xFFFFFFFF80000000
+; ALL-NEXT: cmpl $-2147483648, %esi ## imm = 0x80000000
; ALL-NEXT: cmovnel %edx, %eax
; ALL-NEXT: cmpl $-1, %edi
; ALL-NEXT: cmovnel %edx, %eax
diff --git a/llvm/test/CodeGen/X86/bitreverse.ll b/llvm/test/CodeGen/X86/bitreverse.ll
index 74db6eaea68..62ac8623720 100644
--- a/llvm/test/CodeGen/X86/bitreverse.ll
+++ b/llvm/test/CodeGen/X86/bitreverse.ll
@@ -26,7 +26,7 @@ define i8 @g(i8 %a) {
define <2 x i16> @fold_v2i16() {
; CHECK-LABEL: fold_v2i16:
; CHECK: # BB#0:
-; CHECK-NEXT: movw $-4096, %ax # imm = 0xFFFFFFFFFFFFF000
+; CHECK-NEXT: movw $-4096, %ax # imm = 0xF000
; CHECK-NEXT: movw $240, %dx
; CHECK-NEXT: retl
%b = call <2 x i16> @llvm.bitreverse.v2i16(<2 x i16> <i16 15, i16 3840>)
diff --git a/llvm/test/CodeGen/X86/pr16360.ll b/llvm/test/CodeGen/X86/pr16360.ll
index 984cc17547c..0d2878dc6af 100644
--- a/llvm/test/CodeGen/X86/pr16360.ll
+++ b/llvm/test/CodeGen/X86/pr16360.ll
@@ -6,7 +6,7 @@ define i64 @foo(i32 %sum) {
; CHECK: # BB#0: # %entry
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
; CHECK-NEXT: shrl $2, %eax
-; CHECK-NEXT: orl $-67108864, %eax # imm = 0xFFFFFFFFFC000000
+; CHECK-NEXT: orl $-67108864, %eax # imm = 0xFC000000
; CHECK-NEXT: movl $1073741823, %edx # imm = 0x3FFFFFFF
; CHECK-NEXT: retl
entry:
diff --git a/llvm/test/CodeGen/X86/rem.ll b/llvm/test/CodeGen/X86/rem.ll
index ffa24384c60..cc591e5ac00 100644
--- a/llvm/test/CodeGen/X86/rem.ll
+++ b/llvm/test/CodeGen/X86/rem.ll
@@ -5,7 +5,7 @@ define i32 @test1(i32 %X) {
; CHECK-LABEL: test1:
; CHECK: # BB#0:
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
-; CHECK-NEXT: movl $-2139062143, %edx # imm = 0xFFFFFFFF80808081
+; CHECK-NEXT: movl $-2139062143, %edx # imm = 0x80808081
; CHECK-NEXT: movl %ecx, %eax
; CHECK-NEXT: imull %edx
; CHECK-NEXT: addl %ecx, %edx
@@ -44,7 +44,7 @@ define i32 @test3(i32 %X) {
; CHECK-LABEL: test3:
; CHECK: # BB#0:
; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
-; CHECK-NEXT: movl $-2139062143, %edx # imm = 0xFFFFFFFF80808081
+; CHECK-NEXT: movl $-2139062143, %edx # imm = 0x80808081
; CHECK-NEXT: movl %ecx, %eax
; CHECK-NEXT: mull %edx
; CHECK-NEXT: shrl $7, %edx
diff --git a/llvm/test/CodeGen/X86/sad.ll b/llvm/test/CodeGen/X86/sad.ll
index 24c7fd9cfb1..843699ae887 100644
--- a/llvm/test/CodeGen/X86/sad.ll
+++ b/llvm/test/CodeGen/X86/sad.ll
@@ -16,7 +16,7 @@ define i32 @sad_16i8() nounwind {
; SSE2-NEXT: andq $-64, %rsp
; SSE2-NEXT: subq $128, %rsp
; SSE2-NEXT: pxor %xmm0, %xmm0
-; SSE2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; SSE2-NEXT: movq $-1024, %rax # imm = 0xFC00
; SSE2-NEXT: pxor %xmm1, %xmm1
; SSE2-NEXT: pxor %xmm3, %xmm3
; SSE2-NEXT: pxor %xmm2, %xmm2
@@ -58,7 +58,7 @@ define i32 @sad_16i8() nounwind {
; AVX2-NEXT: andq $-64, %rsp
; AVX2-NEXT: subq $128, %rsp
; AVX2-NEXT: vpxor %ymm0, %ymm0, %ymm0
-; AVX2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX2-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX2-NEXT: vpxor %ymm1, %ymm1, %ymm1
; AVX2-NEXT: .p2align 4, 0x90
; AVX2-NEXT: .LBB0_1: # %vector.body
@@ -89,7 +89,7 @@ define i32 @sad_16i8() nounwind {
; AVX512F-LABEL: sad_16i8:
; AVX512F: # BB#0: # %entry
; AVX512F-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512F-NEXT: .p2align 4, 0x90
; AVX512F-NEXT: .LBB0_1: # %vector.body
; AVX512F-NEXT: # =>This Inner Loop Header: Depth=1
@@ -116,7 +116,7 @@ define i32 @sad_16i8() nounwind {
; AVX512BW-LABEL: sad_16i8:
; AVX512BW: # BB#0: # %entry
; AVX512BW-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512BW-NEXT: .p2align 4, 0x90
; AVX512BW-NEXT: .LBB0_1: # %vector.body
; AVX512BW-NEXT: # =>This Inner Loop Header: Depth=1
@@ -180,7 +180,7 @@ define i32 @sad_32i8() nounwind {
; SSE2-LABEL: sad_32i8:
; SSE2: # BB#0: # %entry
; SSE2-NEXT: pxor %xmm12, %xmm12
-; SSE2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; SSE2-NEXT: movq $-1024, %rax # imm = 0xFC00
; SSE2-NEXT: pxor %xmm4, %xmm4
; SSE2-NEXT: pxor %xmm2, %xmm2
; SSE2-NEXT: pxor %xmm0, %xmm0
@@ -316,7 +316,7 @@ define i32 @sad_32i8() nounwind {
; AVX2-NEXT: andq $-128, %rsp
; AVX2-NEXT: subq $256, %rsp # imm = 0x100
; AVX2-NEXT: vpxor %ymm0, %ymm0, %ymm0
-; AVX2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX2-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX2-NEXT: vpxor %ymm1, %ymm1, %ymm1
; AVX2-NEXT: vpxor %ymm2, %ymm2, %ymm2
; AVX2-NEXT: vpxor %ymm3, %ymm3, %ymm3
@@ -358,7 +358,7 @@ define i32 @sad_32i8() nounwind {
; AVX512F-NEXT: andq $-128, %rsp
; AVX512F-NEXT: subq $256, %rsp # imm = 0x100
; AVX512F-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512F-NEXT: vpxord %zmm1, %zmm1, %zmm1
; AVX512F-NEXT: .p2align 4, 0x90
; AVX512F-NEXT: .LBB1_1: # %vector.body
@@ -397,7 +397,7 @@ define i32 @sad_32i8() nounwind {
; AVX512BW-NEXT: andq $-128, %rsp
; AVX512BW-NEXT: subq $256, %rsp # imm = 0x100
; AVX512BW-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512BW-NEXT: vpxord %zmm1, %zmm1, %zmm1
; AVX512BW-NEXT: .p2align 4, 0x90
; AVX512BW-NEXT: .LBB1_1: # %vector.body
@@ -472,7 +472,7 @@ define i32 @sad_avx64i8() nounwind {
; SSE2: # BB#0: # %entry
; SSE2-NEXT: subq $232, %rsp
; SSE2-NEXT: pxor %xmm8, %xmm8
-; SSE2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; SSE2-NEXT: movq $-1024, %rax # imm = 0xFC00
; SSE2-NEXT: pxor %xmm5, %xmm5
; SSE2-NEXT: pxor %xmm2, %xmm2
; SSE2-NEXT: pxor %xmm1, %xmm1
@@ -764,7 +764,7 @@ define i32 @sad_avx64i8() nounwind {
; AVX2-LABEL: sad_avx64i8:
; AVX2: # BB#0: # %entry
; AVX2-NEXT: vpxor %ymm0, %ymm0, %ymm0
-; AVX2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX2-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX2-NEXT: vpxor %ymm2, %ymm2, %ymm2
; AVX2-NEXT: vpxor %ymm1, %ymm1, %ymm1
; AVX2-NEXT: vpxor %ymm3, %ymm3, %ymm3
@@ -840,7 +840,7 @@ define i32 @sad_avx64i8() nounwind {
; AVX512F-LABEL: sad_avx64i8:
; AVX512F: # BB#0: # %entry
; AVX512F-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512F-NEXT: vpxord %zmm1, %zmm1, %zmm1
; AVX512F-NEXT: vpxord %zmm2, %zmm2, %zmm2
; AVX512F-NEXT: vpxord %zmm3, %zmm3, %zmm3
@@ -893,7 +893,7 @@ define i32 @sad_avx64i8() nounwind {
; AVX512BW-NEXT: andq $-256, %rsp
; AVX512BW-NEXT: subq $512, %rsp # imm = 0x200
; AVX512BW-NEXT: vpxord %zmm0, %zmm0, %zmm0
-; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512BW-NEXT: vpxord %zmm2, %zmm2, %zmm2
; AVX512BW-NEXT: vpxord %zmm3, %zmm3, %zmm3
; AVX512BW-NEXT: vpxord %zmm1, %zmm1, %zmm1
@@ -976,7 +976,7 @@ define i32 @sad_2i8() nounwind {
; SSE2-LABEL: sad_2i8:
; SSE2: # BB#0: # %entry
; SSE2-NEXT: pxor %xmm0, %xmm0
-; SSE2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; SSE2-NEXT: movq $-1024, %rax # imm = 0xFC00
; SSE2-NEXT: movl $65535, %ecx # imm = 0xFFFF
; SSE2-NEXT: movd %ecx, %xmm1
; SSE2-NEXT: .p2align 4, 0x90
@@ -999,7 +999,7 @@ define i32 @sad_2i8() nounwind {
; AVX2-LABEL: sad_2i8:
; AVX2: # BB#0: # %entry
; AVX2-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; AVX2-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX2-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX2-NEXT: vpxor %xmm1, %xmm1, %xmm1
; AVX2-NEXT: .p2align 4, 0x90
; AVX2-NEXT: .LBB3_1: # %vector.body
@@ -1021,7 +1021,7 @@ define i32 @sad_2i8() nounwind {
; AVX512F-LABEL: sad_2i8:
; AVX512F: # BB#0: # %entry
; AVX512F-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512F-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512F-NEXT: vpxor %xmm1, %xmm1, %xmm1
; AVX512F-NEXT: .p2align 4, 0x90
; AVX512F-NEXT: .LBB3_1: # %vector.body
@@ -1043,7 +1043,7 @@ define i32 @sad_2i8() nounwind {
; AVX512BW-LABEL: sad_2i8:
; AVX512BW: # BB#0: # %entry
; AVX512BW-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFFFFFFFFFFFFFC00
+; AVX512BW-NEXT: movq $-1024, %rax # imm = 0xFC00
; AVX512BW-NEXT: vpxor %xmm1, %xmm1, %xmm1
; AVX512BW-NEXT: .p2align 4, 0x90
; AVX512BW-NEXT: .LBB3_1: # %vector.body
diff --git a/llvm/test/CodeGen/X86/sse-intrinsics-fast-isel.ll b/llvm/test/CodeGen/X86/sse-intrinsics-fast-isel.ll
index a20feaf5acd..a762fa07fcd 100644
--- a/llvm/test/CodeGen/X86/sse-intrinsics-fast-isel.ll
+++ b/llvm/test/CodeGen/X86/sse-intrinsics-fast-isel.ll
@@ -1411,7 +1411,7 @@ define void @test_MM_SET_EXCEPTION_MASK(i32 %a0) nounwind {
; X32-NEXT: leal (%esp), %ecx
; X32-NEXT: stmxcsr (%ecx)
; X32-NEXT: movl (%esp), %edx
-; X32-NEXT: andl $-8065, %edx # imm = 0xFFFFFFFFFFFFE07F
+; X32-NEXT: andl $-8065, %edx # imm = 0xE07F
; X32-NEXT: orl %eax, %edx
; X32-NEXT: movl %edx, (%esp)
; X32-NEXT: ldmxcsr (%ecx)
@@ -1423,7 +1423,7 @@ define void @test_MM_SET_EXCEPTION_MASK(i32 %a0) nounwind {
; X64-NEXT: leaq -{{[0-9]+}}(%rsp), %rax
; X64-NEXT: stmxcsr (%rax)
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %ecx
-; X64-NEXT: andl $-8065, %ecx # imm = 0xFFFFFFFFFFFFE07F
+; X64-NEXT: andl $-8065, %ecx # imm = 0xE07F
; X64-NEXT: orl %edi, %ecx
; X64-NEXT: movl %ecx, -{{[0-9]+}}(%rsp)
; X64-NEXT: ldmxcsr (%rax)
@@ -1484,7 +1484,7 @@ define void @test_MM_SET_FLUSH_ZERO_MODE(i32 %a0) nounwind {
; X32-NEXT: leal (%esp), %ecx
; X32-NEXT: stmxcsr (%ecx)
; X32-NEXT: movl (%esp), %edx
-; X32-NEXT: andl $-32769, %edx # imm = 0xFFFFFFFFFFFF7FFF
+; X32-NEXT: andl $-32769, %edx # imm = 0xFFFF7FFF
; X32-NEXT: orl %eax, %edx
; X32-NEXT: movl %edx, (%esp)
; X32-NEXT: ldmxcsr (%ecx)
@@ -1496,7 +1496,7 @@ define void @test_MM_SET_FLUSH_ZERO_MODE(i32 %a0) nounwind {
; X64-NEXT: leaq -{{[0-9]+}}(%rsp), %rax
; X64-NEXT: stmxcsr (%rax)
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %ecx
-; X64-NEXT: andl $-32769, %ecx # imm = 0xFFFFFFFFFFFF7FFF
+; X64-NEXT: andl $-32769, %ecx # imm = 0xFFFF7FFF
; X64-NEXT: orl %edi, %ecx
; X64-NEXT: movl %ecx, -{{[0-9]+}}(%rsp)
; X64-NEXT: ldmxcsr (%rax)
@@ -1564,7 +1564,7 @@ define void @test_MM_SET_ROUNDING_MODE(i32 %a0) nounwind {
; X32-NEXT: leal (%esp), %ecx
; X32-NEXT: stmxcsr (%ecx)
; X32-NEXT: movl (%esp), %edx
-; X32-NEXT: andl $-24577, %edx # imm = 0xFFFFFFFFFFFF9FFF
+; X32-NEXT: andl $-24577, %edx # imm = 0x9FFF
; X32-NEXT: orl %eax, %edx
; X32-NEXT: movl %edx, (%esp)
; X32-NEXT: ldmxcsr (%ecx)
@@ -1576,7 +1576,7 @@ define void @test_MM_SET_ROUNDING_MODE(i32 %a0) nounwind {
; X64-NEXT: leaq -{{[0-9]+}}(%rsp), %rax
; X64-NEXT: stmxcsr (%rax)
; X64-NEXT: movl -{{[0-9]+}}(%rsp), %ecx
-; X64-NEXT: andl $-24577, %ecx # imm = 0xFFFFFFFFFFFF9FFF
+; X64-NEXT: andl $-24577, %ecx # imm = 0x9FFF
; X64-NEXT: orl %edi, %ecx
; X64-NEXT: movl %ecx, -{{[0-9]+}}(%rsp)
; X64-NEXT: ldmxcsr (%rax)
diff --git a/llvm/test/CodeGen/X86/vec_fneg.ll b/llvm/test/CodeGen/X86/vec_fneg.ll
index bc83de5d38c..2dac7f25b40 100644
--- a/llvm/test/CodeGen/X86/vec_fneg.ll
+++ b/llvm/test/CodeGen/X86/vec_fneg.ll
@@ -58,7 +58,7 @@ define <2 x float> @fneg_bitcast(i64 %i) nounwind {
; X32-SSE1-NEXT: movl %esp, %ebp
; X32-SSE1-NEXT: andl $-16, %esp
; X32-SSE1-NEXT: subl $32, %esp
-; X32-SSE1-NEXT: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000
+; X32-SSE1-NEXT: movl $-2147483648, %eax # imm = 0x80000000
; X32-SSE1-NEXT: movl 12(%ebp), %ecx
; X32-SSE1-NEXT: xorl %eax, %ecx
; X32-SSE1-NEXT: movl %ecx, {{[0-9]+}}(%esp)
@@ -71,7 +71,7 @@ define <2 x float> @fneg_bitcast(i64 %i) nounwind {
;
; X32-SSE2-LABEL: fneg_bitcast:
; X32-SSE2: # BB#0:
-; X32-SSE2-NEXT: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000
+; X32-SSE2-NEXT: movl $-2147483648, %eax # imm = 0x80000000
; X32-SSE2-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X32-SSE2-NEXT: xorl %eax, %ecx
; X32-SSE2-NEXT: xorl {{[0-9]+}}(%esp), %eax
diff --git a/llvm/test/CodeGen/X86/vector-bitreverse.ll b/llvm/test/CodeGen/X86/vector-bitreverse.ll
index f33621ea26d..7f116be994e 100644
--- a/llvm/test/CodeGen/X86/vector-bitreverse.ll
+++ b/llvm/test/CodeGen/X86/vector-bitreverse.ll
@@ -615,7 +615,7 @@ define i64 @test_bitreverse_i64(i64 %a) nounwind {
; SSE-NEXT: orq %rcx, %rdx
; SSE-NEXT: movq %rdi, %rcx
; SSE-NEXT: shrq %rcx
-; SSE-NEXT: andl $-2147483648, %ecx # imm = 0xFFFFFFFF80000000
+; SSE-NEXT: andl $-2147483648, %ecx # imm = 0x80000000
; SSE-NEXT: orq %rdx, %rcx
; SSE-NEXT: movq %rdi, %rdx
; SSE-NEXT: shrq $3, %rdx
@@ -871,7 +871,7 @@ define i64 @test_bitreverse_i64(i64 %a) nounwind {
; AVX-NEXT: orq %rcx, %rdx
; AVX-NEXT: movq %rdi, %rcx
; AVX-NEXT: shrq %rcx
-; AVX-NEXT: andl $-2147483648, %ecx # imm = 0xFFFFFFFF80000000
+; AVX-NEXT: andl $-2147483648, %ecx # imm = 0x80000000
; AVX-NEXT: orq %rdx, %rcx
; AVX-NEXT: movq %rdi, %rdx
; AVX-NEXT: shrq $3, %rdx
diff --git a/llvm/test/CodeGen/X86/vector-idiv-sdiv-512.ll b/llvm/test/CodeGen/X86/vector-idiv-sdiv-512.ll
index f91babe644c..7aadeaa771e 100644
--- a/llvm/test/CodeGen/X86/vector-idiv-sdiv-512.ll
+++ b/llvm/test/CodeGen/X86/vector-idiv-sdiv-512.ll
@@ -87,7 +87,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $3, %zmm0, %xmm1
; AVX-NEXT: vpextrd $1, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -96,7 +96,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: vmovd %xmm1, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %edx, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -107,7 +107,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm2, %xmm2
; AVX-NEXT: vpextrd $2, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -117,7 +117,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm2, %xmm2
; AVX-NEXT: vpextrd $3, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -128,7 +128,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $2, %zmm0, %xmm2
; AVX-NEXT: vpextrd $1, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -137,7 +137,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: vmovd %xmm2, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %edx, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -148,7 +148,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -158,7 +158,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -170,7 +170,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $1, %zmm0, %xmm2
; AVX-NEXT: vpextrd $1, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -179,7 +179,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: vmovd %xmm2, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %edx, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -190,7 +190,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -200,7 +200,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -210,7 +210,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $3, %eax, %xmm3, %xmm2
; AVX-NEXT: vpextrd $1, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -219,7 +219,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: vmovd %xmm0, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %edx, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -230,7 +230,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -240,7 +240,7 @@ define <16 x i32> @test_div7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %ecx, %eax
; AVX-NEXT: movl %eax, %ecx
@@ -1162,7 +1162,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $3, %zmm0, %xmm1
; AVX-NEXT: vpextrd $1, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1174,7 +1174,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: subl %edx, %eax
; AVX-NEXT: vmovd %xmm1, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %ecx, %edx
; AVX-NEXT: movl %edx, %esi
@@ -1188,7 +1188,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm2, %xmm2
; AVX-NEXT: vpextrd $2, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1201,7 +1201,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm2, %xmm2
; AVX-NEXT: vpextrd $3, %xmm1, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1215,7 +1215,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $2, %zmm0, %xmm2
; AVX-NEXT: vpextrd $1, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1227,7 +1227,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: subl %edx, %eax
; AVX-NEXT: vmovd %xmm2, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %ecx, %edx
; AVX-NEXT: movl %edx, %esi
@@ -1241,7 +1241,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1254,7 +1254,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1269,7 +1269,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vextracti32x4 $1, %zmm0, %xmm2
; AVX-NEXT: vpextrd $1, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1281,7 +1281,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: subl %edx, %eax
; AVX-NEXT: vmovd %xmm2, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %ecx, %edx
; AVX-NEXT: movl %edx, %esi
@@ -1295,7 +1295,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1308,7 +1308,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm2, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1321,7 +1321,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $3, %eax, %xmm3, %xmm2
; AVX-NEXT: vpextrd $1, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1333,7 +1333,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: subl %edx, %eax
; AVX-NEXT: vmovd %xmm0, %ecx
; AVX-NEXT: movslq %ecx, %rcx
-; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rcx, %rdx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rdx
; AVX-NEXT: addl %ecx, %edx
; AVX-NEXT: movl %edx, %esi
@@ -1347,7 +1347,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $1, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $2, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
@@ -1360,7 +1360,7 @@ define <16 x i32> @test_rem7_16i32(<16 x i32> %a) nounwind {
; AVX-NEXT: vpinsrd $2, %eax, %xmm3, %xmm3
; AVX-NEXT: vpextrd $3, %xmm0, %eax
; AVX-NEXT: cltq
-; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0xFFFFFFFF92492493
+; AVX-NEXT: imulq $-1840700269, %rax, %rcx # imm = 0x92492493
; AVX-NEXT: shrq $32, %rcx
; AVX-NEXT: addl %eax, %ecx
; AVX-NEXT: movl %ecx, %edx
diff --git a/llvm/test/MC/X86/imm-comments.s b/llvm/test/MC/X86/imm-comments.s
index b176d72885b..844bc39ac22 100644
--- a/llvm/test/MC/X86/imm-comments.s
+++ b/llvm/test/MC/X86/imm-comments.s
@@ -18,10 +18,10 @@ movabsq $-9223372036854775808, %rax
# CHECK: movb $-128, %al
# CHECK: movw $32767, %ax # imm = 0x7FFF
-# CHECK: movw $-32768, %ax # imm = 0xFFFFFFFFFFFF8000
+# CHECK: movw $-32768, %ax # imm = 0x8000
# CHECK: movl $2147483647, %eax # imm = 0x7FFFFFFF
-# CHECK: movl $-2147483648, %eax # imm = 0xFFFFFFFF80000000
+# CHECK: movl $-2147483648, %eax # imm = 0x80000000
# CHECK: movabsq $9223372036854775807, %rax # imm = 0x7FFFFFFFFFFFFFFF
# CHECK: movabsq $-9223372036854775808, %rax # imm = 0x8000000000000000
OpenPOWER on IntegriCloud