sext-in-reg.ll 2.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111
  1. ; RUN: llc < %s -march=nvptx -mcpu=sm_20 | FileCheck %s
  2. target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v16:16:16-v32:32:32-v64:64:64-v128:128:128-n16:32:64"
  3. define void @one(i64 %a, i64 %b, i64* %p1, i64* %p2) {
  4. ; CHECK: cvt.s64.s8
  5. ; CHECK: cvt.s64.s8
  6. entry:
  7. %sext = shl i64 %a, 56
  8. %conv1 = ashr exact i64 %sext, 56
  9. %sext1 = shl i64 %b, 56
  10. %conv4 = ashr exact i64 %sext1, 56
  11. %shr = ashr i64 %a, 16
  12. %shr9 = ashr i64 %b, 16
  13. %add = add nsw i64 %conv4, %conv1
  14. store i64 %add, i64* %p1, align 8
  15. %add17 = add nsw i64 %shr9, %shr
  16. store i64 %add17, i64* %p2, align 8
  17. ret void
  18. }
  19. define void @two(i64 %a, i64 %b, i64* %p1, i64* %p2) {
  20. entry:
  21. ; CHECK: cvt.s64.s32
  22. ; CHECK: cvt.s64.s32
  23. %sext = shl i64 %a, 32
  24. %conv1 = ashr exact i64 %sext, 32
  25. %sext1 = shl i64 %b, 32
  26. %conv4 = ashr exact i64 %sext1, 32
  27. %shr = ashr i64 %a, 16
  28. %shr9 = ashr i64 %b, 16
  29. %add = add nsw i64 %conv4, %conv1
  30. store i64 %add, i64* %p1, align 8
  31. %add17 = add nsw i64 %shr9, %shr
  32. store i64 %add17, i64* %p2, align 8
  33. ret void
  34. }
  35. define void @three(i64 %a, i64 %b, i64* %p1, i64* %p2) {
  36. entry:
  37. ; CHECK: cvt.s64.s16
  38. ; CHECK: cvt.s64.s16
  39. %sext = shl i64 %a, 48
  40. %conv1 = ashr exact i64 %sext, 48
  41. %sext1 = shl i64 %b, 48
  42. %conv4 = ashr exact i64 %sext1, 48
  43. %shr = ashr i64 %a, 16
  44. %shr9 = ashr i64 %b, 16
  45. %add = add nsw i64 %conv4, %conv1
  46. store i64 %add, i64* %p1, align 8
  47. %add17 = add nsw i64 %shr9, %shr
  48. store i64 %add17, i64* %p2, align 8
  49. ret void
  50. }
  51. define void @four(i32 %a, i32 %b, i32* %p1, i32* %p2) {
  52. entry:
  53. ; CHECK: cvt.s32.s8
  54. ; CHECK: cvt.s32.s8
  55. %sext = shl i32 %a, 24
  56. %conv1 = ashr exact i32 %sext, 24
  57. %sext1 = shl i32 %b, 24
  58. %conv4 = ashr exact i32 %sext1, 24
  59. %shr = ashr i32 %a, 16
  60. %shr9 = ashr i32 %b, 16
  61. %add = add nsw i32 %conv4, %conv1
  62. store i32 %add, i32* %p1, align 4
  63. %add17 = add nsw i32 %shr9, %shr
  64. store i32 %add17, i32* %p2, align 4
  65. ret void
  66. }
  67. define void @five(i32 %a, i32 %b, i32* %p1, i32* %p2) {
  68. entry:
  69. ; CHECK: cvt.s32.s16
  70. ; CHECK: cvt.s32.s16
  71. %sext = shl i32 %a, 16
  72. %conv1 = ashr exact i32 %sext, 16
  73. %sext1 = shl i32 %b, 16
  74. %conv4 = ashr exact i32 %sext1, 16
  75. %shr = ashr i32 %a, 16
  76. %shr9 = ashr i32 %b, 16
  77. %add = add nsw i32 %conv4, %conv1
  78. store i32 %add, i32* %p1, align 4
  79. %add17 = add nsw i32 %shr9, %shr
  80. store i32 %add17, i32* %p2, align 4
  81. ret void
  82. }
  83. define void @six(i16 %a, i16 %b, i16* %p1, i16* %p2) {
  84. entry:
  85. ; CHECK: cvt.s16.s8
  86. ; CHECK: cvt.s16.s8
  87. %sext = shl i16 %a, 8
  88. %conv1 = ashr exact i16 %sext, 8
  89. %sext1 = shl i16 %b, 8
  90. %conv4 = ashr exact i16 %sext1, 8
  91. %shr = ashr i16 %a, 8
  92. %shr9 = ashr i16 %b, 8
  93. %add = add nsw i16 %conv4, %conv1
  94. store i16 %add, i16* %p1, align 4
  95. %add17 = add nsw i16 %shr9, %shr
  96. store i16 %add17, i16* %p2, align 4
  97. ret void
  98. }