| 12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758 |
- ; RUN: llc < %s -march=nvptx -mcpu=sm_20 | FileCheck --check-prefix=SM20 %s
- ; RUN: llc < %s -march=nvptx -mcpu=sm_35 | FileCheck --check-prefix=SM35 %s
- declare i32 @llvm.nvvm.rotate.b32(i32, i32)
- declare i64 @llvm.nvvm.rotate.b64(i64, i32)
- declare i64 @llvm.nvvm.rotate.right.b64(i64, i32)
- ; SM20: rotate32
- ; SM35: rotate32
- define i32 @rotate32(i32 %a, i32 %b) {
- ; SM20: shl.b32
- ; SM20: sub.s32
- ; SM20: shr.b32
- ; SM20: add.u32
- ; SM35: shf.l.wrap.b32
- %val = tail call i32 @llvm.nvvm.rotate.b32(i32 %a, i32 %b)
- ret i32 %val
- }
- ; SM20: rotate64
- ; SM35: rotate64
- define i64 @rotate64(i64 %a, i32 %b) {
- ; SM20: shl.b64
- ; SM20: sub.u32
- ; SM20: shr.b64
- ; SM20: add.u64
- ; SM35: shf.l.wrap.b32
- ; SM35: shf.l.wrap.b32
- %val = tail call i64 @llvm.nvvm.rotate.b64(i64 %a, i32 %b)
- ret i64 %val
- }
- ; SM20: rotateright64
- ; SM35: rotateright64
- define i64 @rotateright64(i64 %a, i32 %b) {
- ; SM20: shr.b64
- ; SM20: sub.u32
- ; SM20: shl.b64
- ; SM20: add.u64
- ; SM35: shf.r.wrap.b32
- ; SM35: shf.r.wrap.b32
- %val = tail call i64 @llvm.nvvm.rotate.right.b64(i64 %a, i32 %b)
- ret i64 %val
- }
- ; SM20: rotl0
- ; SM35: rotl0
- define i32 @rotl0(i32 %x) {
- ; SM20: shl.b32
- ; SM20: shr.b32
- ; SM20: add.u32
- ; SM35: shf.l.wrap.b32
- %t0 = shl i32 %x, 8
- %t1 = lshr i32 %x, 24
- %t2 = or i32 %t0, %t1
- ret i32 %t2
- }
|