diff options
Diffstat (limited to 'test/CodeGen/AMDGPU/inline-asm.ll')
-rw-r--r-- | test/CodeGen/AMDGPU/inline-asm.ll | 174 |
1 files changed, 168 insertions, 6 deletions
diff --git a/test/CodeGen/AMDGPU/inline-asm.ll b/test/CodeGen/AMDGPU/inline-asm.ll index 9c8d3534f8ad..1f5b8be15e2e 100644 --- a/test/CodeGen/AMDGPU/inline-asm.ll +++ b/test/CodeGen/AMDGPU/inline-asm.ll @@ -1,7 +1,7 @@ -; RUN: llc < %s -march=amdgcn -mcpu=SI -verify-machineinstrs | FileCheck %s -; RUN: llc < %s -march=amdgcn -mcpu=tonga -verify-machineinstrs | FileCheck %s +; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck %s +; RUN: llc -march=amdgcn -mcpu=tonga -verify-machineinstrs < %s | FileCheck %s -; CHECK: {{^}}inline_asm: +; CHECK-LABEL: {{^}}inline_asm: ; CHECK: s_endpgm ; CHECK: s_endpgm define void @inline_asm(i32 addrspace(1)* %out) { @@ -11,13 +11,175 @@ entry: ret void } -; CHECK: {{^}}inline_asm_shader: +; CHECK-LABEL: {{^}}inline_asm_shader: ; CHECK: s_endpgm ; CHECK: s_endpgm -define void @inline_asm_shader() #0 { +define amdgpu_ps void @inline_asm_shader() { entry: call void asm sideeffect "s_endpgm", ""() ret void } -attributes #0 = { "ShaderType"="0" } + +; CHECK: {{^}}branch_on_asm: +; Make sure inline assembly is treted as divergent. +; CHECK: s_mov_b32 s{{[0-9]+}}, 0 +; CHECK: s_and_saveexec_b64 +define void @branch_on_asm(i32 addrspace(1)* %out) { + %zero = call i32 asm "s_mov_b32 $0, 0", "=s"() + %cmp = icmp eq i32 %zero, 0 + br i1 %cmp, label %if, label %endif + +if: + store i32 0, i32 addrspace(1)* %out + br label %endif + +endif: + ret void +} + +; CHECK-LABEL: {{^}}v_cmp_asm: +; CHECK: v_mov_b32_e32 [[SRC:v[0-9]+]], s{{[0-9]+}} +; CHECK: v_cmp_ne_i32_e64 s{{\[}}[[MASK_LO:[0-9]+]]:[[MASK_HI:[0-9]+]]{{\]}}, 0, [[SRC]] +; CHECK-DAG: v_mov_b32_e32 v[[V_LO:[0-9]+]], s[[MASK_LO]] +; CHECK-DAG: v_mov_b32_e32 v[[V_HI:[0-9]+]], s[[MASK_HI]] +; CHECK: buffer_store_dwordx2 v{{\[}}[[V_LO]]:[[V_HI]]{{\]}} +define void @v_cmp_asm(i64 addrspace(1)* %out, i32 %in) { + %sgpr = tail call i64 asm "v_cmp_ne_i32_e64 $0, 0, $1", "=s,v"(i32 %in) + store i64 %sgpr, i64 addrspace(1)* %out + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm: +; CHECK: codeLenInByte = 12 +define void @code_size_inline_asm(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "v_nop_e64", ""() + ret void +} + +; All inlineasm instructions are assumed to be the maximum size +; CHECK-LABEL: {{^}}code_size_inline_asm_small_inst: +; CHECK: codeLenInByte = 12 +define void @code_size_inline_asm_small_inst(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "v_nop_e32", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst: +; CHECK: codeLenInByte = 20 +define void @code_size_inline_asm_2_inst(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect " + v_nop_e64 + v_nop_e64 + ", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_2_inst_extra_newline: +; CHECK: codeLenInByte = 20 +define void @code_size_inline_asm_2_inst_extra_newline(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect " + v_nop_e64 + + v_nop_e64 + ", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_0_inst: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_0_inst(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_1_comment(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; comment", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_newline_1_comment: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_newline_1_comment(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect " +; comment", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_1_comment_newline: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_1_comment_newline(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; comment +", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_2_comments_line(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; first comment ; second comment", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_2_comments_line_nospace: +; CHECK: codeLenInByte = 4 +define void @code_size_inline_asm_2_comments_line_nospace(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; first comment;second comment", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments0: +; CHECK: codeLenInByte = 20 +define void @code_size_inline_asm_mixed_comments0(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; comment + v_nop_e64 ; inline comment +; separate comment + v_nop_e64 + + ; trailing comment + ; extra comment + ", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments1: +; CHECK: codeLenInByte = 20 +define void @code_size_inline_asm_mixed_comments1(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "v_nop_e64 ; inline comment +; separate comment + v_nop_e64 + + ; trailing comment + ; extra comment + ", ""() + ret void +} + +; CHECK-LABEL: {{^}}code_size_inline_asm_mixed_comments_operands: +; CHECK: codeLenInByte = 20 +define void @code_size_inline_asm_mixed_comments_operands(i32 addrspace(1)* %out) { +entry: + call void asm sideeffect "; comment + v_add_i32_e32 v0, vcc, v1, v2 ; inline comment +; separate comment + v_bfrev_b32_e32 v0, 1 + + ; trailing comment + ; extra comment + ", ""() + ret void +} |