llvm-toolchain/debian/patches/add-arm-vcmpz-test.patch

318 lines
9.1 KiB
Diff

From b5caa68fb2444f0fc0e459be3b7a3da88a68b566 Mon Sep 17 00:00:00 2001
From: David Green <david.green@arm.com>
Date: Tue, 1 Nov 2022 15:00:56 +0000
Subject: [PATCH] [ARM] Tests for various NEON vector compares. NFC
---
llvm/test/CodeGen/ARM/vcmpz.ll | 301 +++++++++++++++++++++++++++++++++
1 file changed, 301 insertions(+)
create mode 100644 llvm/test/CodeGen/ARM/vcmpz.ll
diff --git a/llvm/test/CodeGen/ARM/vcmpz.ll b/llvm/test/CodeGen/ARM/vcmpz.ll
new file mode 100644
index 000000000000..f800346a6b56
--- /dev/null
+++ b/llvm/test/CodeGen/ARM/vcmpz.ll
@@ -0,0 +1,301 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=armv8-eabi -mattr=+neon | FileCheck %s
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_eq(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_eq:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q0, q0, q1
+; CHECK-NEXT: bx lr
+ %2 = icmp eq <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_ne(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_ne:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q8, q0, q1
+; CHECK-NEXT: vmvn q0, q8
+; CHECK-NEXT: bx lr
+ %2 = icmp ne <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_slt(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_slt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.s32 q0, q1, q0
+; CHECK-NEXT: bx lr
+ %2 = icmp slt <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_sle(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_sle:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.s32 q0, q1, q0
+; CHECK-NEXT: bx lr
+ %2 = icmp sle <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_sgt(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_sgt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.s32 q0, q0, q1
+; CHECK-NEXT: bx lr
+ %2 = icmp sgt <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_sge(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_sge:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.s32 q0, q0, q1
+; CHECK-NEXT: bx lr
+ %2 = icmp sge <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_ult(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_ult:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.u32 q0, q1, q0
+; CHECK-NEXT: bx lr
+ %2 = icmp ult <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_ule(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_ule:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.u32 q0, q1, q0
+; CHECK-NEXT: bx lr
+ %2 = icmp ule <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_ugt(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_ugt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.u32 q0, q0, q1
+; CHECK-NEXT: bx lr
+ %2 = icmp ugt <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_uge(<4 x i32> %0, <4 x i32> %b) {
+; CHECK-LABEL: vcmpz_uge:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.u32 q0, q0, q1
+; CHECK-NEXT: bx lr
+ %2 = icmp uge <4 x i32> %0, %b
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_eq(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_eq:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp eq <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_ne(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_ne:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q8, q0, #0
+; CHECK-NEXT: vmvn q0, q8
+; CHECK-NEXT: bx lr
+ %2 = icmp ne <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_slt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_slt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vclt.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp slt <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_sle(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_sle:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcle.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sle <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_sgt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_sgt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sgt <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_sge(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_sge:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sge <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_ult(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_ult:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vmov.i32 q0, #0x0
+; CHECK-NEXT: bx lr
+ %2 = icmp ult <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+;define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_ule(<4 x i32> %0) {
+; %2 = icmp ule <4 x i32> %0, zeroinitializer
+; %3 = sext <4 x i1> %2 to <4 x i32>
+; ret <4 x i32> %3
+;}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_ugt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_ugt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q8, q0, #0
+; CHECK-NEXT: vmvn q0, q8
+; CHECK-NEXT: bx lr
+ %2 = icmp ugt <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zr_uge(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zr_uge:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vmov.i8 q0, #0xff
+; CHECK-NEXT: bx lr
+ %2 = icmp uge <4 x i32> %0, zeroinitializer
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_eq(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_eq:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp eq <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_ne(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_ne:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q8, q0, #0
+; CHECK-NEXT: vmvn q0, q8
+; CHECK-NEXT: bx lr
+ %2 = icmp ne <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_slt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_slt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcgt.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp slt <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_sle(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_sle:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcge.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sle <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_sgt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_sgt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vclt.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sgt <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_sge(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_sge:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vcle.s32 q0, q0, #0
+; CHECK-NEXT: bx lr
+ %2 = icmp sge <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_ult(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_ult:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vceq.i32 q8, q0, #0
+; CHECK-NEXT: vmvn q0, q8
+; CHECK-NEXT: bx lr
+ %2 = icmp ult <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_ule(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_ule:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vmov.i8 q0, #0xff
+; CHECK-NEXT: bx lr
+ %2 = icmp ule <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_ugt(<4 x i32> %0) {
+; CHECK-LABEL: vcmpz_zl_ugt:
+; CHECK: @ %bb.0:
+; CHECK-NEXT: vmov.i32 q0, #0x0
+; CHECK-NEXT: bx lr
+ %2 = icmp ugt <4 x i32> zeroinitializer, %0
+ %3 = sext <4 x i1> %2 to <4 x i32>
+ ret <4 x i32> %3
+}
+
+;define arm_aapcs_vfpcc <4 x i32> @vcmpz_zl_uge(<4 x i32> %0) {
+; %2 = icmp uge <4 x i32> zeroinitializer, %0
+; %3 = sext <4 x i1> %2 to <4 x i32>
+; ret <4 x i32> %3
+;}