@@ -36,7 +36,7 @@ define <vscale x 8 x half> @famin_u_f16(<vscale x 8 x i1> %pg, <vscale x 8 x hal
3636; CHECK: // %bb.0:
3737; CHECK-NEXT: famin z0.h, p0/m, z0.h, z1.h
3838; CHECK-NEXT: ret
39- %r = call <vscale x 8 x half > @llvm.aarch64.sve.famin.u.nxv8f16 (<vscale x 8 x i1 > %pg , <vscale x 8 x half > %a , <vscale x 8 x half > %b )
39+ %r = call <vscale x 8 x half > @llvm.aarch64.sve.famin.u.nxv8f16 (<vscale x 8 x i1 > %pg , <vscale x 8 x half > %b , <vscale x 8 x half > %a )
4040 ret <vscale x 8 x half > %r
4141}
4242
@@ -45,7 +45,7 @@ define <vscale x 4 x float> @famin_u_f32(<vscale x 4 x i1> %pg, <vscale x 4 x fl
4545; CHECK: // %bb.0:
4646; CHECK-NEXT: famin z0.s, p0/m, z0.s, z1.s
4747; CHECK-NEXT: ret
48- %r = call <vscale x 4 x float > @llvm.aarch64.sve.famin.u.nxv4f32 (<vscale x 4 x i1 > %pg , <vscale x 4 x float > %a , <vscale x 4 x float > %b )
48+ %r = call <vscale x 4 x float > @llvm.aarch64.sve.famin.u.nxv4f32 (<vscale x 4 x i1 > %pg , <vscale x 4 x float > %b , <vscale x 4 x float > %a )
4949 ret <vscale x 4 x float > %r
5050}
5151
@@ -54,7 +54,7 @@ define <vscale x 2 x double> @famin_u_f64(<vscale x 2 x i1> %pg, <vscale x 2 x d
5454; CHECK: // %bb.0:
5555; CHECK-NEXT: famin z0.d, p0/m, z0.d, z1.d
5656; CHECK-NEXT: ret
57- %r = call <vscale x 2 x double > @llvm.aarch64.sve.famin.u.nxv2f64 (<vscale x 2 x i1 > %pg , <vscale x 2 x double > %a , <vscale x 2 x double > %b )
57+ %r = call <vscale x 2 x double > @llvm.aarch64.sve.famin.u.nxv2f64 (<vscale x 2 x i1 > %pg , <vscale x 2 x double > %b , <vscale x 2 x double > %a )
5858 ret <vscale x 2 x double > %r
5959}
6060
@@ -90,7 +90,7 @@ define <vscale x 8 x half> @famax_u_f16(<vscale x 8 x i1> %pg, <vscale x 8 x hal
9090; CHECK: // %bb.0:
9191; CHECK-NEXT: famax z0.h, p0/m, z0.h, z1.h
9292; CHECK-NEXT: ret
93- %r = call <vscale x 8 x half > @llvm.aarch64.sve.famax.u.nxv8f16 (<vscale x 8 x i1 > %pg , <vscale x 8 x half > %a , <vscale x 8 x half > %b )
93+ %r = call <vscale x 8 x half > @llvm.aarch64.sve.famax.u.nxv8f16 (<vscale x 8 x i1 > %pg , <vscale x 8 x half > %b , <vscale x 8 x half > %a )
9494 ret <vscale x 8 x half > %r
9595}
9696
@@ -99,7 +99,7 @@ define <vscale x 4 x float> @famax_u_f32(<vscale x 4 x i1> %pg, <vscale x 4 x fl
9999; CHECK: // %bb.0:
100100; CHECK-NEXT: famax z0.s, p0/m, z0.s, z1.s
101101; CHECK-NEXT: ret
102- %r = call <vscale x 4 x float > @llvm.aarch64.sve.famax.u.nxv4f32 (<vscale x 4 x i1 > %pg , <vscale x 4 x float > %a , <vscale x 4 x float > %b )
102+ %r = call <vscale x 4 x float > @llvm.aarch64.sve.famax.u.nxv4f32 (<vscale x 4 x i1 > %pg , <vscale x 4 x float > %b , <vscale x 4 x float > %a )
103103 ret <vscale x 4 x float > %r
104104}
105105
@@ -108,7 +108,7 @@ define <vscale x 2 x double> @famax_u_f64(<vscale x 2 x i1> %pg, <vscale x 2 x d
108108; CHECK: // %bb.0:
109109; CHECK-NEXT: famax z0.d, p0/m, z0.d, z1.d
110110; CHECK-NEXT: ret
111- %r = call <vscale x 2 x double > @llvm.aarch64.sve.famax.u.nxv2f64 (<vscale x 2 x i1 > %pg , <vscale x 2 x double > %a , <vscale x 2 x double > %b )
111+ %r = call <vscale x 2 x double > @llvm.aarch64.sve.famax.u.nxv2f64 (<vscale x 2 x i1 > %pg , <vscale x 2 x double > %b , <vscale x 2 x double > %a )
112112 ret <vscale x 2 x double > %r
113113}
114114
0 commit comments