Disable TD tests on CUDA backends due to failure. (#17041)
See Issue https://github.com/openxla/iree/issues/17039
diff --git a/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir b/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
index 9293fd4..dfd1d92 100644
--- a/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: *
func.func @double_matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<16x16xf16>, %second : tensor<16x8xf16>) -> tensor<16x8xf16> {
%c0 = arith.constant 0.0 : f16
%0 = tensor.empty() : tensor<16x16xf16>
diff --git a/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
index c1b1c29..ceaa510 100644
--- a/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: *
#map = affine_map<(d0, d1) -> (d0, d1)>
func.func @matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<8x16xf16>, %bias : tensor<16x8xf16>) -> tensor<16x8xf16> {
%c0 = arith.constant 0.0 : f16
diff --git a/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
index f01f07c..2444f07 100644
--- a/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: *
#map = affine_map<(d0, d1) -> (d0, d1)>
#map1 = affine_map<(d0, d1) -> (d0)>
func.func @matmul_reduction(%lhs : tensor<16x16xf16>, %rhs : tensor<16x16xf16>) -> tensor<16x16xf16> {
@@ -9,7 +11,9 @@
%1 = linalg.fill ins(%c0 : f16) outs(%0 : tensor<16x16xf16>) -> tensor<16x16xf16>
%2 = linalg.matmul_transpose_b ins(%lhs, %rhs : tensor<16x16xf16>, tensor<16x16xf16>)
outs(%1 : tensor<16x16xf16>) -> tensor<16x16xf16>
- %6 = linalg.generic {indexing_maps = [#map, #map1], iterator_types = ["parallel", "reduction"]}
+ %6 = linalg.generic {indexing_maps// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: *
+ = [#map, #map1], iterator_types = ["parallel", "reduction"]}
ins(%2 : tensor<16x16xf16>) outs(%init : tensor<16xf16>) {
^bb0(%in: f16, %out: f16):
%20 = arith.maximumf %in, %out : f16
diff --git a/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
index 5c0e240..02bb863 100644
--- a/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: *
func.func @matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<16x8xf16>) -> tensor<16x8xf16> {
%c0 = arith.constant 0.0 : f16
%0 = tensor.empty() : tensor<16x8xf16>