Disable TD tests on CUDA backends due to failure. (#17041)

See Issue https://github.com/openxla/iree/issues/17039
diff --git a/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir b/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
index 9293fd4..dfd1d92 100644
--- a/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/double_mma_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: * 
 func.func @double_matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<16x16xf16>, %second : tensor<16x8xf16>) -> tensor<16x8xf16> {
   %c0 = arith.constant 0.0 : f16
   %0 = tensor.empty() : tensor<16x16xf16>
diff --git a/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
index c1b1c29..ceaa510 100644
--- a/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_elemwise_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: * 
 #map = affine_map<(d0, d1) -> (d0, d1)>
 func.func @matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<8x16xf16>, %bias : tensor<16x8xf16>) -> tensor<16x8xf16> {
   %c0 = arith.constant 0.0 : f16
diff --git a/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
index f01f07c..2444f07 100644
--- a/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_reduction_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: * 
 #map = affine_map<(d0, d1) -> (d0, d1)>
 #map1 = affine_map<(d0, d1) -> (d0)>
 func.func @matmul_reduction(%lhs : tensor<16x16xf16>, %rhs : tensor<16x16xf16>) -> tensor<16x16xf16> {
@@ -9,7 +11,9 @@
   %1 = linalg.fill ins(%c0 : f16) outs(%0 : tensor<16x16xf16>) -> tensor<16x16xf16>
   %2 = linalg.matmul_transpose_b ins(%lhs, %rhs : tensor<16x16xf16>, tensor<16x16xf16>)
       outs(%1 : tensor<16x16xf16>) -> tensor<16x16xf16>
-  %6 = linalg.generic {indexing_maps = [#map, #map1], iterator_types = ["parallel", "reduction"]}
+  %6 = linalg.generic {indexing_maps// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: * 
+ = [#map, #map1], iterator_types = ["parallel", "reduction"]}
         ins(%2 : tensor<16x16xf16>) outs(%init : tensor<16xf16>) {
         ^bb0(%in: f16, %out: f16):
           %20 = arith.maximumf %in, %out : f16
diff --git a/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir b/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
index 5c0e240..02bb863 100644
--- a/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
+++ b/tests/transform_dialect/cuda/mma_using_layout_analysis.mlir
@@ -1,3 +1,5 @@
+// XFAILED due to failure with https://github.com/openxla/iree/pull/16665 (Issues: https://github.com/openxla/iree/issues/17039)
+// XFAIL: * 
 func.func @matmul(%lhs : tensor<16x16xf16>, %rhs : tensor<16x8xf16>) -> tensor<16x8xf16> {
   %c0 = arith.constant 0.0 : f16
   %0 = tensor.empty() : tensor<16x8xf16>