)]}'
{
  "commit": "0af34bd2ab8a2cb480a78375ae013db3f13194dd",
  "tree": "f4c8ce3b76f97b0128c34132134d66bf559a9c61",
  "parents": [
    "8d9638e688cdf99293cac13c76b5f0d87277a7e8"
  ],
  "author": {
    "name": "Max191",
    "email": "44243577+Max191@users.noreply.github.com",
    "time": "Thu Dec 21 12:33:26 2023 -0500"
  },
  "committer": {
    "name": "GitHub",
    "email": "noreply@github.com",
    "time": "Thu Dec 21 17:33:26 2023 +0000"
  },
  "message": "[GlobalOpt] Add quantized matmul reassociation support for f16 types (#15964)\n\nThis adds support for reassociating f16 typed quantized matmuls, fixing\r\na bug reported in https://github.com/openxla/iree/issues/15661.",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "7ec2388020b4d584d528d5e1ba430676a83877d1",
      "old_mode": 33188,
      "old_path": "compiler/src/iree/compiler/GlobalOptimization/FuseDequantizationMatmul.cpp",
      "new_id": "aa3473b5cbf158c699fa8c5174f58b3e87ad2fb3",
      "new_mode": 33188,
      "new_path": "compiler/src/iree/compiler/GlobalOptimization/FuseDequantizationMatmul.cpp"
    },
    {
      "type": "modify",
      "old_id": "52f6e09ccad4e4bba1951e94f7913c94a70fb413",
      "old_mode": 33188,
      "old_path": "compiler/src/iree/compiler/GlobalOptimization/test/fuse_dequantization_matmul.mlir",
      "new_id": "2b596897aababa74faac49fc7e0e9f7b01bff8a0",
      "new_mode": 33188,
      "new_path": "compiler/src/iree/compiler/GlobalOptimization/test/fuse_dequantization_matmul.mlir"
    }
  ]
}
