)]}'
{
  "commit": "781be38b91e88628134fa11d0fa28fdc287c2e6e",
  "tree": "c1a005c345ba008faf4f6c5d5c0ddbe3df994701",
  "parents": [
    "6a82eb5065742fa4498b0bc85eb5f01271b20a65"
  ],
  "author": {
    "name": "zjgarvey",
    "email": "47986913+zjgarvey@users.noreply.github.com",
    "time": "Tue Jul 16 10:33:18 2024 -0700"
  },
  "committer": {
    "name": "GitHub",
    "email": "noreply@github.com",
    "time": "Tue Jul 16 10:33:18 2024 -0700"
  },
  "message": "Add torch-fuse-quantized-ops pass to the torch-to-iree pipeline (#17908)\n\nThe torch to iree pipeline currently does not use\r\n`--torch-fuse-quantized-ops`, which is the cause of significant\r\ndiscrepancies between model testing with iree-compile from torch IR and\r\nmodel testing which first lowers to linalg with torch-mlir before\r\ncompiling. Together with `--torch-fuse-quantized-ops`, a newer pass\r\n`--torch-scalarize-shapes` is added to the `torch-to-iree` pipeline to\r\nkeep in line with the\r\n`--torch-backend-to-linalg-on-tensors-backend-pipeline`.\r\n\r\n---------\r\n\r\nSigned-off-by: zjgarvey \u003czjgarvey@gmail.com\u003e",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "2dc4a9291d0c6788705cce18906959245f49e40f",
      "old_mode": 33188,
      "old_path": "compiler/plugins/input/Torch/InputConversion/Passes.cpp",
      "new_id": "293921892b26d601db0bdf63b1b60886a39f52de",
      "new_mode": 33188,
      "new_path": "compiler/plugins/input/Torch/InputConversion/Passes.cpp"
    }
  ]
}
