)]}'
{
  "commit": "9adee084ecf1244a55bca2b45d5a5b6e36b7ec6e",
  "tree": "9866621a732288db3d3f2b6109d25aa07ff6c89e",
  "parents": [
    "20fd5b23ac305918cf391158e931573c3507b101"
  ],
  "author": {
    "name": "RJ Ascani",
    "email": "rjascani@google.com",
    "time": "Wed Feb 28 10:45:56 2024 -0800"
  },
  "committer": {
    "name": "GitHub",
    "email": "noreply@github.com",
    "time": "Wed Feb 28 18:45:56 2024 +0000"
  },
  "message": "Remove experimental flag for per-channel FC quantization (#2482)\n\nPer-channel quantization in fully connected layers are still not supported by TFLM, but the converter now has proper support so we can remove the flag.\n\nBUG\u003dcl/610755484",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "04bbb324729c6ac3f817fc38710d8e5958ee288f",
      "old_mode": 33188,
      "old_path": "tensorflow/lite/micro/tools/requantize_flatbuffer_test.py",
      "new_id": "4d80991b0b119fb47ac75951d9e7a8d33143390d",
      "new_mode": 33188,
      "new_path": "tensorflow/lite/micro/tools/requantize_flatbuffer_test.py"
    }
  ]
}
