)]}'
{
  "commit": "96fff35325e519cc76ffacf22e57e4c393d4446f",
  "tree": "56d287fb561e70c42a79c9be294d744d478efe57",
  "parents": [
    "3821b92c1faf7e7058feeb0048511c946a841105"
  ],
  "author": {
    "name": "Marta Gómez Macías",
    "email": "mgmacias@google.com",
    "time": "Sun May 28 14:15:53 2023"
  },
  "committer": {
    "name": "GitHub",
    "email": "noreply@github.com",
    "time": "Sun May 28 14:15:53 2023"
  },
  "message": "gh-105017: Include CRLF lines in strings and column numbers (#105030)\n\nCo-authored-by: Pablo Galindo \u003cpablogsal@gmail.com\u003e",
  "tree_diff": [
    {
      "type": "modify",
      "old_id": "293592b3fd13db49e51026d1df501dc6d0cc1ce4",
      "old_mode": 33188,
      "old_path": "Lib/test/test_tokenize.py",
      "new_id": "cd11dddd0fe51abf95b1a34b1de93f54c6f34959",
      "new_mode": 33188,
      "new_path": "Lib/test/test_tokenize.py"
    },
    {
      "type": "add",
      "old_id": "0000000000000000000000000000000000000000",
      "old_mode": 0,
      "old_path": "/dev/null",
      "new_id": "02d653c2d658eb04e1a9f5db47489ac904b85d83",
      "new_mode": 33188,
      "new_path": "Misc/NEWS.d/next/Core and Builtins/2023-05-27-21-50-48.gh-issue-105017.4sDyDV.rst"
    },
    {
      "type": "modify",
      "old_id": "b031a6f5d440e85a833e47e2ea4ac99fd08dda29",
      "old_mode": 33188,
      "old_path": "Parser/pegen.c",
      "new_id": "b9894dd0acc54644683e3e7b3358b6a516dbe821",
      "new_mode": 33188,
      "new_path": "Parser/pegen.c"
    },
    {
      "type": "modify",
      "old_id": "a84c2492b6b17a085b4558afcdda54fcb9bc5fbb",
      "old_mode": 33188,
      "old_path": "Parser/tokenizer.c",
      "new_id": "59c817293fbfcd9c6bbd8aaee9d7972e024fc59e",
      "new_mode": 33188,
      "new_path": "Parser/tokenizer.c"
    },
    {
      "type": "modify",
      "old_id": "019f533ef2a2601671ecd96f82ef72110e9534fb",
      "old_mode": 33188,
      "old_path": "Parser/tokenizer.h",
      "new_id": "02749e355da81241ec152d89d9ec6a7ee9921ff3",
      "new_mode": 33188,
      "new_path": "Parser/tokenizer.h"
    },
    {
      "type": "modify",
      "old_id": "01c2215366a73633a7aede0ee52a45c6127823f6",
      "old_mode": 33188,
      "old_path": "Python/Python-tokenize.c",
      "new_id": "4eced66b6177085c946155ab07c42675903f0aee",
      "new_mode": 33188,
      "new_path": "Python/Python-tokenize.c"
    }
  ]
}
