)]}'
{
  "id": "13b666ca3b79ed631ce92546edbb1072e8b27139",
  "repo": "external/github.com/python/cpython",
  "revision": "72dde1016493c52abe857fc4a7bf6c40138b4114",
  "path": "Parser/tokenizer.c"
}
