)]}'
{
  "id": "612cb23545f4d884c652fbccf1244c3230c1bb11",
  "repo": "external/github.com/python/cpython",
  "revision": "f205f1000a2d7f8b044caf281041b3705f293480",
  "path": "Parser/tokenizer.c"
}
