)]}'
{
  "id": "ae3874b09de48e030e0250aff4bcf2068a1b528b",
  "repo": "external/github.com/python/cpython",
  "revision": "7076bef8ba8836a19d5033f4ceb8eb9837de2301",
  "path": "Parser/tokenizer.c"
}
