)]}'
{
  "id": "463c0e00ca1411af5624ab82881011a42bbc992c",
  "repo": "external/github.com/python/cpython",
  "revision": "3c67ec394faac79d260804d569a18fab43018af0",
  "path": "Parser/tokenizer.c"
}
