)]}'
{
  "id": "b783e807844cc495d165879208a063bfcef384c4",
  "repo": "external/github.com/python/cpython",
  "revision": "2b110f97132668bbf19622de4b97ac022db98f74",
  "path": "Parser/tokenizer.c"
}
