)]}'
{
  "id": "a47b537f554b5ab7f4104589b76fde56b0c598e1",
  "repo": "external/github.com/python/cpython",
  "revision": "64d036c8fe3cd4be0a3cbd23942ced14235adaeb",
  "path": "Parser/tokenizer.h"
}
