)]}'
{
  "id": "e10972cc99d05e59a662b59d0853490bd9eda015",
  "repo": "external/github.com/python/cpython",
  "revision": "4348a25665b2f09f76a605bab507b4edacc4dd24",
  "path": "Parser/tokenizer.h"
}
