)]}'
{
  "id": "bfae39f311fc3f70fa0270f0d81986127e645986",
  "repo": "external/github.com/python/cpython",
  "revision": "fd71b9e9d496caa510dec56a9b69966558d6ba5d",
  "path": "Parser/tokenizer.h"
}
