)]}'
{
  "id": "e9423333e248bdd2d993932348d5bebb06476bf7",
  "repo": "external/github.com/python/cpython",
  "revision": "02bedcd36dda52f3b97a78ebaa7ecd6e94df7c1d",
  "path": "Parser/tokenizer.h"
}
