)]}'
{
  "id": "8fded37ad998cbe17352458ec6739c5a66c29607",
  "repo": "external/github.com/python/cpython",
  "revision": "40986ed30e3ab7ebc2aea3ba0c28fe3fb565b29c",
  "path": "Parser/tokenizer.h"
}
