)]}'
{
  "id": "1ce6eeba8ca59b01562aca2648ebc77cca1263ce",
  "repo": "external/github.com/python/cpython",
  "revision": "2205da43a6bb68b641d97fa7832b71c8a57efb77",
  "path": "Parser/tokenizer.h"
}
