)]}'
{
  "id": "92669bfd8a1607fbe450ee9f904f949dac9075f4",
  "repo": "external/github.com/python/cpython",
  "revision": "2ec70102066fe5534f1a62e8f496d2005e1697db",
  "path": "Parser/tokenizer.h"
}
