)]}'
{
  "id": "61f0a6138f676bed1b8f87af2e97288e384e288c",
  "repo": "external/github.com/python/cpython",
  "revision": "ec80ffe39acc41ab43bec33d6fa9900e8eff37e3",
  "path": "Parser/tokenizer.h"
}
