)]}'
{
  "id": "f15e2523f7a88bc23c27a9e9614283b74b14e12e",
  "repo": "external/github.com/python/cpython",
  "revision": "482f908410060eebed6c5cdadf967d4a0b22d114",
  "path": "Parser/tokenizer.h"
}
