)]}'
{
  "id": "2be3bf2e3158999a4ba3b23c99f768c54b9f9e14",
  "repo": "external/github.com/python/cpython",
  "revision": "ef0faa5c59cc5252f55cf347c26706337eeb9ba1",
  "path": "Parser/tokenizer.h"
}
