)]}'
{
  "id": "952265eb923f9d0189e7d8f744d28d659afe9768",
  "repo": "external/github.com/python/cpython",
  "revision": "7ad6f74fcf9db1ccfeaf0986064870d8d3887300",
  "path": "Parser/tokenizer.c"
}
