)]}'
{
  "id": "ce8129ddd123f0406ef9e88c932a4e24813489e8",
  "repo": "external/github.com/python/cpython",
  "revision": "863b7496e79e044245ca7687df941be840d02d78",
  "path": "Parser/tokenizer.c"
}
